SELF-ADAPTABLE ACCELERATORS HAVING ALTERNATING PRODUCTION/OPTIMIZING MODES

    公开(公告)号:US20240362031A1

    公开(公告)日:2024-10-31

    申请号:US18308275

    申请日:2023-04-27

    CPC classification number: G06F9/44505 G06F11/3495

    Abstract: Systems and methods are provided for an accelerator system that includes a baseline (production) accelerator, optimizing accelerator, and control hardware accelerator, and an operation of alternatingly switching the production/optimizing accelerators between production and optimizing. With two production/optimizing accelerators, at any given point in time, one accelerator adapts while another accelerator processes data. Once the second accelerator starts doing a better job (e.g., has adapted to data drift), the accelerators change their modes, and the trainable accelerator becomes the “optimized” one. The accelerators do this non-stop, thus maintaining redundancy, providing expected quality of service (QOS) and adapting to data/concept drift.

    SYSTEMS AND METHODS FOR INTELLIGENT DATA SHUFFLING FOR HIGH-PERFORMANCE DISTRIBUTED MACHINE LEARNING TRAINING

    公开(公告)号:US20220067577A1

    公开(公告)日:2022-03-03

    申请号:US17010744

    申请日:2020-09-02

    Abstract: Systems and methods are provided for data shuffling for distributed machine learning training, including each training node in the network receiving a shard of training data, wherein the training data set is divided into shards having data items. Each data item is assigned to a working set such that each of the working set includes data items from multiple shards. The training nodes perform training using the data items of a first working set that are in each node's shard. Upon completion of the training using the data items of the first working set, the training nodes performing training using the data items of a second working set that are in their shards; and while the training nodes are performing training on their respective subsets of shards of the second working set, the nodes randomly shuffling data items in the first working set to create a shuffled first working set.

    Adjustable Precision for Multi-Stage Compute Processes

    公开(公告)号:US20200042287A1

    公开(公告)日:2020-02-06

    申请号:US16052218

    申请日:2018-08-01

    Abstract: Disclosed techniques provide for dynamically changing precision of a multi-stage compute process. For example, changing neural network (NN) parameters on a per-layer basis depending on properties of incoming data streams and per-layer performance of an NN among other considerations. NNs include multiple layers that may each be calculated with a different degree of accuracy and therefore, compute resource overhead (e.g., memory, processor resources, etc.). NNs are usually trained with 32-bit or 16-bit floating-point numbers. Once trained, an NN may be deployed in production. One approach to reduce compute overhead is to reduce parameter precision of NNs to 16 or 8 for deployment. The conversion to an acceptable lower precision is usually determined manually before deployment and precision levels are fixed while deployed. Disclosed techniques and implementations address automatic rather than manual determination or precision levels for different stages and dynamically adjusting precision for each stage at run-time.

Patent Agency Ranking