Attached accelerator selection and placement

    公开(公告)号:US11494621B2

    公开(公告)日:2022-11-08

    申请号:US16020788

    申请日:2018-06-27

    Abstract: Implementations detailed herein include description of a computer-implemented method. In an implementation, the method at least includes receiving an application instance configuration, an application of the application instance to utilize a portion of an attached accelerator during execution of a machine learning model and the application instance configuration including an arithmetic precision of the machine learning model to be used in determining the portion of the accelerator to provision; provisioning the application instance and the portion of the accelerator attached to the application instance, wherein the application instance is implemented using a physical compute instance in a first location, wherein the portion of the accelerator is implemented using a physical accelerator in the second location; loading the machine learning model onto the portion of the accelerator; and performing inference using the loaded machine learning model of the application using the portion of the accelerator on the attached accelerator.

    Image compression and decompression using embeddings

    公开(公告)号:US10652565B1

    公开(公告)日:2020-05-12

    申请号:US15782725

    申请日:2017-10-12

    Abstract: A processing device receives a representation of an image, wherein the image has a first size and the representation has a second size that is smaller than the first size, the representation having been generated from the image by a first portion of a first trained machine learning model. The processing device processes the representation of the image using a second portion of the trained machine learning model to generate a reconstruction of the image and then outputs the reconstruction of the image.

    Incremental block compression
    28.
    发明授权

    公开(公告)号:US10579591B1

    公开(公告)日:2020-03-03

    申请号:US15385740

    申请日:2016-12-20

    Abstract: Techniques for performing incremental block compression using a processor are described herein. The processor receives a request to compress input data, the request including compression parameters for the compression and a target block size. The processor divides the input data into portions. The processor iteratively compresses the input data to an output block, until compressing another portion of data would increase a file size of the output block over a threshold value that is based at least on the target block size.

    Deep learning processing of video
    29.
    发明授权

    公开(公告)号:US10460175B1

    公开(公告)日:2019-10-29

    申请号:US15624258

    申请日:2017-06-15

    Abstract: A method and system for processing multiple frames of a video by a neural network are provided. Two frames of a video may be analyzed to determine if at least a portion of the layer-by-layer processing by a neural network can be skipped or terminated. Processing of a first frame of the video is performed by the neural network. A next frame of the video is processed by the neural network, such that processing of fewer layers (or sets of operations) of the neural network is performed if the first frame and the second frame are substantially similar.

    Random access to decompressed blocks

    公开(公告)号:US10366026B1

    公开(公告)日:2019-07-30

    申请号:US15390250

    申请日:2016-12-23

    Abstract: A system comprises a data storage, a decompression accelerator configured to decompress compressed data and thereby generate decompressed data, and a direct memory access (DMA) engine coupled to the data storage and the decompression accelerator. The DMA engine comprises a buffer for storage of a plurality of descriptors containing configuration parameters for a block of compressed data to be retrieved from the data storage and decompressed by the decompression accelerator, wherein at least one of the descriptors comprises a threshold value. The DMA engine, in accordance with one or more of the descriptors, is configured to read compressed data from data storage and transmit the threshold value and the compressed data to the decompression accelerator. The decompression accelerator is configured to decompress the compressed data until the threshold value is reached and then to abort further data decompression and to assert a stop transaction signal to the DMA engine.

Patent Agency Ranking