INLINE DECOMPRESSION
    1.
    发明申请

    公开(公告)号:US20240413839A1

    公开(公告)日:2024-12-12

    申请号:US18807785

    申请日:2024-08-16

    Abstract: Techniques and apparatuses to decompress data that has been stack compressed is described. Stack compression refers to compression of data in one or more dimensions. For uncompressed data blocks that are very sparse, i.e., data blocks that contain many zeros, stack compression can be effective. In stack compression, uncompressed data block is compressed into compressed data block by removing one or more zero words from the uncompressed data block. A map metadata that maps the zero words of the uncompressed data block is generated during compression. With the use of the map metadata, the compressed data block can be decompressed to restore the uncompressed data block.

    SINGLE INSTRUCTION MULTIPLE DATA (SIMD) SPARSE DECOMPRESSION WITH VARIABLE DENSITY

    公开(公告)号:US20240118902A1

    公开(公告)日:2024-04-11

    申请号:US18339797

    申请日:2023-06-22

    CPC classification number: G06F9/3887 G06F9/30178

    Abstract: An aspect of the disclosure relates to a data processing system, including: an input medium configured to include a first set of blocks of data including a first set of block of compressed data and a first set of metadata, respectively; an output medium configured to include a first set of blocks of decompressed data each having a predetermined number of decompressed elements; and a set of single instruction multiple data (SIMD) processors configured to: access the first set of blocks of data from the input medium, respectively; decompress the first set of blocks of compressed data to generate the first set of blocks of decompressed data based on the first set of metadata, respectively; and provide the first set of blocks of decompressed data to the output medium, respectively.

    SPARSITY-BASED NEURAL NETWORK MAPPING TO COMPUTING UNITS IN A SYSTEM-ON-CHIP

    公开(公告)号:US20220284271A1

    公开(公告)日:2022-09-08

    申请号:US17194202

    申请日:2021-03-05

    Abstract: A method for an artificial neural network includes receiving a set of input values to be convolved with multiple kernels via multiple computing units. One or more thermally-stressed computing units of the multiple computing units are determined. The multiple kernels are mapped to the multiple computing units of a system-on-chip (SOC) based on the one or more thermally-stressed computing units. A convolution is performed on the set of input values and a most sparse kernel of the multiple kernels on the most thermally-stressed computing unit.

    INLINE DECOMPRESSION
    4.
    发明申请

    公开(公告)号:US20210351789A1

    公开(公告)日:2021-11-11

    申请号:US16870873

    申请日:2020-05-08

    Abstract: Techniques and apparatuses to decompress data that has been stack compressed is described. Stack compression refers to compression of data in one or more dimensions. For uncompressed data blocks that are very sparse, i.e., data blocks that contain many zeros, stack compression can be effective. In stack compression, uncompressed data block is compressed into compressed data block by removing one or more zero words from the uncompressed data block. A map metadata that maps the zero words of the uncompressed data block is generated during compression. With the use of the map metadata, the compressed data block can be decompressed to restore the uncompressed data block.

    INLINE DECOMPRESSION
    5.
    发明公开

    公开(公告)号:US20230223954A1

    公开(公告)日:2023-07-13

    申请号:US17997619

    申请日:2021-05-07

    CPC classification number: H03M7/70 H03M7/3066 H03M7/6005 G06N3/0495

    Abstract: Techniques and apparatuses to decompress data that has been stack compressed is described. Stack compression refers to compression of data in one or more dimensions. For uncompressed data blocks that are very sparse, i.e., data blocks that contain many zeros, stack compression can be effective. In stack compression, uncompressed data block is compressed into compressed data block by removing one or more zero words from the uncompressed data block. A map metadata that maps the zero words of the uncompressed data block is generated during compression. With the use of the map metadata, the compressed data block can be decompressed to restore the uncompressed data block.

    SPLIT NETWORK ACCELERATION ARCHITECTURE
    6.
    发明申请

    公开(公告)号:US20200250545A1

    公开(公告)日:2020-08-06

    申请号:US16783047

    申请日:2020-02-05

    Abstract: A method for accelerating machine learning on a computing device is described. The method includes hosting a neural network in a first inference accelerator and a second inference accelerator. The neural network split between the first inference accelerator and the second inference accelerator. The method also includes routing intermediate inference request results directly between the first inference accelerator and the second inference accelerator. The method further includes generating a final inference request result from the intermediate inference request results.

    MEMORY STORAGE FORMAT FOR SUPPORTING MACHINE LEARNING ACCELERATION

    公开(公告)号:US20240095872A1

    公开(公告)日:2024-03-21

    申请号:US17946753

    申请日:2022-09-16

    CPC classification number: G06F3/064 G06F3/0604 G06F3/0673

    Abstract: A processor-implemented method for a memory storage format to accelerate machine learning (ML) on a computing device is described. The method includes receiving an image in a first layer storage format of a neural network. The method also includes assigning addresses to image pixels of each of three channels of the first layer storage format for accessing the image pixels in a blocked ML storage acceleration format. The method further includes storing the image pixels in the blocked ML storage acceleration format according to the assigned addresses of the image pixels. The method also includes accelerating inference video processing of the image according to the assigned addresses for the image pixels corresponding to the blocked ML storage acceleration format.

Patent Agency Ranking