Caching techniques for deep learning accelerator

    公开(公告)号:US12094531B2

    公开(公告)日:2024-09-17

    申请号:US17146314

    申请日:2021-01-11

    CPC classification number: G11C11/54 G06F12/0862 G06F12/0897 G06N3/063 G06N3/08

    Abstract: Systems, devices, and methods related to a Deep Learning Accelerator and memory are described. For example, the accelerator can have processing units to perform at least matrix computations of an artificial neural network via execution of instructions. The processing units have a local memory store operands of the instructions. The accelerator can access a random access memory via a system buffer, or without going through the system buffer. A fetch instruction can request an item, available at a memory address in the random access memory, to be loaded into the local memory at a local address. The fetch instruction can include a hint for the caching of the item in the system buffer. During execution of the instruction, the hint can be used to determine whether to load the item through the system buffer or to bypass the system buffer in loading the item.

    Telemetry-capable memory sub-system

    公开(公告)号:US12061551B2

    公开(公告)日:2024-08-13

    申请号:US17896883

    申请日:2022-08-26

    CPC classification number: G06F12/0815 G06F2212/1032 G06F2212/305

    Abstract: An access counter associated with a segment of a memory device is maintained. An access notification for a first line of the segment is received. An access type associated with the access notification is identified. A first value of the access counter is changed by a second value based on the access type. Based on the first value of the access counter, a memory management scheme is implemented.

    Delta predictions for page scheduling

    公开(公告)号:US12007899B2

    公开(公告)日:2024-06-11

    申请号:US17867371

    申请日:2022-07-18

    CPC classification number: G06F12/0882 G06F12/0646

    Abstract: Disclosed in some examples are improved address prediction and memory preloading that leverages next-delta prediction and/or far-delta prediction for scheduling using a DNN. Previous memory access sequence data that identify one or more memory addresses previously accessed by one or more processors of a system may be processed and then converted into a sequence of delta values. The sequence of delta values are then mapped to one or more classes that are then input to a DNN. The DNN then outputs a predicted future class identifier sequence that represents addresses that the DNN predicts will be accessed by the processor in the future. The predicted future class identifier sequence is then converted back to a predicted delta value sequence and back into a set of one or more predicted addresses.

    ACCESS HEATMAP IMPLEMENTATIONS AT A HOST DEVICE

    公开(公告)号:US20230393744A1

    公开(公告)日:2023-12-07

    申请号:US17831242

    申请日:2022-06-02

    CPC classification number: G06F3/0611 G06F3/0659 G06F3/0653 G06F3/0679

    Abstract: Methods, systems, and devices for access heatmap implementations at a host device are described. A host device may leverage access operation monitoring that is performed at a memory device, including various examples of signaling and management of monitoring configurations. For example, a memory device may maintain a storage location for tracking access operation occurrence, for which access operations of a given address may be mapped to multiple fields, and for which each field may be associated with access operations of a respective subset of the addresses. In some examples, such registers may be configured or accessed based on indications (e.g., commands, requests) from a host device, which may support dynamic access operation monitoring that is responsive to various operating conditions. In some examples, the host device may perform evaluations based on such minimum values associated with respective addresses to determine a distribution of data across various portions of memory.

    Adaptive cache
    56.
    发明授权

    公开(公告)号:US11693775B2

    公开(公告)日:2023-07-04

    申请号:US17657922

    申请日:2022-04-04

    CPC classification number: G06F12/0815 G06F12/0804 G06F12/0864 G06F2212/1044

    Abstract: Described apparatuses and methods form adaptive cache lines having a configurable capacity from hardware cache lines having a fixed capacity. The adaptive cache lines can be formed in accordance with a programmable cache-line parameter. The programmable cache-line parameter can specify a capacity for the adaptive cache lines. The adaptive cache lines may be formed by combining respective groups of fixed-capacity hardware cache lines. The quantity of fixed-capacity hardware cache lines included in respective adaptive cache lines may be based on the programmable cache-line parameter. The programmable cache-line parameter can be selected in accordance with characteristics of the cache workload.

    Cache Memory Addressing
    57.
    发明申请

    公开(公告)号:US20230061668A1

    公开(公告)日:2023-03-02

    申请号:US17823480

    申请日:2022-08-30

    Abstract: Described apparatuses and methods order memory address portions advantageously for cache-memory addressing. An address bus can have a smaller width than a memory address. The multiple bits of the memory address can be separated into most-significant bits (MSB) and least-significant bits (LSB) portions. The LSB portion is provided to a cache first. The cache can process the LSB portion before the MSB portion is received. The cache can use index bits of the LSB portion to index into an array of memory cells and identify multiple corresponding tags. The cache can also check the corresponding tags against lower tag bits of the LSB portion. A partial match may be labeled as a predicted hit, and a partial miss may be labeled as an actual miss, which can initiate a data fetch. With the remaining tag bits from the MSB portion, the cache can confirm or refute the predicted hit.

    SELECTIVE CACHE LINE MEMORY ENCRYPTION

    公开(公告)号:US20230058668A1

    公开(公告)日:2023-02-23

    申请号:US17843571

    申请日:2022-06-17

    Abstract: A cache memory can maintain multiple cache lines and each cache line can include a data field, an encryption status attribute, and an encryption key attribute. The encryption status attribute can indicate whether the data field in the corresponding cache line includes encrypted or unencrypted data and the encryption key attribute can include an encryption key identifier for the corresponding cache line. In an example, a cryptographic controller can access keys from a key table to selectively encrypt or unencrypt cache data. Infrequently accessed cache data can be maintained as encrypted data, and more frequently accessed cache data can be maintained as unencrypted data. In some examples, different cache lines in the same cache memory can be maintained as encrypted or unencrypted data, and different cache lines can use respective different encryption keys.

    Adaptive Address Tracking
    59.
    发明申请

    公开(公告)号:US20230052043A1

    公开(公告)日:2023-02-16

    申请号:US17815516

    申请日:2022-07-27

    Abstract: Described apparatuses and methods track access metadata pertaining to activity within respective address ranges. The access metadata can be used to inform prefetch operations within the respective address ranges. The prefetch operations may involve deriving access patterns from access metadata covering the respective ranges. Suitable address range sizes for accurate pattern detection, however, can vary significantly from region to region of the address space based on, inter alia, workloads produced by programs utilizing the regions. Advantageously, the described apparatuses and methods can adapt the address ranges covered by the access metadata for improved prefetch performance. A data structure may be used to manage the address ranges in which access metadata are tracked. The address ranges can be adapted to improve prefetch performance through low-overhead operations implemented within the data structure. The data structure can encode hierarchical relationships that ensure the resulting address ranges are distinct.

    MEMORY INSTRUCTION FOR MEMORY TIERS

    公开(公告)号:US20220342568A1

    公开(公告)日:2022-10-27

    申请号:US17238791

    申请日:2021-04-23

    Abstract: Various embodiments provide for one or more processor instructions and memory instructions that enable a memory sub-system to copy, move, or swap data across (e.g., between) different memory tiers of the memory sub-system, where each of the memory tiers is associated with different memory locations (e.g., different physical memory locations) on one or more memory devices of the memory sub-system.

Patent Agency Ranking