Apparatus and method to reduce bandwidth and latency overheads of probabilistic caches

    公开(公告)号:US12124371B2

    公开(公告)日:2024-10-22

    申请号:US17214356

    申请日:2021-03-26

    申请人: Intel Corporation

    摘要: An apparatus and method to reduce bandwidth and latency associated with probabilistic caches. For example, one embodiment of a processor comprises: a plurality of cores to execute instructions and process data, one or more of the cores to generate a request for a first cache line; a cache controller comprising cache lookup logic to determine a first way of a cache in which to search for the first cache line based on a first set of tag bits comprising one or more bits associated with the first cache line; the cache lookup logic to compare a second set of tag bits of the first cache line with a third set of tag bits of an existing cache line stored in the first way, wherein if the second set of tag bits and the third set of tag bits to not match, then the cache lookup logic to determine that the first cache line is not in the first way and to compare a fourth set of tag bits of the first cache line with a fifth set of tag bits of the existing cache line, wherein responsive to a match between the fourth set of tag bits and the fifth set of tag bits, the cache lookup logic to determine that the first cache line is stored in a second way and to responsively read the first cache line from the second way.

    Techniques for storing data and tags in different memory arrays

    公开(公告)号:US12079135B2

    公开(公告)日:2024-09-03

    申请号:US18209967

    申请日:2023-06-14

    申请人: Rambus Inc.

    发明人: Frederick A. Ware

    摘要: A memory controller includes logic circuitry to generate a first data address identifying a location in a first external memory array for storing first data, a first tag address identifying a location in a second external memory array for storing a first tag, a second data address identifying a location in the second external memory array for storing second data, and a second tag address identifying a location in the first external memory array for storing a second tag. The memory controller includes an interface that transfers the first data address and the first tag address for a first set of memory operations in the first and the second external memory arrays. The interface transfers the second data address and the second tag address for a second set of memory operations in the first and the second external memory arrays.

    Processor micro-operations cache architecture for intermediate instructions

    公开(公告)号:US12061907B2

    公开(公告)日:2024-08-13

    申请号:US17704127

    申请日:2022-03-25

    摘要: Various example embodiments for supporting processor capabilities are presented herein. Various example embodiments may be configured to support a micro-architecture for a micro-operations cache (UC) of a processor. Various example embodiments for supporting a micro-architecture for a UC of a processor may be configured to implement the UC of a processor using an intermediate vector UC (IV-UC). Various example embodiments for supporting an IV-UC for a processor may be configured to support a processor including an IV-UC where the IV-UC includes a micro-operations cache (UC) configured to store a cache line including sets of micro-operations (UOPs) from instructions decoded by the processor and an intermediate vector cache (IVC) configured to store indications of locations of the sets of UOPs in the cache line of the UC for intermediate instructions of the cache line of the UC.

    PARALLEL PROCESSING HAZARD MITIGATION AVOIDANCE

    公开(公告)号:US20240264974A1

    公开(公告)日:2024-08-08

    申请号:US18640044

    申请日:2024-04-19

    申请人: Ascenium, Inc.

    发明人: Peter Foley

    IPC分类号: G06F15/80 G06F12/0895

    CPC分类号: G06F15/8007 G06F12/0895

    摘要: Techniques for parallel processing based on hazard mitigation avoidance are disclosed. An array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operation hazard mitigation is enabled. The hazard mitigation is enabled by a control word tag. The control word tag supports memory access precedence information and is provided by the compiler at compile time. A hazardless memory access operation is executed. The hazardless memory access operation is determined by the compiler, and the hazardless memory access operation is designated by a unique set of precedence information contained in the tag. The tag is modified during runtime by hardware.

    MULTI-MODE INDEXED CACHE IN A PROCESSOR
    7.
    发明公开

    公开(公告)号:US20240241832A1

    公开(公告)日:2024-07-18

    申请号:US18097421

    申请日:2023-01-16

    IPC分类号: G06F12/0864 G06F12/0895

    CPC分类号: G06F12/0864 G06F12/0895

    摘要: Various example embodiments for supporting processor capabilities are presented herein. Various example embodiments for supporting processor capabilities may be configured to support a multi-mode indexed cache for a processor. Various example embodiments for supporting a multi-mode indexed cache for a processor may be configured to support a multi-mode indexed cache configured as a set associative cache having a plurality of sets, where the cache is configured to support multiple indexing modes for indexing memory blocks such that, for a memory operation for a memory block, the multiple indexing modes are configured to cause selection of different ones of the plurality of sets of the cache for the memory operation for the given memory block.