-
公开(公告)号:US20250036571A1
公开(公告)日:2025-01-30
申请号:US18715719
申请日:2021-12-03
Inventor: John D. DAVIS , Steven WALLACH , Gerald ZURASKI
IPC: G06F12/0895 , G06F12/0846 , G06F13/16
Abstract: The present invention relates to a method for storing and accessing an operand in a memory unit of a computer, and specifically, to a method for accessing a data operand stored in a main memory unit with N independently controlled channels a memory cache and a tag store, wherein the method comprises the steps of comparing a physical address with the tag store, the cache tags comprising N status bits, referencing the channels, and a method for writing a data operand in a main memory unit with N independently controlled channels.
-
公开(公告)号:US20250028675A1
公开(公告)日:2025-01-23
申请号:US18791963
申请日:2024-08-01
Applicant: Intel Corporation
Inventor: JOYDEEP RAY , SELVAKUMAR PANNEER , SAURABH TANGRI , BEN ASHBAUGH , SCOTT JANUS , ABHISHEK APPU , VARGHESE GEORGE , RAVISHANKAR IYER , NILESH JAIN , PATTABHIRAMAN K , ALTUG KOKER , MIKE MACPHERSON , JOSH MASTRONARDE , ELMOUSTAPHA OULD-AHMED-VALL , JAYAKRISHNA P. S , ERIC SAMSON
IPC: G06F15/78 , G06F7/544 , G06F7/575 , G06F7/58 , G06F9/30 , G06F9/38 , G06F9/50 , G06F12/02 , G06F12/06 , G06F12/0802 , G06F12/0804 , G06F12/0811 , G06F12/0862 , G06F12/0866 , G06F12/0871 , G06F12/0875 , G06F12/0882 , G06F12/0888 , G06F12/0891 , G06F12/0893 , G06F12/0895 , G06F12/0897 , G06F12/1009 , G06F12/128 , G06F15/80 , G06F17/16 , G06F17/18 , G06N3/08 , G06T1/20 , G06T1/60 , G06T15/06 , H03M7/46
Abstract: Embodiments described herein include software, firmware, and hardware that provides techniques to enable deterministic scheduling across multiple general-purpose graphics processing units. One embodiment provides a multi-GPU architecture with uniform latency. One embodiment provides techniques to distribute memory output based on memory chip thermals. One embodiment provides techniques to enable thermally aware workload scheduling. One embodiment provides techniques to enable end to end contracts for workload scheduling on multiple GPUs.
-
公开(公告)号:US20250028646A1
公开(公告)日:2025-01-23
申请号:US18524412
申请日:2023-11-30
Applicant: Samsung Electronics Co., Ltd.
Inventor: Qirui Yang , Bridget Davis , Devasena Inupakutika , Peter Chen , Bradley Carpenter
IPC: G06F12/0895 , G06F12/123
Abstract: A method for data storage includes receiving, by a storage device, a read request for data, the read request being associated with a request size, determining that a first cache area associated with a first portion of the data is in a first portion of a cache, the first cache area having a first size that is smaller than the request size, determining that a second cache area associated with a second portion of the data is in a second portion of the cache, the second cache area having a second size that is smaller than the request size and differently sized than the first size, and based on the read request, reading the first portion of the data from the first cache area and reading the second portion of the data from the second cache area.
-
公开(公告)号:US12197331B2
公开(公告)日:2025-01-14
申请号:US18487196
申请日:2023-10-16
Applicant: TEXAS INSTRUMENTS INCORPORATED
Inventor: Abhijeet Ashok Chachad , David Matthew Thompson , Naveen Bhoria , Pete Michael Hippleheuser
IPC: G06F12/0811 , G06F9/30 , G06F9/38 , G06F9/46 , G06F9/54 , G06F11/30 , G06F12/0808 , G06F12/0815 , G06F12/0817 , G06F12/0831 , G06F12/084 , G06F12/0895 , G06F12/128 , G06F13/16
Abstract: An apparatus includes a CPU core and a L1 cache subsystem including a L1 main cache, a L1 victim cache, and a L1 controller. The apparatus includes a L2 cache subsystem including a L2 main cache, a shadow L1 main cache, a shadow L1 victim cache, and a L2 controller configured to receive a read request from the L1 controller as a single transaction. Read request includes a read address, a first indication of an address and a coherence state of a cache line A to be moved from the L1 main cache to the L1 victim cache to allocate space for data returned in response to the read request, and a second indication of an address and a coherence state of a cache line B to be removed from the L1 victim cache in response to the cache line A being moved to the L1 victim cache.
-
公开(公告)号:US20250013284A1
公开(公告)日:2025-01-09
申请号:US18894180
申请日:2024-09-24
Applicant: TEXAS INSTRUMENTS INCORPORATED
Inventor: Oluleye Olorode , Mehrdad Nourani
IPC: G06F1/3234 , G06F12/0811 , G06F12/0846 , G06F12/0895
Abstract: Disclosed embodiments relate to a dNap architecture that accurately transitions cache lines to full power state before an access to them. This ensures that there are no additional delays due to waking up drowsy lines. Only cache lines that are determined by the DMC to be accessed in the immediate future are fully powered while others are put in drowsy mode. As a result, we are able to significantly reduce leakage power with no cache performance degradation and minimal hardware overhead, especially at higher associativities. Up to 92% static/leakage power savings are accomplished with minimal hardware overhead and no performance tradeoff.
-
公开(公告)号:US20250004981A1
公开(公告)日:2025-01-02
申请号:US18793247
申请日:2024-08-02
Applicant: Intel Corporation
Inventor: Abhishek R. Appu , Altug Koker , Aravindh Anantaraman , Elmoustapha Ould-Ahmed-Vall , Valentin Andrei , Nicolas Galoppo Von Borries , Varghese George , Mike Macpherson , Subramaniam Maiyuran , Joydeep Ray , Lakshminarayana Striramassarma , Scott Janus , Brent Insko , Vasanth Ranganathan , Kamal Sinha , Arthur Hunter , Prasoonkumar Surti , David Puffer , James Valerio , Ankur N. Shah
IPC: G06F15/78 , G06F7/544 , G06F7/575 , G06F7/58 , G06F9/30 , G06F9/38 , G06F9/50 , G06F12/02 , G06F12/06 , G06F12/0802 , G06F12/0804 , G06F12/0811 , G06F12/0862 , G06F12/0866 , G06F12/0871 , G06F12/0875 , G06F12/0882 , G06F12/0891 , G06F12/0893 , G06F12/0895 , G06F12/0897 , G06F12/1009 , G06F12/128 , G06F15/80 , G06F17/16 , G06F17/18 , G06N3/08 , G06T1/20 , G06T1/60 , G06T15/06 , H03M7/46
Abstract: Methods and apparatus relating to techniques for multi-tile memory management. In an example, a graphics processor includes an interposer, a first chiplet coupled with the interposer, the first chiplet including a graphics processing resource and an interconnect network coupled with the graphics processing resource, cache circuitry coupled with the graphics processing resource via the interconnect network, and a second chiplet coupled with the first chiplet via the interposer, the second chiplet including a memory-side cache and a memory controller coupled with the memory-side cache. The memory controller is configured to enable access to a high-bandwidth memory (HBM) device, the memory-side cache is configured to cache data associated with a memory access performed via the memory controller, and the cache circuitry is logically positioned between the graphics processing resource and a chiplet interface.
-
公开(公告)号:US12182062B1
公开(公告)日:2024-12-31
申请号:US17961833
申请日:2022-10-07
Applicant: Intel Corporation
Inventor: Abhishek R. Appu , Altug Koker , Aravindh Anantaraman , Elmoustapha Ould-Ahmed-Vall , Valentin Andrei , Nicolas Galoppo Von Borries , Varghese George , Mike Macpherson , Subramaniam Maiyuran , Joydeep Ray , Lakshminarayanan Striramassarma , Scott Janus , Brent Insko , Vasanth Ranganathan , Kamal Sinha , Arthur Hunter , Prasoonkumar Surti , David Puffer , James Valerio , Ankur N. Shah
IPC: G06F15/78 , G06F7/544 , G06F7/575 , G06F7/58 , G06F9/30 , G06F9/38 , G06F9/50 , G06F12/02 , G06F12/06 , G06F12/0802 , G06F12/0804 , G06F12/0811 , G06F12/0862 , G06F12/0866 , G06F12/0871 , G06F12/0875 , G06F12/0882 , G06F12/0891 , G06F12/0893 , G06F12/0895 , G06F12/0897 , G06F12/1009 , G06F12/128 , G06F15/80 , G06F17/16 , G06F17/18 , G06T1/20 , G06T1/60 , H03M7/46 , G06N3/08 , G06T15/06
Abstract: Methods and apparatus relating to techniques for multi-tile memory management. In an example, a graphics processor includes an interposer, a first chiplet coupled with the interposer, the first chiplet including a graphics processing resource and an interconnect network coupled with the graphics processing resource, cache circuitry coupled with the graphics processing resource via the interconnect network, and a second chiplet coupled with the first chiplet via the interposer, the second chiplet including a memory-side cache and a memory controller coupled with the memory-side cache. The memory controller is configured to enable access to a high-bandwidth memory (HBM) device, the memory-side cache is configured to cache data associated with a memory access performed via the memory controller, and the cache circuitry is logically positioned between the graphics processing resource and a chiplet interface.
-
公开(公告)号:US20240411698A1
公开(公告)日:2024-12-12
申请号:US18329677
申请日:2023-06-06
Applicant: Truist Bank
Inventor: Jonathan Randall Howarth , Szu Yu Zoe Chao , Brian Stewart Brown , Thomas Prescott Atherton , Earlie Clark, IV
IPC: G06F12/0895
Abstract: A tag accumulator tool includes, in an expanded mode, a proffered dataset of items for user selection, and a displayed accumulation area including a tag cache having, for each one of the selected data items, a respective corresponding tag. In a reduced mode, the tag accumulator tool displays an enumeration of the selected data items. A query box is operable to, upon entry of a character string, display a redacted list of selectable items including only those items of the proffered dataset at least partially matching the character string. Each tag in the accumulation area may include an annul tool the activation of which removes the tag from the accumulation area, effectively returning the corresponding data item to an unselected status.
-
公开(公告)号:US20240354254A1
公开(公告)日:2024-10-24
申请号:US18759068
申请日:2024-06-28
Applicant: Lodestar Licensing Group LLC
Inventor: Richard C. Murphy
IPC: G06F12/0864 , G06F9/30 , G06F12/0811 , G06F12/084 , G06F12/0895 , G11C7/10 , G11C11/4091 , G11C11/4096 , G11C11/4094 , G11C19/00
CPC classification number: G06F12/0864 , G06F9/30036 , G06F12/0811 , G06F12/084 , G06F12/0895 , G11C7/1006 , G11C11/4091 , G11C11/4096 , G06F2212/1012 , G06F2212/1044 , G06F2212/283 , G06F2212/6032 , G11C11/4094 , G11C19/00
Abstract: The present disclosure includes apparatuses and methods for compute enabled cache. An example apparatus comprises a compute component, a memory and a controller coupled to the memory. The controller configured to operate on a block select and a subrow select as metadata to a cache line to control placement of the cache line in the memory to allow for a compute enabled cache.
-
10.
公开(公告)号:US20240345990A1
公开(公告)日:2024-10-17
申请号:US18626775
申请日:2024-04-04
Applicant: Intel Corporation
Inventor: Lakshminarayanan Striramassarma , Prasoonkumar Surti , Varghese George , Ben Ashbaugh , Aravindh Anantaraman , Valentin Andrei , Abhishek Appu , Nicolas Galoppo Von Borries , Altug Koker , Mike Macpherson , Subramaniam Maiyuran , Nilay Mistry , Elmoustapha Ould-Ahmed-Vall , Selvakumar Panneer , Vasanth Ranganathan , Joydeep Ray , Ankur Shah , Saurabh Tangri
IPC: G06F15/78 , G06F7/544 , G06F7/575 , G06F7/58 , G06F9/30 , G06F9/38 , G06F9/50 , G06F12/02 , G06F12/06 , G06F12/0802 , G06F12/0804 , G06F12/0811 , G06F12/0862 , G06F12/0866 , G06F12/0871 , G06F12/0875 , G06F12/0882 , G06F12/0888 , G06F12/0891 , G06F12/0893 , G06F12/0895 , G06F12/0897 , G06F12/1009 , G06F12/128 , G06F15/80 , G06F17/16 , G06F17/18 , G06N3/08 , G06T1/20 , G06T1/60 , G06T15/06 , H03M7/46
CPC classification number: G06F15/7839 , G06F7/5443 , G06F7/575 , G06F7/588 , G06F9/3001 , G06F9/30014 , G06F9/30036 , G06F9/3004 , G06F9/30043 , G06F9/30047 , G06F9/30065 , G06F9/30079 , G06F9/3887 , G06F9/5011 , G06F9/5077 , G06F12/0215 , G06F12/0238 , G06F12/0246 , G06F12/0607 , G06F12/0802 , G06F12/0804 , G06F12/0811 , G06F12/0862 , G06F12/0866 , G06F12/0871 , G06F12/0875 , G06F12/0882 , G06F12/0888 , G06F12/0891 , G06F12/0893 , G06F12/0895 , G06F12/0897 , G06F12/1009 , G06F12/128 , G06F15/8046 , G06F17/16 , G06F17/18 , G06T1/20 , G06T1/60 , H03M7/46 , G06F9/3802 , G06F9/3818 , G06F9/3867 , G06F2212/1008 , G06F2212/1021 , G06F2212/1044 , G06F2212/302 , G06F2212/401 , G06F2212/455 , G06F2212/60 , G06N3/08 , G06T15/06
Abstract: Multi-tile Memory Management for Detecting Cross Tile Access, Providing Multi-Tile Inference Scaling with multicasting of data via copy operation, and Providing Page Migration are disclosed herein. In one embodiment, a graphics processor for a multi-tile architecture includes a first graphics processing unit (GPU) having a memory and a memory controller, a second graphics processing unit (GPU) having a memory and a cross-GPU fabric to communicatively couple the first and second GPUs. The memory controller is configured to determine whether frequent cross tile memory accesses occur from the first GPU to the memory of the second GPU in the multi-GPU configuration and to send a message to initiate a data transfer mechanism when frequent cross tile memory accesses occur from the first GPU to the memory of the second GPU.
-
-
-
-
-
-
-
-
-