-
公开(公告)号:US20230393788A1
公开(公告)日:2023-12-07
申请号:US18454693
申请日:2023-08-23
Applicant: NVIDIA Corporation
Inventor: Nilandrish Chatterjee , James Michael O'Connor , Donghyuk Lee , Gaurav Uttreja , Wishwesh Anil Gandhi
CPC classification number: G06F3/0659 , G06F3/0604 , G06F12/0607 , G06F12/10 , H01L25/18 , G06F2212/657 , G06F2212/151 , G06F2212/154 , G06F3/0673
Abstract: A combined on-package and off-package memory system uses a custom base-layer within which are fabricated one or more dedicated interfaces to off-package memories. An on-package processor and on-package memories are also directly coupled to the custom base-layer. The custom base-layer includes memory management logic between the processor and memories (both off and on package) to steer requests. The memories are exposed as a combined memory space having greater bandwidth and capacity compared with either the off-package memories or the on-package memories alone. The memory management logic services requests while maintaining quality of service (QoS) to satisfy bandwidth requirements for each allocation. An allocation may include any combination of the on and/or off package memories. The memory management logic also manages data migration between the on and off package memories.
-
公开(公告)号:US20210373774A1
公开(公告)日:2021-12-02
申请号:US16888116
申请日:2020-05-29
Applicant: NVIDIA Corporation
Inventor: Ram Rangan , Patrick Richard Brown , Wishwesh Anil Gandhi , Steven James Heinrich , Mathias Heyer , Emmett Michael Kilgariff , Praveen Krishnamurthy , Dong Han Ryu
Abstract: Some systems compress data utilized by a user mode software without the user mode software being aware of any compression taking place. To maintain that illusion, such systems prevent user mode software from being aware of and/or accessing the underlying compressed states of the data. While such an approach protects proprietary compression techniques used in such systems from being deciphered, such restrictions limit the ability of user mode software to use the underlying compressed forms of the data in new ways. Disclosed herein are various techniques for allowing user-mode software to access the underlying compressed states of data either directly or indirectly. Such techniques can be used, for example, to allow various user-mode software on a single system or on multiple systems to exchange data in the underlying compression format of the system(s) even when the user mode software is unable to decipher the compression format.
-
公开(公告)号:US20210349639A1
公开(公告)日:2021-11-11
申请号:US16866813
申请日:2020-05-05
Applicant: NVIDIA Corporation
Inventor: Ram Rangan , Suryakant Patidar , Praveen Krishnamurthy , Wishwesh Anil Gandhi
Abstract: Accesses between a processor and its external memory is reduced when the processor internally maintains a compressed version of values stored in the external memory. The processor can then refer to the compressed version rather than access the external memory. One compression technique involves maintaining a dictionary on the processor mapping portions of a memory to values. When all of the values of a portion of memory are uniform (e.g., the same), the value is stored in the dictionary for that portion of memory. Thereafter, when the processor needs to access that portion of memory, the value is retrieved from the dictionary rather than from external memory. Techniques are disclosed herein to extend, for example, the capabilities of such dictionary-based compression so that the amount of accesses between the processor and its external memory are further reduced.
-
公开(公告)号:US10430915B2
公开(公告)日:2019-10-01
申请号:US15879377
申请日:2018-01-24
Applicant: NVIDIA Corporation
Inventor: Andrei Khodakovsky , Kirill A. Dmitriev , Rouslan L. Dimitrov , Tzyywei Hwang , Wishwesh Anil Gandhi , Lacky Vasant Shah
Abstract: One or more copy commands are scheduled for locating one or more pages of data in a local memory of a graphics processing unit (GPU) for more efficient access to the pages of data during rendering. A first processing unit that is coupled to a first GPU receives a notification that an access request count has reached a specified threshold. The first processing unit schedules a copy command to copy the first page of data to a first memory circuit of the first GPU from a second memory circuit of the second GPU. The copy command is included within a GPU command stream.
-
公开(公告)号:US20190206023A1
公开(公告)日:2019-07-04
申请号:US15857330
申请日:2017-12-28
Applicant: NVIDIA Corporation
Inventor: Rouslan L. Dimitrov , Kirill A. Dmitriev , Andrei Khodakovsky , Tzyywei Hwang , Wishwesh Anil Gandhi , Lacky Vasant Shah
IPC: G06T1/60 , G06T1/20 , G06F12/1009 , G06T15/00
CPC classification number: G06T1/60 , G06F9/44 , G06F12/1009 , G06T1/20 , G06T15/005
Abstract: A method for rendering graphics frames allocates rendering work to multiple graphics processing units (GPUs) that are configured to allow access to pages of data stored in locally attached memory of a peer GPU. The method includes the steps of generating, by a first GPU coupled to a first memory circuit, one or more first memory access requests to render a first primitive for a first frame, where at least one of the first memory access requests targets a first page of data that physically resides within a second memory circuit coupled to a second GPU. The first GPU requests the first page of data through a first data link coupling the first GPU to the second GPU and a register circuit within the first GPU accumulates an access request count for the first page of data. The first GPU notifies a driver that the access request count has reached a specified threshold.
-
-
-
-