-
公开(公告)号:US11556344B2
公开(公告)日:2023-01-17
申请号:US17035484
申请日:2020-09-28
Applicant: XILINX, INC.
Inventor: Millind Mittal , Jaideep Dastidar
IPC: G06F9/38 , G06F9/30 , G06F9/54 , G06F12/0891
Abstract: Embodiments herein describe transferring ownership of data (e.g., cachelines or blocks of data comprising multiple cachelines) from a host to hardware in an I/O device. In one embodiment, the host and I/O device (e.g., an accelerator) are part of a cache-coherent system where ownership of data can be transferred from a home agent (HA) in the host to a local HA in the I/O device—e.g., a computational slave agent (CSA). That way, a function on the I/O device (e.g., an accelerator function) can request data from the local HA without these requests having to be sent to the host HA. Further, the accelerator function can indicate whether the local HA tracks the data on a cacheline-basis or by a data block (e.g., multiple cachelines). This provides flexibility that can reduce overhead from tracking the data, depending on the function's desired use of the data.
-
公开(公告)号:US11477049B2
公开(公告)日:2022-10-18
申请号:US16053384
申请日:2018-08-02
Applicant: Xilinx, Inc.
Inventor: Millind Mittal , Kiran S. Puranik , Jaideep Dastidar
Abstract: A method and a system for transparently overlaying a logical transport network over an existing physical transport network is disclosed. The system designates a virtual channel located in a first transaction layer of a network conforming to a first network protocol. The system assembles a transaction layer packet in a second logical transaction layer of a second network protocol that is also recognizable by the first transaction layer. The system transfers the transaction layer packet from the second transaction layer to the virtual channel. The system transmits the transaction layer packet over the first transaction layer using the designated virtual channel over the network.
-
公开(公告)号:US10970217B1
公开(公告)日:2021-04-06
申请号:US16422647
申请日:2019-05-24
Applicant: XILINX, INC.
Inventor: Jaideep Dastidar , Millind Mittal
IPC: G06F12/06 , G06F12/08 , G06F12/02 , G06F9/50 , G06F12/0831 , G06F13/42 , G06F9/4401 , G06F13/40
Abstract: Embodiments disclosed herein provide a domain aware data migration scheme between processing elements, memory, and various caches in a CC-NUMA system. The scheme creates domain awareness in data migration operations, such as Direct Cache Transfer (DCT) operation, stashing operation, and in the allocation of policies of snoop filters and private, shared, or inline caches. The scheme defines a hardware-software interface to communicate locality information (also referred herein as affinity information or proximity information) and subsequent hardware behavior for optimal data migration, thus overcoming traditional CC-NUMA limitations.
-
公开(公告)号:US20200341941A1
公开(公告)日:2020-10-29
申请号:US16396540
申请日:2019-04-26
Applicant: Xilinx, Inc.
Inventor: Jaideep Dastidar , Millind Mittal
Abstract: Examples herein describe a peripheral I/O device with a hybrid gateway that permits the device to have both I/O and coherent domains. As a result, the compute resources in the coherent domain of the peripheral I/O device can communicate with the host in a similar manner as CPU-to-CPU communication in the host. The dual domains in the peripheral I/O device can be leveraged for machine learning (ML) applications. While an I/O device can be used as an ML accelerator, these accelerators previously only used an I/O domain. In the embodiments herein, compute resources can be split between the I/O domain and the coherent domain where a ML engine is in the I/O domain and a ML model is in the coherent domain. An advantage of doing so is that the ML model can be coherently updated using a reference ML model stored in the host.
-
公开(公告)号:US10698824B1
公开(公告)日:2020-06-30
申请号:US16141704
申请日:2018-09-25
Applicant: Xilinx, Inc.
Inventor: Millind Mittal , Jaideep Dastidar
IPC: G06F3/00 , G06F12/0815
Abstract: Disclosed systems and methods include in each agent, an agent layer, a link layer, and a port layer. The agent layer looks-up a port identifier in an address-to-port identifier map in response to a request directed to another agent and submits the request to the port layer. The link layer includes a plurality of links, and each link buffers communications from and to the agent layer. The port layer looks-up, in response to the request from the agent layer, a link identifier and chip identifier and writes the request to one of the links identified by the link identifier and associated with the chip identifier. The port layer also reads requests from the links and submits communications to a transport layer circuit based on the requests read from the links and associated chip identifiers.
-
公开(公告)号:US20200044895A1
公开(公告)日:2020-02-06
申请号:US16053384
申请日:2018-08-02
Applicant: Xilinx, Inc.
Inventor: Millind Mittal , Kiran S. Puranik , Jaideep Dastidar
Abstract: A method and a system for transparently overlaying a logical transport network over an existing physical transport network is disclosed. The system designates a virtual channel located in a first transaction layer of a network conforming to a first network protocol. The system assembles a transaction layer packet in a second logical transaction layer of a second network protocol that is also recognizable by the first transaction layer. The system transfers the transaction layer packet from the second transaction layer to the virtual channel. The system transmits the transaction layer packet over the first transaction layer using the designated virtual channel over the network.
-
公开(公告)号:US12045187B2
公开(公告)日:2024-07-23
申请号:US18206045
申请日:2023-06-05
Applicant: XILINX, INC.
Inventor: Jaideep Dastidar , Millind Mittal
CPC classification number: G06F13/4022 , G06F9/30043 , G06F13/1663 , G06F13/1668 , G06F2209/5011 , G06F2213/0038
Abstract: An integrated circuit (IC) for adaptive memory expansion scheme is proposed, which comprises: a home agent comprising a first memory expansion pool and a second memory expansion pool; a first port connecting the home agent to a first memory expansion device, where the first memory expansion device comprises a first memory pool; a second port connecting the home agent to a second memory expansion device, where the second memory expansion device comprises a second memory pool; a first address table mapping the first memory expansion pool to the first memory pool based on a size of the first memory expansion pool or a size of the first memory pool; and a second address table mapping the second memory expansion pool to the second memory pool based on a size of the second memory expansion pool or a size of the second memory pool.
-
公开(公告)号:US11983575B2
公开(公告)日:2024-05-14
申请号:US17903084
申请日:2022-09-06
Applicant: XILINX, INC.
Inventor: Millind Mittal , Jaideep Dastidar
IPC: G06F9/50 , G06F9/38 , G06F9/455 , G06F12/0815
CPC classification number: G06F9/5061 , G06F9/3877 , G06F9/45545 , G06F9/45558 , G06F9/5044 , G06F12/0815 , G06F2009/4557
Abstract: The embodiments herein describe a virtualization framework for cache coherent accelerators where the framework incorporates a layered approach for accelerators in their interactions between a cache coherent protocol layer and the functions performed by the accelerator. In one embodiment, the virtualization framework includes a first layer containing the different instances of accelerator functions (AFs), a second layer containing accelerator function engines (AFE) in each of the AFs, and a third layer containing accelerator function threads (AFTs) in each of the AFEs. Partitioning the hardware circuitry using multiple layers in the virtualization framework allows the accelerator to be quickly re-provisioned in response to requests made by guest operation systems or virtual machines executing in a host. Further, using the layers to partition the hardware permits the host to re-provision sub-portions of the accelerator while the remaining portions of the accelerator continue to operate as normal.
-
公开(公告)号:US11386031B2
公开(公告)日:2022-07-12
申请号:US16894446
申请日:2020-06-05
Applicant: XILINX, INC.
Inventor: Millind Mittal , Jaideep Dastidar
Abstract: Embodiments herein describe techniques for separating data transmitted between I/O functions in an integrated component and a host into separate data paths. In one embodiment, data packets are transmitted using a direct data path that bypasses a switch in the integrated component. In contrast, configuration packets (e.g., hot-swap, hot-add, hot-remove data, some types of descriptors, etc.) are transmitted to the switch which then forwards the configuration packets to their destination. The direct path for the data packets does not rely on switch connectivity (and its accompanying latency) to transport bandwidth sensitive traffic between the host and the I/O functions, and instead avoids (e.g., bypasses) the bandwidth, resource, store/forward, and latency properties of the switch. Meanwhile, the software compatibility attributes, such as hot plug attributes (which are not latency or bandwidth sensitive), continue to be supported by using the switch to provide a configuration data path.
-
公开(公告)号:US11372769B1
公开(公告)日:2022-06-28
申请号:US16555138
申请日:2019-08-29
Applicant: XILINX, INC.
Inventor: Millind Mittal , Jaideep Dastidar
IPC: G06F12/0891 , G06F12/0815 , G06F9/50 , G06F3/06
Abstract: The embodiments herein describe a multi-tenant cache that implements fine-grained allocation of the entries within the cache. Each entry in the cache can be allocated to a particular tenant—i.e., fine-grained allocation—rather than having to assign all the entries in a way to a particular tenant. If the tenant does not currently need those entries (which can be tracked using counters), the entries can be invalidated (i.e., deallocated) and assigned to another tenant. Thus, fine-grained allocation provides a flexible allocation of entries in a hardware cache that permits an administrator to reserve any number of entries for a particular tenant, but also permit other tenants to use this bandwidth when the reserved entries are not currently needed by the tenant.
-
-
-
-
-
-
-
-
-