摘要:
A method, computer program product, and a data processing system for posting and retrieving WQEs to a shared receive queue in a manner that alleviates head-of-line blocking issues is provided. The present invention provides a shared receive queue and a posting routine for allowing non-sequential work request postings in the shared receive queue. Additionally, a fetch routine is provided for non-sequential fetching of work queue elements from the shared receive queue.
摘要:
A method for receiving data in a network acceleration architecture for use with TCP (transport control protocol), iSCSI (Internet Small Computer System Interface) and/or RDMA (Remote Direct Memory Access) over TCP, including providing a hardware acceleration engine, called a streamer, adapted for communication with and processing data from a consumer application in a system that supports TCP, iSCSI and RDMA over TCP, providing a software protocol processor adapted for carrying out TCP implementation, the software control processor being called a TCE (TCP Control Engine), wherein the streamer and the TCE are adapted to operate asynchronously and independently of one another, and transmitting a TCP segment with the streamer.
摘要:
A network acceleration architecture for use with TCP, iSCSI and/or RDMA over TCP, including a hardware acceleration engine adapted for communication with and processing data from a consumer application in a system that supports TCP, iSCSI and RDMA over TCP, a software protocol processor adapted for carrying out TCP implementation, and an asynchronous dual-queue interface for exchanging information between the hardware acceleration engine and the software protocol processor, wherein the hardware acceleration engine and the software protocol processor are adapted to operate asynchronously and independently of one another.
摘要翻译:一种用于TCP,iSCSI和/或RDMA over TCP的网络加速架构,包括适用于与支持TCP,iSCSI和RDMA over TCP的系统中的消费者应用程序通信和处理数据的硬件加速引擎,软件协议处理器 适用于执行TCP实现,以及用于在硬件加速引擎和软件协议处理器之间交换信息的异步双队列接口,其中硬件加速引擎和软件协议处理器适于彼此异步地且独立地操作。
摘要:
A method for receiving data in a network acceleration architecture for use with TCP (transport control protocol), iSCSI (Internet Small Computer System Interface) and RDMA (Remote Direct Memory Access) over TCP, including providing a hardware acceleration engine, called a streamer, adapted for communication with and processing data from a consumer application in a system that supports TCP, iSCSI and RDMA over TCP, providing a software protocol processor adapted for carrying out TCP implementation, the software control processor being called a TCE (TCP Control Engine), wherein the streamer and the TCE are adapted to operate asynchronously and independently of one another, and receiving an inbound TCP segment with the streamer.
摘要翻译:一种用于在TCP(传输控制协议),iSCSI(因特网小型计算机系统接口)和RDMA(远程直接存储器访问)TCP上使用的网络加速架构中接收数据的方法,包括提供称为流传输器的硬件加速引擎, 适用于与支持TCP,iSCSI和RDMA over TCP的系统中的消费者应用程序进行通信和处理数据,提供适于执行TCP实现的软件协议处理器,称为TCE(TCP控制引擎)的软件控制处理器, 其中所述流送器和所述TCE适于彼此异步地且独立地操作,并且与所述流送器接收入站TCP段。
摘要:
A method for receiving data in a network acceleration architecture for use with TCP (transport control protocol), iSCSI (Internet Small Computer System Interface) and RDMA (Remote Direct Memory Access) over TCP, including providing a hardware acceleration engine, called a streamer, adapted for communication with and processing data from a consumer application in a system that supports TCP, iSCSI and RDMA over TCP, providing a software protocol processor adapted for carrying out TCP implementation, the software control processor being called a TCE (TCP Control Engine), and providing an asynchronous dual-queue interface for exchanging information between the streamer and TCE, wherein the streamer and TCE are adapted to operate asynchronously and independently of one another.
摘要翻译:一种用于在TCP(传输控制协议),iSCSI(因特网小型计算机系统接口)和RDMA(远程直接存储器访问)TCP上使用的网络加速架构中接收数据的方法,包括提供称为流传输器的硬件加速引擎, 适用于与支持TCP,iSCSI和RDMA over TCP的系统中的消费者应用程序进行通信和处理数据,提供适于执行TCP实现的软件协议处理器,称为TCE(TCP控制引擎)的软件控制处理器, 以及提供用于在流传送器和TCE之间交换信息的异步双队列接口,其中,所述流送器和TCE适于彼此异步地且彼此独立地操作。
摘要:
A method, system and computer program product that allows a System Image within a multiple System Image Virtual Server to maintain isolation from the other system images while directly exposing a portion, or all, of its associated System Memory to a shared PCI Adapter without the need for each I/O operation to be analyzed and verified by a component trusted by the LPAR manager.
摘要:
A method, computer program product, and distributed data processing system that allows a single physical I/O adapter, such as a PCI, PCI-X, or PCI-E adapter, to validate that a direct memory access address referenced by an incoming I/O transaction that was initiated through a memory mapped I/O operation are associated with a virtual adapter or virtual resource that is referenced by the incoming memory mapped I/O operation is provided. Specifically, the present invention is directed to a mechanism for sharing conventional PCI (Peripheral Component Interconnect) I/O adapters, PCI-X I/O Adapters, PCI-Express I/O Adapters, and, in general, any I/O adapter that uses a memory mapped I/O interface for communications. A single physical I/O adapter validates that one or more direct memory access addresses referenced by an incoming I/O transaction initiated through a memory mapped I/O operation are associated with a virtual adapter or virtual resource that is referenced by the incoming memory mapped I/O operation.
摘要:
A method, computer program product, and distributed data processing system that allows a single physical I/O adapter to validate that a memory mapped I/O address referenced by an incoming I/O operation is associated with a virtual host that initiated the incoming memory mapped I/O operation is provided. Specifically, the present invention is directed to a mechanism for sharing a PCI family I/O adapter and, in general, any I/O adapter that uses a memory mapped I/O interface for communications. A mechanism is provided that allows a single physical I/O adapter to validate that a memory mapped I/O address referenced by an incoming memory mapped I/O operation used to initiate an I/O transaction is associated with a virtual host that initiated the incoming memory mapped I/O operation.
摘要:
A method, computer program product, and distributed data processing system that enables host software or firmware to allocate virtual resources to one or more system images from a single physical I/O adapter, such as a PCI, PCI-X, or PCI-E adapter, is provided. Adapter resource groups are assigned to respective system images. An adapter resource group is exclusively available to the system image to which the adapter resource group assignment was made. Assignment of adapter resource groups may be made per a relative resource assignment or an absolute resource assignment. In another embodiment, adapter resource groups are assigned to system images on a first come, first served basis.
摘要:
A mechanism for handling event notifications or interrupts in a logically partitioned computing system having IO adapters that support adapter virtualization are provided. A virtual adapter associated with a physical IO adapter detects an event, identifies a logical partition associated with the event, and writes an event notification entry in a partition interrupt control block (PICB) of the logical partition memory space. The virtual adapter notifies shared control point logic on the physical IO adapter of the update to the PICB which may then write an entry to a logical partition manager interrupt control block (LPAR manager ICB) identifying the update to the PICB and the operating system associated with the PICB. The LPAR manager may then inform the operating system of the logical partition of the updates to the PICB which may then read the entries in the PICB, process them and inform appropriate application instances of the events.