摘要:
A hosting partition update mechanism allows updating I/O capability of a logically-partitioned computer system in a way that minimally affects the performance and availability of I/O in the computer system. When an update is needed, a new hosting partition is created with the desired update(s). I/O adapters in the current hosting partition are then migrated to the new hosting partition. The migration of an I/O adapter from the current hosting partition to the new hosting partition is relatively fast, thereby minimally impacting system performance and availability of I/O. Once all of the I/O adapters have been migrated to the new hosting partition, the current hosting partition may be kept as a backup, or may be eliminated. Providing a new or backup hosting partition allows updates to be performed in the new or backup hosting partition in a non-disruptive manner while the current hosting partition continues to service I/O requests.
摘要:
A hosting partition update mechanism allows updating I/O capability of a logically-partitioned computer system in a way that minimally affects the performance and availability of I/O in the computer system. When an update is needed, a new hosting partition is created with the desired update(s). I/O adapters in the current hosting partition are then migrated to the new hosting partition. The migration of an I/O adapter from the current hosting partition to the new hosting partition is relatively fast, thereby minimally impacting system performance and availability of I/O. Once all of the I/O adapters have been migrated to the new hosting partition, the current hosting partition may be kept as a backup, or may be eliminated. Providing a new or backup hosting partition allows updates to be performed in the new or backup hosting partition in a non-disruptive manner while the current hosting partition continues to service I/O requests.
摘要:
A method and apparatus are provided for mapping multiple address spaces into a single bus, such as a single peripheral component interconnect (PCI) bus. The single bus is coupled to a first processor complex and a second processor complex. An original address of a memory access is shifted to a unique address space for each originator/target of an operation. The shifted address is used on the single bus. Then the shifted address is shifted back to the original address for completing the operation on a destination bus. The original address of a memory access is shifted to a unique address space for each originator/target of an operation using a respective predefined value (+X1, +X2, or +X3) for shifting the original address above a predefined boundary for each originator/target of the operation. Shifting back the shifted address to the original address for completing the operation on the destination bus uses a respective predefined value (−X1, −X2, or −X3) for the shifted back address to the original address for completing the operation on the destination bus. Using the shifted address on the single bus utilizes a dual address cycle (DAC) of the single bus for the shifted address. Completing the operation on a destination bus utilizes a single address cycle (SAC) of the destination bus for the shifted back address to the original address.
摘要:
An input/output bus architecture that includes: an input/output bus; an input/output device connected to the input/output bus; a main processor, connected to the input/output bus, for executing a device driver corresponding to the input/output device, the device driver generating load/store commands for the input/output device; and a load/store assist engine, connected to the input/output bus and yet independent of the main processor, for loading/storing data to/from the input/output device according to the load/store commands from the device driver. The load/store assist engine decouples the main processor from latencies associated with execution of the load/store commands. The device driver is reassigned to the main processor, rather than being found in a device that is external to the main processor, such as an input/output processor.
摘要:
A method, apparatus and computer program product implement InfiniBand (IB) network topology simplification. A Subnet Manager (SM) of an IB subnet sends a subnet discovery request to each switch requesting the number of ports that are attached to the switch. Each of the switches and target channel adapters (TCAs) within the IB subnet includes a Subnet Management Agent (SMA). The Subnet Management Agent (SMA) of the receiving switch responds to the SM indicating a sufficient number of ports on the switch to support at least one port for each TCA. Each TCA supports at least two local IDs (LIDs).
摘要:
Methods and systems for discovering and managing devices connected to InfiniBand ports are provided. The discovery may be performed by an end node, such that the end node interoperates with all standard InfiniBand components. Specific actions that are vendor unique, and potentially not compliant with the InfiniBand architecture, may not be done until after the discovery is complete and it has been verified that the noncompliant action will only be directed to entities known to be capable of processing them. These actions may include assuming the configuration responsibilities that would have been performed by the Subnet Manager in a standard InfiniBand network.
摘要:
A method, apparatus, and computer program product are provided for implementing a receive function over an interconnect network, such as InfiniBand. A virtual lane (VL) with a pending packet for a queue pair (QP) is selected. Then the pending packet is checked for an exceptional condition. Responsive to identifying the exceptional condition for the pending packet, a state bit is set for freezing the selected VL; and an interrupt is generated to firmware. Responsive to receiving the interrupt, the firmware determines a cause for freezing the selected VL and performs a responsive action. For example, the responsive action performed by firmware includes firmware performing an error recovery procedure (ERP) for the QP; firmware updating a state for the QP; or firmware performing application unique processing for the QP.
摘要:
A method, apparatus, and computer program product are provided for implementing global to local queue pair translation in a network transport layer. A global queue pair number is identified. The global queue pair number is translated to a smaller local queue pair number. The local queue pair number is used for storing local queue pair context data for outbound header generation and inbound header checking. Upper layers of the network protocol above the network transport layer are allowed to use the global queue pair numbers.
摘要:
A method, apparatus and computer program product are provided for implementing a transmit queue. A queue pair context memory is provided for storing a set of pointers for each queue pair. The set of pointers are used to control the transmit queue for receiving, processing, and sending messages. Responsive to identifying an error for a queue pair, a limit pointer enable bit and a limit pointer to identify a last request for processing after the error are stored in the queue pair context memory for the queue pair.
摘要:
An I/O system including a processor complex and system main memory connected to I/O adapters via I/O adapters and I/O bus. A message transport mechanism and method stores an upstream message queue and a downstream message queue in system main memory. Queue addresses are stored both in system main memory and designated registers of I/O adapters. The I/O adapters utilize the queue addresses to manage the transfer of downstream command messages and to send upstream response messages to the system main memory via direct memory access across the I/O bus.