Abstract:
System and method for policing logical network traffic flows using a ternary content addressable memory (TCAM). An exemplary embodiment can provide a network port that is associated with a plurality of entries of a TCAM, where each TCAM entry contains a value. Further, each TCAM entry can be assigned a priority and associated with at least one allow/drop action. A predefined set of values can be retrieved from at least one header field of a data packet processed by the network port. Each value in the predefined set of values can be aggregated into a search value, and the search value can be compared to the value contained in each TCAM entry. When a match is found between the search value and the value contained in a TCAM entry, the allow/drop action associated with the matching TCAM entry can be performed.
Abstract:
System and method providing resource access control within a single partition in a network switch environment in a high performance computing environment. A resource request portion of an IB packet includes data identifying a resource request for a requested resource, and a context identification portion of the packet includes data identifying a context of the resource request. Access rights to a controlled resource are determined based on the requested resource relative to the controlled resource and according to a comparison of the context of the resource request relative to one or more valid request context values. A resource table storing valid Q_Key and P_Key context values may be indexed by a Dest. QP of the packet header to determine selective access to the controlled resource and/or to data relating to the controlled resource.
Abstract:
Implementations of the present disclosure involve an apparatus, device, component, and/or method for a hardware efficient flow to port affinity management table for link aggregation for a network fabric switch with Ethernet Gateway functionality. Rather than maintaining a state per traffic flow list, the present disclosure utilizes a handle or hash value derived from the traffic flow and associates an output port state to the hash value. The output port state for the hash value is further associated with a portlist that is based on at least a traffic flow policy of the server or group of servers associated with the traffic flow. In addition, the management table may be adjusted based on state changes to one or more of the output ports such that, if a port becomes unavailable, the management table may be adjusted to account for the unavailability of the port.
Abstract:
Implementations of the present disclosure involve an apparatus, device, component, and/or method for a virtual output queue linked list management scheme for a high-performance network switch. In general, the linked list management scheme utilizes one or more look-ahead links associated with one or more descriptors in the linked list of descriptors that describe the storage of the incoming data packets to the switch. The look-ahead links allow the switch to schedule reads of memory locations included in the descriptors at the same speed at which the data packets are stored in memory.
Abstract:
A system and method can support packet switching in a network environment. A networking device, such as a network switch, which includes a crossbar fabric, can be associated with a plurality of input ports and a plurality of output ports. Furthermore, the networking device can detect a link state change at an output port that is associated with the networking device. Then, the networking device can notify one or more input ports, via the output port, of the link state change at the output port.
Abstract:
The disclosed embodiments provide a system that operates a switch fabric. During operation, the system creates a set of virtual links for a first set of physical ports in the switch fabric. Next, the system uses a set of virtual port numbers to create a set of mappings between the set of virtual links and a second set of physical ports in the switch fabric, wherein the second set of physical ports is connected to the first set of physical ports using a single physical link. The system then uses the mappings to provide non-blocking behavior in traffic from the first set of physical ports to the second set of physical ports.
Abstract:
The disclosed embodiments provide a system that operates a switch fabric. During operation, the system creates a set of virtual links for a first set of physical ports in the switch fabric. Next, the system uses a set of virtual port numbers to create a set of mappings between the set of virtual links and a second set of physical ports in the switch fabric, wherein the second set of physical ports is connected to the first set of physical ports using a single physical link. The system then uses the mappings to provide non-blocking behavior in traffic from the first set of physical ports to the second set of physical ports.
Abstract:
Implementations of the present disclosure involve an apparatus, device, component, and/or method for a hardware efficient flow to port affinity management table for link aggregation for a network fabric switch with Ethernet Gateway functionality. Rather than maintaining a state per traffic flow list, the present disclosure utilizes a handle or hash value derived from the traffic flow and associates an output port state to the hash value. The output port state for the hash value is further associated with a portlist that is based on at least a traffic flow policy of the server or group of servers associated with the traffic flow. In addition, the management table may be adjusted based on state changes to one or more of the output ports such that, if a port becomes unavailable, the management table may be adjusted to account for the unavailability of the port.
Abstract:
Implementations of the present disclosure involve an apparatus, device, component, and/or method for a virtual output queue linked list management scheme for a high-performance network switch. In general, the linked list management scheme utilizes one or more look-ahead links associated with one or more descriptors in the linked list of descriptors that describe the storage of the incoming data packets to the switch. The look-ahead links allow the switch to schedule reads of memory locations included in the descriptors at the same speed at which the data packets are stored in memory.
Abstract:
An input/output (I/O) switch fabric includes first physical ports that convey multiple network flows. Moreover, classifiers in the I/O switch fabric separate packets for network flows associated with different types of service. Then, the I/O switch fabric conveys the packets to different virtual switch ports without interference between the separated packets associated with different network flows. Furthermore, second physical ports in the I/O switch fabric output the packets, where a given second physical port outputs packets for at least some of the network flows associated with different types of service. In this way, the given second physical port can output packets having: the same source and destination; different sources and the same destination; or the same source and different destinations.