Abstract:
Disclosed are various embodiments for facilitating network flows in a networked environment. In various embodiments, a switch transmits data using an egress port that comprises an egress queue. The switch sets a congestion notification threshold for the egress queue. The switch generates a drain rate metric based at least in part on a drain rate for the egress queue, and the congestion notification threshold is adjusted based at least in part on the drain rate metric.
Abstract:
Disclosed are various embodiments that relate to a network switch. The network switch obtains a network state metric, the network state metric quantifying a network traffic congestion associated with a switch. The network switch identifies a synchronous time stamp associated with the network state metric and generates an network state reporting message, the network state reporting message comprising the network state metric and the synchronous time stamp. The network state reporting message may be transmitted to a monitoring system.
Abstract:
Network devices add annotation information to network packets as they travel through the network devices. The network devices may be switches, routers, bridges, hubs, or any other network device. The annotation information may be information specific to the network devices, as opposed to simply the kinds of information available at application servers that receive the network packets. As just a few examples, the annotation information may include switch buffer levels, routing delay, routing parameters affecting the packet, switch identifiers, power consumption, and heat, moisture, or other environmental data.
Abstract:
To reduce latency in a network device that buffer packets in different queues based on class of service, packets received from a network are stored in physical queues according to a class of service associated with the packets and a class of service associated with each of the physical queues. The physical queues are scheduled based quality of service requirements of their associated class of service. The physical queues are shadowed by virtual queues, and whether congestion exists in at least one of the virtual queues is determined. Packets departing from at least one of the physical queues are marked when congestion exists in at least one of the virtual queues. The service rate of the virtual queues is set to be less than or equal to a port link rate of the network device.
Abstract:
Aspects of a switch fabric end-to-end congestion avoidance mechanism are presented. Aspects of a system for end-to-end congestion avoidance in a switch fabric may include at least one circuit that enables reception of a congestion notification message that specifies a traffic flow identifier. The circuitry may enable increase or decrease of a current rate for transmission of data link layer (DLL) protocol data units (PDU) associated with the specified traffic flow identifier as a response to the reception of the congestion notification message.
Abstract:
In various embodiments, a system includes a switch comprising a resource that is shared between multiple objects. The switch comprises circuitry that determines a congestion metric for the switch in response to an amount of used of the resource by the objects. The circuitry determines a feedback parameter that is responsive to the congestion metric. The circuitry generates a congestion notification message that comprises a congestion feedback value responsive to the feedback parameter.In further embodiments, a system includes a switch that processes data for a first data link layer access network. The switch includes circuitry that identifies whether a received packet originated from a source device that shares the same network layer access layer as the switch. If the source device shares the same network layer access network as the switch, the circuitry generates a congestion notification message comprising a congestion feedback value for the switch.
Abstract:
Processing techniques in a network switch help reduce latency in the delivery of data packets to a recipient. The processing techniques include speculative flow status messaging, for example. The speculative flow status messaging may alert an egress tile or output port of an incoming packet before the incoming packet is fully received. The processing techniques may also include implementing a separate accelerated credit pool which provides controlled push capability for the ingress tile or input port to send packets to the egress tile or output port without waiting for a bandwidth credit from the egress tile or output port.
Abstract:
Disclosed are various embodiments for providing a data packet with timestamp information. A data packet is generated such that it comprises a payload and a header. The payload comprises a first timestamp field that comprises data indicating when a network device processed the data packet. The payload also comprises a body data field and a body data protocol field. The body data protocol field comprises data identifying a protocol used by body data in the body data field. The header comprises a payload protocol field that comprises data identifying that the payload comprises timestamp data.
Abstract:
Disclosed are various embodiments that relate to a network switch. The switch determines whether a network packet is associated with a packet processing context, the packet processing context specifying a condition of handling network packets processed in the switch. The switch determines debug metadata for the network packet in response to the network packet being associated with the packet processing context; and the debug metadata is stored in a capture buffer.
Abstract:
A system and method for adjusting an energy efficient Ethernet (EEE) control policy using measured power savings. An EEE-enabled device can be designed to report EEE event data. This reported EEE event data can be used to quantify the actual EEE benefits of the EEE-enabled device, debug the EEE-enabled device, and adjust the EEE control policy.