Abstract:
A system and method is provided for sending congestion notification messages through L3 networks. For example, a data packet is received at a first switch in a first fabric block of an L3 network, and the first switch performs source MAC tagging of the data packet. The data packet is then forwarded to a second switch in a second fabric block of the L3 network, and the source MAC tag is maintained by the second switch and any intermediate switches. The second switch determines, in response to receiving the data packet, whether it is congested, and generates a notification message if it is congested. The notification message is L2 forwarded to the first fabric block, and further forwarded from the first switch to a source of the data packet using ACL matching.
Abstract:
Aspects of the present disclosure relate to detecting and repairing permanently pauses on a flow controlled fabric. In one aspect, one or more computing devices, such as a switch or a centralized controller, may detect whether a port of a network device receives one or more pause messages. The pause messages may instruct the network device to pause data transmission. Further, the one or more computing devices may determine a period of time during which the port receives the one or more pause messages and identify the port as a permanently paused port based on the determined period of time. The one or more computing devices may then reconfigure the permanently paused port to stop complying with the one or more pause messages.
Abstract:
Systems and methods for increasing bandwidth in a computer network are provided. A computer network can include a first lower level switch having a first port and a second port. The computer network can include a second lower level switch having a first port and a second port. The computer network can include an upper level switch having respective ports directly coupled to ports of the first and second lower level switches. A third port of the upper level switch can couple to a first port of a passive optical splitter. The passive optical splitter can have second and third ports coupled to respective ports of the first and second lower level switches. The passive optical splitter can be configured to transmit signals received at its first port as output signals on both of its second and third ports.
Abstract:
The present disclosure provides for the determination of bandwidth allocation of inter-block traffic in a data center network. It employs a number of optimization objectives and a heuristic water-filling strategy to avoid producing unnecessary paths and to avoid determining paths that would be unavailable when actually needed. Allocation may be adjusted incrementally upon node and link failure, for instance to perform only the minimal allocation changes necessary. If demand between a source and a destination cannot be satisfied, a decomposition process may be used to allocate remaining demand. One aspect constructs a graph for route computation based on inter-block topology. Here, the graph initially starts with a highest level of abstraction with each node representing a middle block, and gradually reduces the abstraction level to identify paths of mixed abstraction level to satisfy additional demand.
Abstract:
A system and method is provided for sending congestion notification messages through L3 networks and implementation of QCN in L3 switches. For example, according to this system and method, an L3 switch receives one or more data packets, and determines, based on the received one or more data packets, whether the L3 switch is congested. If the L3 switch is congested, it generates a congestion notification message, the congestion notification message including an Internet Protocol (IP) header, the IP header identifying a source of the one or more received data packets as its destination. The L3 switch sends the congestion notification message to the source of the one or more received data packets using information in the IP header.
Abstract:
A system and method is provided for sending congestion notification messages through L3 networks. For example, a data packet is received at a first switch in a first fabric block of an L3 network, and the first switch performs source MAC tagging of the data packet. The data packet is then forwarded to a second switch in a second fabric block of the L3 network, and the source MAC tag is maintained by the second switch and any intermediate switches. The second switch determines, in response to receiving the data packet, whether it is congested, and generates a notification message if it is congested. The notification message is L2 forwarded to the first fabric block, and further forwarded from the first switch to a source of the data packet using ACL matching.
Abstract:
A method includes, receiving a plurality of data flows. A respective data flow includes a respective source address and a respective destination address. The method further includes generating, without regard to priorities associated with the plurality of data flows, an ordering of the plurality of data flows; and iteratively modifying, without regard to the priorities, the ordering of the plurality of data flows by applying a randomization algorithm to the plurality of data flows, until a cost associated with path assignments for the ordering of the plurality of data flows satisfies a predetermined condition. A respective path assignment for a respective data flow specifies a respective path from a respective source address to a respective destination address. The method also includes executing the data flows based on the path assignments for the ordering of the plurality of data flows having the cost that satisfies the predetermined condition.
Abstract:
A technique for load balancing in a multi-topology network selects a network path having a favorable bandwidth*hop count product. Accordingly, shorter paths are prioritized while longer paths are used for traffic load balancing when appropriate. For example, a telecommunications network may employ a Clos-style topology in which hosts are connected to each other through multiple hierarchical levels of forwarding devices. One or more high level switches may be removed, creating isolated blocks of forwarding devices. The blocks may be connected by the remaining original high level Clos-style topology switch(es) and a second type of network topology. In an exemplary embodiment, traffic may be scheduled over routes with the following decreasing priorities: (1) the shortest path, using either or both network topologies; (2) a path using only a single network topology type; (3) non-shortest paths between directly-connected blocks; and (4) non-shortest paths between indirectly-connected blocks.
Abstract:
The present disclosure provides for the determination of bandwidth allocation of inter-block traffic in a data center network. It employs a number of optimization objectives and a heuristic water-filling strategy to avoid producing unnecessary paths and to avoid determining paths that would be unavailable when actually needed. Allocation may be adjusted incrementally upon node and link failure, for instance to perform only the minimal allocation changes necessary. If demand between a source and a destination cannot be satisfied, a decomposition process may be used to allocate remaining demand. One aspect constructs a graph for route computation based on inter-block topology. Here, the graph initially starts with a highest level of abstraction with each node representing a middle block, and gradually reduces the abstraction level to identify paths of mixed abstraction level to satisfy additional demand.
Abstract:
Exemplary embodiments determine respective capacities of network links in a multi-stage network. Specifically, the capacities of the upstream links for a given network element are determined based on the capacities of the downstream links for that network element. According to the various embodiments, a network element determines the amount of data, i.e. traffic, which may be assigned to downstream links based on the determined capacities. A network element may not receive more traffic than the total direct downstream capacity of the network element. Thus, optimum throughput may be attained for the entire network.