Abstract:
In one embodiment, longest prefix matching (LPM) operations are performed on a value in multiple interspersed prefix length search spaces to determine an overall longest prefix matching result in a packet switching system. A first LPM lookup unit performs a first LPM operation on the particular lookup value in a first search space finding a first longest matching prefix, and a second LPM lookup unit performs a second LPM operation on the particular lookup value in a second search space finding a second longest matching prefix. The longer of the first and second longest matching prefixes determines the overall LPM. In one embodiment, the first search space and the second search space include non-default route prefixes with interspersed prefix lengths matching a same value, such as, but not limited to the particular lookup value (e.g., a destination address of a packet).
Abstract:
Aspects of the disclosed technology address limitations relating to packet replication for multi-destination traffic, by providing methods for performing hardware-based replication in network infrastructure devices, such as switches. In some aspects, application specific integrated circuits (ASICs) resident in physical devices can be used to perform packet replication. Depending on implementation, a hardware-based replication process can include steps for receiving a first packet that includes a first outer header containing first address information, receiving a second packet including a second outer header containing a hardware replication flag, forwarding the first packet to all virtual tunnel endpoints (VTEPs) connected with the TOR switch, and performing hardware replication for the second packet based on the hardware replication flag to generate one or more unicast packets. Systems and machine readable media are also provided.
Abstract:
Systems, methods, and non-transitory computer-readable storage media for performing hierarchical routing are disclosed. The method includes identifying routes in a computer network and arranging those routes in two separate routing tables. The first routing table is stored on a first module and the second routing table is stored on a second module.
Abstract:
Various embodiments are disclosed for increasing Layer-3 LPM (longest prefix match) routing database in a network platform. In some embodiments, chipsets in fabric modules (FMs) can be partitioned into multiple banks. Network traffic can be directed towards a corresponding bank in the FMs by using a LPM table on a line card (LC). Entries in the LPM table on the LC can be programmed either statically or dynamically based upon LPM routes that are dynamically learned.
Abstract:
Embodiments provide a method, network device, and computer program product for optimizing traffic in a link-state network distributed across a plurality of sites. The method, network device, and computer program product include receiving a multi-destination message at a first node within the link-state network. Additionally, the method, network device, and computer program product include identifying a plurality of multi-destination trees within the network, each tree having a respective root node. The method, network device, and computer program product determine a plurality of cost values corresponding to each of the plurality of multi-destination trees, based on one or more links along a shortest path from the first node to the respective root node of each tree. Based on the determined cost values, the method, network device, and computer program product select one or more of the multi-destination trees and transmit the multi-destination message using the selected tree.
Abstract:
A plurality of equal cost paths through a network from a source node to a destination node are determined. A maximum bandwidth capacity for each link of each of the plurality of equal cost paths is determined, and a smallest capacity link for each of the plurality of equal cost paths is determined from the maximum capacity bandwidths for each link. An aggregated maximum bandwidth from the source node to the destination node is determined by aggregating the smallest capacity links for each of the plurality of equal cost paths. Traffic is sent from the source node along each of the plurality of equal cost paths according to a value of a capacity for the smallest capacity link for each of the plurality of equal cost paths, wherein a total of the sent traffic does not exceed the aggregated maximum bandwidth.
Abstract:
Systems, methods, and non-transitory computer-readable storage media for performing hierarchical routing are disclosed. The method includes identifying routes in a computer network and arranging those routes in two separate routing tables. The first routing table is stored on a first module and the second routing table is stored on a second module.
Abstract:
A plurality of line cards with each line card having a respective network forwarding engine and a respective outgoing interface (OIF) list and at least one fabric module communicatively coupled with each line card with each fabric module can have a respective network forwarding engine. The local OIF list can be asymmetrically programmed. The network forwarding engine of a line card can be configured to receive a multicast packet, compare a multicast address associate with the received multicast packet with entries in the local OIF list of the line card and forward the received multicast packet to at least one interface associated with the multicast address in response to the comparison resulting in a match.
Abstract:
Various embodiments are disclosed for increasing Layer-3 LPM (longest prefix match) routing database in a network platform. In some embodiments, chipsets in fabric modules (FMs) can be partitioned into multiple banks. Network traffic can be directed towards a corresponding bank in the FMs by using a LPM table on a line card (LC). Entries in the LPM table on the LC can be programmed either statically or dynamically based upon LPM routes that are dynamically learned.
Abstract:
Techniques are provided for generating and updating flooding tree paths in a network. At a particular node device in a network, a first flooding tree is generated by performing a first shortest path first (SPF) operation from a first selected node device in the network to a plurality of other node devices in the network. A second flooding tree is generated by performing a second SPF operation from a second selected node device in the network to the plurality of other node devices in the network. A network topology change event is detected in either the first or second flooding tree, and a packet sequence exchange is initiated between the particular node device and another node device in the network in response to the detected network topology change. The first and second flooding trees are then updated based on information obtained during the packet sequence exchange.