摘要:
In one example, a system includes a first computing device configured to execute a virtual machine, wherein the virtual machine is communicatively coupled to a virtual private network (VPN) via a first attachment circuit using a first set of network parameters, stop execution of the virtual machine, and create checkpoint data for the virtual machine, and a second computing device configured to execute the virtual machine, using at least some of the checkpoint data, and to cause the virtual machine to become communicatively coupled to the VPN via a second attachment circuit using a second set of network parameters different from the first set of network parameters. The system may further include a first provider edge (PE) routing device communicatively coupled to the first computing device via the first attachment circuit, and a second PE routing device communicatively coupled to the second computing device via the second attachment circuit.
摘要:
In general, techniques are described for providing feedback loops for service engineered paths. A service node comprising an interface and a control unit may implement the techniques. The interface receives traffic via a path configured within a network to direct the traffic from an ingress network device of the path to the service node. The control unit applies one or more services to the traffic received via the path and generates service-specific information related to the application of the one or more services to the traffic. The interface then sends the service-specific information to at least one network device configured to forward the traffic via the path so that the at least one network device configured to forward the traffic via the path is able to adapt the path based on the service-specific information.
摘要:
A software reload is executed. The hardware associated with the network device continues to forward network traffic during the software reload. Also, a kernel of the network device operates unaffected in a protected address space throughout the software reload. Further, the kernel preserves local checkpointed and shared memory data. Application processes running on the network node are shut down gracefully. The reloaded software is brought up and the network device is resynchronized.
摘要:
In a tiered system of networks capable of processing distributed hash table (DHT) Put and Get messages, if a sending network of a Put or Get is at a tier below a subject tier at which a subject network receiving the message is located, the message is forwarded to networks at tiers below the subject tier. On the other hand, if the sending network is at a tier equal to or above the subject tier, the message is forwarded to all networks in the system other than the sending network.
摘要:
A loopback mechanism uses frame relay OAM loopback frames which eliminate the need for location identifiers and reduces overhead and complexity. The frames replace source and destination loopback location identifiers by a loopback indicator field, a send node count field and a return node count field. The originating node of the loopback procedure sets the number of intermediate frame relay nodes to be traversed. The intermediate frame relay nodes simply increment or decrement the node count, depending on the direction of the loopback frame.
摘要:
In one example, a network device receives a packet to be forwarded according to a label switching protocol, determines a service to be performed on the packet by a service network device, sends a label request message to the service network device, wherein the label request message indicates support for labels having a particular length, wherein the particular length is larger than twenty bits (e.g., forty bits), and wherein the label request message specifies the service to be performed on the packet, receives, in response to the label request message, a label mapping message defining a label of the particular length, appends the label to the packet to form a Multi-Protocol Label Switching (MPLS)-encapsulated packet, and forwards the MPLS-encapsulated packet according to the label switching protocol.
摘要:
In general, techniques are described for extending a path computation element (PCE) communication protocol (PCEP) to support messages that enable PCEs to actively modify Multi-Protocol Label Switching (MPLS) for Traffic Engineering Label Switched Paths (TE LSPs) in and across network domains. In one example, an LSP database of a router includes configuration data for one or more LSPs configured in the router and further includes LSP state information specifying a current state of all LSPs of the router. A path computation client (PCC) of a router establishes an extended PCEP session and synchronizes LSP state information to a stateful PCE using the extended PCEP session. Subsequently, the stateful PCE sends an LSP update request to the PCC in the extended PCEP session, wherein the LSP update request includes one or more updated parameters for the LSP. The PCC then re-signals the LSP through the network according to the updated parameters.
摘要:
In a tiered system of networks capable of processing distributed hash table (DHT) Put and Get messages, if a sending network of a Put or Get is at a tier below a subject tier at which a subject network receiving the message is located, the message is forwarded to networks at tiers below the subject tier. On the other hand, if the sending network is at a tier equal to or above the subject tier, the message is forwarded to all networks in the system other than the sending network.
摘要:
In general, techniques are described for using routing information obtained by operation of network routing protocols to dynamically generate network and cost maps for an application-layer traffic optimization (ALTO) service. For example, an ALTO server of an autonomous system (AS) receives routing information from routers of the AS by listening for routing protocol updates outputted by the routers and uses the received topology information to dynamically generate a network map of PIDs that reflects a current topology of the AS and/or of the broader network that includes the AS. Additionally, the ALTO server dynamically calculates inter-PID costs using received routing information that reflects current link metrics. The ALTO server then assembles the inter-PID costs into a cost map that the ALTO server may provide, along with the network map, to clients of the ALTO service.
摘要:
Using the ALTO Service, networking applications can request through the ALTO protocol information about the underlying network topology from the ISP or Content Provider. The ALTO Service provides information such as network resource preferences with the goal of modifying network resource consumption patterns while maintaining or improving application performance. This document describes, in one example, an ALTO server that implements enhancements to the ALTO service to assign a PID-type attribute to each of a set of one or more PIDs each associated with a subset of one or more endpoints of a network, wherein a PID-type attribute specifies a type for the subset of endpoints associated with the PID. The ALTO server generates an ALTO network map that includes a PID entry to describe each of the PIDs, wherein each PID entry includes a PID-type field that stores the assigned PID-type attribute for the PID described by the PID entry.