Abstract:
A content delivery node receives data packets carrying content from an upstream source of content, and writes segments of the received content directly to a memory buffer of a memory using direct memory access (DMA) data transfers. The node derives, for each segment, respective segment-specific metadata based on contents of the segment, and stores the respective segment-specific metadata in the memory. The node receives from multiple downstream client devices respective requests for the same content. Each request includes client-specific information. Responsive to the requests, the node: identifies one or more segments that satisfy the requests; generates, for each client device, client-specific metadata using the client-specific information and the segment-specific metadata for the one or more segments; constructs, for each client, a client-specific data packet that includes the one or more segments and the client-specific metadata; and transmits the client-specific data packets to the downstream client devices.
Abstract:
A method is provided in one example embodiment and includes, for each of a plurality of individual storage units collectively comprising a virtual storage unit, mapping an internal address of the storage unit to a unique IP address, wherein each of the storage units comprises a block of storage on one of a plurality of physical storage devices and wherein the IP address includes a virtual storage unit number identifying the virtual storage unit; receiving from a client a request to perform an operation on at least one of the data storage units, wherein the request identifies the internal address of the at least one of the data storage units; translating the internal address of the at least one of the data storage unit to the unique IP address of the at least one of the data storage units; and performing the requested operation on the at least one of the data storage units.
Abstract:
A system includes an access node having an associated identifier. The access node is configured to insert the identifier into a network connection request. The system includes an IP edge service node connected to the access node and configured to receive the network connection request. The IP edge service node is further configured to store the inserted identifier and to insert the identifier into an Internet protocol version 6 (IPv6) address request transmitted according to dynamic host configuration protocol version 6 (DHCPv6) through an established network connection based on the network connection request. The system further includes a server configured to receive the IPv6 address request and assign an IPv6 address based on the IPv6 address request.
Abstract:
In one embodiment, for each port of an access node in an access-based computer network, one access server of a plurality of access servers is configured as a preferred access server for that port. Upon receiving a session initiation message at a particular port, the access node forwards the session initiation message to one or more of the access servers based on the configured preferred access server for the particular port.
Abstract:
A system includes an access node having an associated identifier. The access node is configured to insert the identifier into a network connection request. The system includes an IP edge service node connected to the access node and configured to receive the network connection request. The IP edge service node is further configured to store the inserted identifier and to insert the identifier into an Internet protocol version 6 (IPv6) address request transmitted according to dynamic host configuration protocol version 6 (DHCPv6) through an established network connection based on the network connection request. The system further includes a server configured to receive the IPv6 address request and assign an IPv6 address based on the IPv6 address request.
Abstract:
This disclosure describes techniques for providing a distributed scalable architecture for Network Address Translation (NAT) systems with high availability and mitigations for flow breakage during failover events. The NAT servers may include functionality to serve as fast-path servers and/or slow-path servers. A fast-path server may include a NAT worker that includes a cache of NAT mappings to perform stateful network address translation and to forward packets with minimal latency. A slow-path server may include a mapping server that creates new NAT mappings, depreciates old ones, and answers NAT worker state requests. The NAT system may use virtual mapping servers (VMSs) running on primary physical servers with state duplicated VMSs on different physical failover servers. Additionally, the NAT servers may implement failover solutions for dynamically allocated routeable address/port pairs assigned to new sessions by assigning new outbound address/port pairs when a session starts and broadcasting pairing information.
Abstract:
This disclosure describes methods and systems to externally manage network-to-network interconnect configuration data in conjunction with a centralized database subsystem. An example of the methods includes receiving and storing, in the centralized database subsystem, data indicative of user intent to interconnect at least a first network and a second network. The example method further includes, based at least in part on the data indicative of user intent, determining and storing, in the centralized database subsystem, a network intent that corresponds to the user intent. The example method further includes providing data indicative of the network intent from the centralized database subsystem to a first data plane adaptor, associated with the first network, and a second data plane adaptor, associated with the second network.
Abstract:
A method is provided in one example embodiment and includes, for each of a plurality of individual storage units collectively comprising a virtual storage unit, mapping an internal address of the storage unit to a unique IP address, wherein each of the storage units comprises a block of storage on one of a plurality of physical storage devices and wherein the IP address includes a virtual storage unit number identifying the virtual storage unit; receiving from a client a request to perform an operation on at least one of the data storage units, wherein the request identifies the internal address of the at least one of the data storage units; translating the internal address of the at least one of the data storage unit to the unique IP address of the at least one of the data storage units; and performing the requested operation on the at least one of the data storage units.
Abstract:
Aspects of the subject technology provide state-less load-balancing using sequence numbers to identify traffic flows. In some implementations, a process of the technology can include steps for receiving, by a load-balancer, a first packet from a source device including a request to access the service provided by a server coupled to the load-balancer, determining a load for each of the servers, wherein each server is associated with a unique set of sequence numbers, and forwarding the request to a target server selected based on its corresponding load, and wherein the request is configured to cause the target server to issue a reply to the source device. Systems and machine-readable media are also provided.
Abstract:
Techniques for zero-loss workload mobility with segment routing for virtual machines are presented. The techniques include receiving, by a virtual router, an electronic message destined for a first virtual machine running on a first physical machine and checking a first virtual machine state for the first virtual machine. In response to determining that it is associated with a running state indicating the first physical machine, inserting a segment routing header including an indication of the source virtual machine, the first physical machine, and the first virtual machine. In response to determining that it is associated with a migration state, inserting, by the virtual router, a segment routing header indicating the source virtual machine, an END.S for the first physical machine, the first virtual machine; and an END.SBUF for a second physical machine. The message is then routed based at least in part on the inserted segment routing header.