摘要:
A load-balanced network architecture is disclosed in which a traffic flow at a given network node is split into a plurality of parts, and the parts are distributed to respective ones of the plurality of nodes that are designated as participating in a load balancing process for the traffic flow. Each of at least a subset of the participating nodes receiving one of the parts routes at least a portion of its received part to one or more destination nodes.
摘要:
A load-balanced network architecture is disclosed in which a traffic flow deliverable from a source node to a destination node via intermediate nodes is split into parts, and the parts are distributed to respective ones of the intermediate nodes. Path delay differences for the parts are substantially equalized by delay adjustment at one or more of the intermediate nodes, and packets of one or more of the parts are scheduled for routing from respective ones of the intermediate nodes to the destination node based on arrival times of the packets at the source node.
摘要:
A load-balanced network architecture is disclosed in which a traffic flow at a given network node is split into a plurality of parts, and the parts are distributed to respective ones of the plurality of nodes that are designated as participating in a load balancing process for the traffic flow. Each of at least a subset of the participating nodes receiving one of the parts routes at least a portion of its received part to one or more destination nodes.
摘要:
A fast shared path allocation technique is disclosed. Network nodes are pre-configured such that data from multiple data sources or multiple primary data paths may be sent via a shared secondary data path. Merge nodes merge input from a plurality of input ports onto an output port. The merge nodes implement a blocking function such that upon receipt of a signal from one of the input ports, the signals from the other input ports are blocked from reaching the output port. Upon a triggering event indicating a need to allocate the shared path, the data is first sent to the merge node where it is appropriately merged onto the output link and transmitted towards its destination. Only after the data has been sent does the merge node block the remaining input ports from reaching the output port. This blocking may be performed automatically by the merge node or by conventional network signaling.
摘要:
Network design techniques and techniques for routing virtually-concatenated data traffic in a network in a manner which distributes delay to intermediate nodes of the network are disclosed. For example, in one aspect of the invention, a technique for routing virtually-concatenated data traffic in a network comprising a plurality of nodes comprises, for a given traffic demand to be routed from a source node to a destination node in the network, the following steps/operations. Two or more paths are determined to route the given traffic demand. Each of the two or more paths correspond to a member of a virtually-concatenated group. At least one path of the two or more paths comprises the source node, the destination node and at least one other node coupled between the source node and the destination node. Further, at least a subset of the source node, the destination node and the one other node buffer at least a portion of the given traffic demand such that a delay is distributed over the at least one path. The given traffic demand is routed over the two or more determined paths. The at least one path is preferably the shorter of the two or more determined paths.
摘要:
A method for location tracking on a distributed basis using multiple locations, which utilizes a pairwise application of distance constraints and vicinities for determining locations. The location of a particular node is represented by a group of points (as opposed to a single point) defined by the vicinity.
摘要:
Virtually-concatenated data traffic is routed in a network comprising a plurality of nodes, the plurality of nodes including one or more nodes that are differential delay enabled and one or more nodes that are not differential delay enabled. For a given traffic demand to be routed from a source node to a destination node in the network, at least one route is determined for routing the demand between an intermediate node that is differential delay enabled and one of the source node and the destination node that is not differential delay enabled. Also, a set of routes is determined for routing the demand between the intermediate node that is differential delay enabled and at least one other node of the network, that may or may not be differential delay enabled, with each of the routes in the set corresponding to a member of a virtually-concatenated group. The given traffic demand is routed from the source node to the destination node, utilizing the at least one route and the set of routes.
摘要:
Multi-path routing techniques using intra-flow splitting are disclosed. For example, a technique for processing traffic flows at a node in a network comprises the following steps/operations. At least one traffic flow is obtained. The at least one traffic flow comprises multiple packets or bytes. The at least one flow is split into at least two sub-flows, wherein each of the at least two sub-flows comprises a portion of the multiple packets or bytes. The packets or bytes of the at least two sub-flows are respectively routed on at least two paths in the network.
摘要:
Techniques and systems for designing a network providing communication and location identification services are described. A solution point comprising parameters for each of a plurality of base stations is generated. A coverage and locatability performance value for the solution point is computed, as well as derivatives of the performance value. The coverage and locatability performance value and its derivatives are used to indicate favorable directions for searching for subsequent solution points, and subsequent solution points are generated and compared against previous solution points until an optimum solution point is found. The coverage and locatability performance value is a weighted sum of coverage and locatability values for each point in the service area of the network, with coverage values representing forward and reverse link quality and locatability values representing the probability that a point will experience an acceptable power level from at least four base stations.
摘要:
Techniques for single-failure protection in load-balanced network architectures are disclosed. For example, in one aspect of the invention, a technique for processing a traffic flow in a communication network comprising a plurality of nodes, the traffic flow being deliverable from a source node to at least one destination node via one or more intermediate nodes, comprises the following steps/operations. The traffic flow is split at the source node into a plurality of parts. The parts are distributed to respective ones of the intermediate nodes such that the parts are routed from the source node to the at least one destination node in a disjoint manner.