摘要:
Systems and methods of controlling and optimizing system latency in order to minimize system latency between remotely operating computing devices over a communications network are disclosed. A method of controlling and optimizing latency over a computer network from a first computing device to a second computing device includes identifying a condition associated with the transmission of a video signal, receiving the video signal at a computing device, determining whether the condition is met after receiving the video signal, and displaying the video signal on a display device when the condition is met.
摘要:
Systems and methods of controlling and optimizing system latency in order to minimize system latency between remotely operating computing devices over a communications network are disclosed. A method of controlling and optimizing latency over a computer network from a first computing device to a second computing device includes identifying a condition associated with the transmission of a video signal, receiving the video signal at a computing device, determining whether the condition is met after receiving the video signal, and displaying the video signal on a display device when the condition is met.
摘要:
A server computer that is connected over a network to one or more remote client computers includes specialized hardware for handling computationally-intensive signal processing operations involving video, audio, and/or other signals. As a result, users of the client computers experience little or no delay when accessing over the network applications that are hosted on the server.
摘要:
A server computer that is connected over a network to one or more remote client computers includes specialized hardware for handling computationally-intensive signal processing operations involving video, audio, and/or other signals. As a result, users of the client computers experience little or no delay when accessing over the network applications that are hosted on the server.
摘要:
The present invention is based on a ring architecture wherein each chip in the multiplicity of chips is connected in a chain so that it can receive packets from the next previous chip in the chain and forward packets to the next following chip in the chain, whereby to form a ring. Before a packet is placed on the ring by a device it is provided with a ‘masking’ field which identifies each of the other devices and the ports within any such device to which the packet must be sent in order to achieve a multicast. When a device in its turn receives such a packet it will examine the masking field to determine whether the field identifies that device. If so, the packet, which may be held in temporary storage, may be processed for forwarding from a port of that device. The device will strip from the masking field the identification of the respective device and transmit the packet with a thus modified masking field to the next device in turn. If a packet does not contain any indication of another device in the ring, the packet will not be transmitted and therefore will naturally terminate.
摘要:
Detection of a signature in a data packet comprises performing a pre-classification of the packet, using header information and particularly a 5-tuple access control list, into one of a multiplicity of flows and directing the payload of the packet to a respective one of a multiplicity of deterministic finite state machines each of which stores a plurality of signatures as a sequence of states and acts only on the respective flow.
摘要:
A network stack includes a plurality of network units each of which includes a multiplicity of ports for receiving and forwarding addressed data packets, at least two cascade ports and a switching engine for forwarding received packets to at least one port in accordance with address data in the packets and a cascade connection including, for each of two opposite directions around the stack, at least one unidirectional path for data packets composed of links each between a respective cascade port on a network unit and a corresponding cascade port on the next network unit.
摘要:
Network units such as switches for use in a cascaded stack are organised to provide a cascade connection in the form of a dual unidirectional connection so that, in its ordinary configuration, there is at least one and preferably more than one unidirectional ring for each direction around the cascade, each ring including a respective port on each unit. For each ring, each port on a unit is connected by a respective link to a corresponding port on the preceding unit and the following unit. The units provide a self-healing operation in the event of various kinds of operational failure. The self-healing operation includes loop-back of packets in units adjacent the failure and bypass of a packet switching process for other units. The units include control logic for passing control frames containing status information relating to the units and links between them and for co-operation with a CPU to control a switching engine to perform the self-healing operation in accordance with that status information. The units forward on the cascade packets with headers that identify a destination port and the unit on which that port is located and also indicate which units have and have not been traversed by a packet.
摘要:
In one embodiment, a method is described for detecting an operational failure between the network unit and an adjacent network unit in the stack; controlling the switching engine to redirect packets which would otherwise be sent from a particular port to the adjacent network unit to be forwarded from another port to be sent to a different network unit in the stack; and entering the switching engine into a bypass mode in response to control data indicating an operational failure between at least two other network units in the stack to cause packets to be forwarded without being re-directed by the switching engine.
摘要:
A network stack includes a plurality of network units each of which includes a multiplicity of ports for receiving and forwarding addressed data packets, at least two cascade ports and a switching engine for forwarding received packets to at least one port in accordance with address data in the packets and a cascade connection including, for each of two opposite directions around the stack, at least one unidirectional path for data packets composed of links each between a respective cascade port on a network unit and a corresponding cascade port on the next network unit.