摘要:
A system for sharing a computational resource by buffering multiple requests from multiple devices to a memory (e.g. a multi-port RAM or FIFO) in a single clock cycle. The system includes a memory having a first write port and a second write port. A first request input is coupled to the first write port. A second request input is coupled to the second write port. A controller is coupled to the memory. The controller is configured to control the memory to store a first request into the memory via the first write port and a second request into the memory via the second write port. The first and second requests are received via the first and second request inputs and stored into the memory in one clock cycle. Requests are removed from the memory sequentially at a rate that is determined by the shared computational resource.
摘要:
In one embodiment of the invention, a method for prioritizing network packets, includes: comparing a packet with at least one copy rule; and if the packet matches the copy rule, then buffering the packet in a queue. The method further includes: processing the packet after buffering the packet in the queue.
摘要:
A method of selective direct memory access (DMA) in a computer system having a network interface card (NIC), a processor subsystem, and a memory accessible to the subsystem, including the steps of the NIC receiving a plurality of packets from an external source; the NIC decoding a portion of each packet and determining whether the packet should be accepted by the computer system based on a predefined rule; and if the packet is accepted, selectively truncating the packet based on the predefined rule, and storing the truncated packet in the memory.
摘要:
Systems, devices, and methods, including executable instructions are provided for content addressable memory (CAM). One method includes defining the CAM into an array of data words having M rows and N columns, with each of N and M being greater than one. The data words of the CAM are arranged according to a 2-dimensional priority scheme. Data words outside a selected 1×M column are masked to be ignored in determining a match, and the CAM is searched. Each search includes N compare cycles and each compare cycle having a different 1×M column selected. A highest priority match per compare cycle is pipelined from a priority encoder with the pipelined matches arranged to communicate a priority order in a first dimension of the 2-dimensional priority scheme.
摘要:
In an embodiment of the invention, a method and system for a per-port penalty queue system in a network device includes: selecting a state for a port in the network device; wherein the selected state comprises either a normal state or a restricted state; wherein the normal state permits a packet received at the port to be copied to a first queue; and wherein the restricted state causes the packet to be copied to a penalty queue which has lower priority than the first queue or causes the packet to not be copied to a queue. In another embodiment of the invention, a method and system permit using the port state for modifying a forwarding decision for a packet, so that the penalized packet will use a sub-optimal or less optimal routing path to the packet destination. In another embodiment of the invention, a method and system permit using the port state as a search key into an access control list (ACL) operation related to packet forwarding decisions or packet filtering decisions.
摘要:
One embodiment disclosed relates to a method for remote mirroring of network traffic. A data packet to be remotely mirrored is received by an entry device. The entry device is pre-configured with a destination Internet Protocol (IP) address to which to mirror the data packet. An IP header is generated and added to IP encapsulate the data packet. The IP header includes the aforementioned destination IP address. The IP-encapsulated packet is forwarded to an exit device associated with the destination IP address. Subsequently, the exit device may decapsulate the IP-encapsulated packet to reproduce the original data packet.
摘要:
Dynamic hardware address assignment to network devices in a switch mesh is described. In an example, a network device in a switch mesh dynamically assigns a hardware address to a network interface. The network device receives an address request from a client through the switch mesh. The network device calculates an index value for the client based on data in the address request. The network device selects the hardware address of a plurality of hardware addresses assigned to the network interface based on the index value. The switch mesh includes a plurality of traffic paths respectively for the plurality of hardware addresses. The network device sends the hardware address in a reply to the client through the switch mesh.
摘要:
Dynamic hardware address assignment to network devices in a switch mesh is described. In an example, a network device in a switch mesh dynamically assigns a hardware address to a network interface. The network device receives an address request from a client through the switch mesh. The network device calculates an index value for the client based on data in the address request. The network device selects the hardware address of a plurality of hardware addresses assigned to the network interface based on the index value. The switch mesh includes a plurality of traffic paths respectively for the plurality of hardware addresses. The network device sends the hardware address in a reply to the client through the switch mesh.
摘要:
A method of selective direct memory access (DMA) in a computer system having a network interface card (NIC), a processor subsystem, and a memory accessible to the subsystem, including the steps of the NIC receiving a plurality of packets from an external source; the NIC decoding a portion of each packet and determining whether the packet should be accepted by the computer system based on a predefined rule; and if the packet is accepted, selectively truncating the packet based on the predefined rule, and storing the truncated packet in the memory.
摘要:
An embodiment of the invention evaluates the suitability of routes for use in a network route cache. An embodiment of the invention allows the use of network routing caches to be optimized such that routes that are not used to route to a large number of destinations can be offloaded into a less expensive host route cache.