Efficient implementation of error correction code scheme
    94.
    发明授权
    Efficient implementation of error correction code scheme 失效
    有效执行纠错码方案

    公开(公告)号:US06681340B2

    公开(公告)日:2004-01-20

    申请号:US09792533

    申请日:2001-02-23

    IPC分类号: G06F1110

    CPC分类号: H04L1/0043 H04L1/0063

    摘要: A method and system for efficiently implementing an error correction code scheme. In one embodiment of the present invention, a system comprises a processor configured to process frames of data. Each frame of data may be associated with a frame control block. The processor comprises a first queue configured to store one or more frame control blocks associated with one or more frames of data. The processor further comprises a second queue configured to store one or more frame control blocks not associated with a frame of data. The one or more frame control blocks associated with one or more frames of data in the first queue comprise a bit for storing a parity bit. The one or more frame control blocks in the second queue comprise a plurality of bits for storing a code of an error correction code scheme.

    摘要翻译: 一种用于有效实施纠错码方案的方法和系统。 在本发明的一个实施例中,系统包括被配置为处理数据帧的处理器。 数据帧可以与帧控制块相关联。 处理器包括被配置为存储与一个或多个数据帧相关联的一个或多个帧控制块的第一队列。 处理器还包括被配置为存储与数据帧不相关联的一个或多个帧控制块的第二队列。 与第一队列中的一个或多个数据帧相关联的一个或多个帧控制块包括用于存储奇偶校验位的位。 第二队列中的一个或多个帧控制块包括用于存储纠错码方案的代码的多个比特。

    Network processor for multiprotocol data flows
    95.
    发明授权
    Network processor for multiprotocol data flows 有权
    用于多协议数据流的网络处理器

    公开(公告)号:US06671280B1

    公开(公告)日:2003-12-30

    申请号:US09535794

    申请日:2000-03-29

    IPC分类号: H04L1256

    摘要: A method for integrating Asynchronous Transfer Mode (ATM) and frame-based traffic flows within a telecommunications network is disclosed. The telecommunications network includes a network processor having upside processing means for delivering an incoming flow from the telecommunications network to a switch and downside processing means for delivering outgoing network traffic from the switch to the telecommunications network. The incoming flow is initially received at the upside processing means as a frame-based flow. The incoming flow may be characterized as belonging to a group having frame-based flows and ATM flows. In response to the receipt of the incoming flow, the incoming flow is determined if it is destined for a legacy, ATM-only device. The incoming flow is then processed according to the determined routing requirements and the incoming flow characterization before delivering the incoming flow to the switch.

    摘要翻译: 公开了一种在电信网络内集成异步传输模式(ATM)和基于帧的业务流的方法。 电信网络包括具有上行处理装置的网络处理器,用于将来自电信网络的输入流传送到交换机,以及下行处理装置,用于将来自交换机的输出网络业务传送到电信网络。 最初在上行处理装置处接收输入流作为基于帧的流。 输入流可以被表征为属于具有基于帧的流和ATM流的组。 响应于接收到的流入,确定进入流是否发往传统的仅ATM设备。 然后根据确定的路由要求和输入流特性,将传入流量传送到交换机之前处理进入流。

    Queue manager for a buffer
    98.
    发明授权
    Queue manager for a buffer 失效
    队列管理器为缓冲区

    公开(公告)号:US06557053B1

    公开(公告)日:2003-04-29

    申请号:US09477179

    申请日:2000-01-04

    IPC分类号: G06F1314

    CPC分类号: G06F13/1673

    摘要: A bandwidth conserving queue manager for a FIFO buffer is provided, preferably on an ASIC chip and preferably including separate DRAM storage that maintains a FIFO queue which can extend beyond the data storage space of the FIFO buffer to provide additional data storage space as needed. FIFO buffers are used on the ASIC chip to store and retrieve multiple queue entries. As long as the total size of the queue does not exceed the storage available in the buffers, no additional data storage is needed. However, when some predetermined amount of the buffer storage space in the FIFO buffers is exceeded, data are written to and read from the additional data storage, and preferably in packets which are of optimum size for maintaining peak performance of the data storage device and which are written to the data storage device in such a way that they are queued in a first-in, first-out (FIFO) sequence of addresses. Preferably, the data are written to and are read from the DRAM in burst mode.

    摘要翻译: 提供了用于FIFO缓冲器的带宽保存队列管理器,优选地在ASIC芯片上,并且优选地包括分离的DRAM存储器,其维持FIFO队列,其可以超出FIFO缓冲器的数据存储空间,以根据需要提供附加的数据存储空间。 在ASIC芯片上使用FIFO缓冲器来存储和检索多个队列条目。 只要队列的总大小不超过缓冲区中可用的存储空间,则不需要额外的数据存储。 然而,当超过FIFO缓冲器中的一些预定量的缓冲存储空间时,数据被写入附加数据存储器并从其中读出,并且优选地是具有用于保持数据存储设备的峰值性能的最佳尺寸的数据包,以及哪个 被写入数据存储设备,使得它们以先入先出(FIFO)地址序列排队。 优选地,以突发模式将数据写入DRAM并从DRAM读取。

    Controller for multiple instruction thread processors
    99.
    发明授权
    Controller for multiple instruction thread processors 失效
    多指令线程处理器的控制器

    公开(公告)号:US08006244B2

    公开(公告)日:2011-08-23

    申请号:US10915983

    申请日:2004-08-11

    IPC分类号: G06F9/46

    CPC分类号: G06F9/3802 G06F9/3851

    摘要: A mechanism controls a multi-thread processor so that when a first thread encounters a latency event for a first predefined time interval temporary control is transferred to an alternate execution thread for duration of the first predefined time interval and then back to the original thread. The mechanism grants full control to the alternate execution thread when a latency event for a second predefined time interval is encountered. The first predefined time interval is termed short latency event whereas the second time interval is termed long latency event.

    摘要翻译: 机制控制多线程处理器,使得当第一线程遇到第一预定义时间间隔的等待时间事件时,临时控制在第一预定义时间间隔的持续时间内被传送到备用执行线程,然后返回到原始线程。 当遇到第二个预定义时间间隔的延迟事件时,机制将授权对备用执行线程的完全控制。 第一预定时间间隔称为短延迟事件,而第二时间间隔称为长延迟事件。

    Network processor which makes thread execution control decisions based on latency event lengths
    100.
    发明授权
    Network processor which makes thread execution control decisions based on latency event lengths 失效
    基于延迟事件长度的线程执行控制决策的网络处理器

    公开(公告)号:US07093109B1

    公开(公告)日:2006-08-15

    申请号:US09542189

    申请日:2000-04-04

    IPC分类号: G06F9/46

    CPC分类号: G06F9/3851 G06F9/3802

    摘要: A control mechanism is established between a network processor and a tree search coprocessor to deal with latencies in accessing the data such as information formatted in a tree structure. A plurality of independent instruction execution threads are queued to enable them to have rapid access to the shared memory. If execution of a thread becomes stalled due to a latency event, full control is granted to the next thread in the queue. The grant of control is temporary when a short latency event occurs or full when a long latency event occurs. Control is returned to the original thread when a short latency event is completed. Each execution thread utilizes an instruction prefetch buffer that collects instructions for idle execution threads when the instruction bandwidth is not fully utilized by an active execution thread. The thread execution control is governed by the collective functioning of a FIFO, an arbiter and a thread control state machine.

    摘要翻译: 在网络处理器和树形搜索协处理器之间建立一种控制机制来处理访问诸如以树结构格式化的信息的数据的延迟。 排队多个独立的指令执行线程使其能够快速访问共享存储器。 如果由于延迟事件导致线程执行失败,则会对队列中的下一个线程授予完全控制权。 当长时间延迟事件发生时,发生短延迟事件或满时,授权控制是暂时的。 当短暂延迟事件完成时,控制返回到原始线程。 每个执行线程使用指令预取缓冲器,当指令带宽未被活动执行线程充分利用时,该指令预取缓冲器收集空闲执行线程的指令。 线程执行控制由FIFO,仲裁器和线程控制状态机的集合功能决定。