Latency insensitive FIFO signaling protocol
    1.
    发明授权
    Latency insensitive FIFO signaling protocol 有权
    延迟不敏感的FIFO信令协议

    公开(公告)号:US07454538B2

    公开(公告)日:2008-11-18

    申请号:US11128135

    申请日:2005-05-11

    IPC分类号: G06F3/00

    摘要: Data from a source domain operating at a first data rate is transferred to a FIFO in another domain operating at a different data rate. The FIFO buffers data before transfer to a sink for further processing or storage. A source side counter tracks space available in the FIFO. In disclosed examples, the initial counter value corresponds to FIFO depth. The counter decrements in response to a data ready signal from the source domain, without delay. The counter increments in response to signaling from the sink domain of a read of data off the FIFO. Hence, incrementing is subject to the signaling latency between domains. The source may send one more beat of data when the counter indicates the FIFO is full. The last beat of data is continuously sent from the source until it is indicated that a FIFO position became available; effectively providing one more FIFO position.

    摘要翻译: 来自以第一数据速率运行的源域的数据被传送到以不同数据速率工作的另一个域中的FIFO。 FIFO在传输到宿之前缓冲数据以进一步处理或存储。 源端计数器跟踪FIFO中可用的空间。 在公开的示例中,初始计数器值对应于FIFO深度。 响应于来自源域的数据就绪信号,计数器无延迟地递减。 响应于来自接收器域的信令从FIFO读取数据,计数器递增。 因此,增量受到域之间的信令等待时间的限制。 当计数器指示FIFO已满时,源可能再发送一次数据。 数据的最后一次节拍从源头连续发送到指示FIFO位置可用为止; 有效提供一个FIFO位置。

    Latency insensitive FIFO signaling protocol
    2.
    发明授权
    Latency insensitive FIFO signaling protocol 有权
    延迟不敏感的FIFO信令协议

    公开(公告)号:US07725625B2

    公开(公告)日:2010-05-25

    申请号:US12179970

    申请日:2008-07-25

    IPC分类号: G06F3/00

    摘要: Data from a source domain operating at a first data rate is transferred to a FIFO in another domain operating at a different data rate. The FIFO buffers data before transfer to a sink for further processing or storage. A source side counter tracks space available in the FIFO. In disclosed examples, the initial counter value corresponds to FIFO depth. The counter decrements in response to a data ready signal from the source domain, without delay. The counter increments in response to signaling from the sink domain of a read of data off the FIFO. Hence, incrementing is subject to the signaling latency between domains. The source may send one more beat of data when the counter indicates the FIFO is full. The last beat of data is continuously sent from the source until it is indicated that a FIFO position became available; effectively providing one more FIFO position.

    摘要翻译: 来自以第一数据速率运行的源域的数据被传送到以不同数据速率工作的另一个域中的FIFO。 FIFO在传输到宿之前缓冲数据以进一步处理或存储。 源端计数器跟踪FIFO中可用的空间。 在公开的示例中,初始计数器值对应于FIFO深度。 响应于来自源域的数据就绪信号,计数器无延迟地递减。 响应于来自接收器域的信令从FIFO读取数据,计数器递增。 因此,增量受到域之间的信令等待时间的限制。 当计数器指示FIFO已满时,源可能再发送一次数据。 数据的最后一次节拍从源头连续发送到指示FIFO位置可用为止; 有效提供一个FIFO位置。

    Latency Insensitive FIFO Signaling Protocol
    3.
    发明申请
    Latency Insensitive FIFO Signaling Protocol 有权
    延迟不敏感的FIFO信令协议

    公开(公告)号:US20080281996A1

    公开(公告)日:2008-11-13

    申请号:US12179970

    申请日:2008-07-25

    IPC分类号: G06F13/38 G06F3/00

    摘要: Data from a source domain operating at a first data rate is transferred to a FIFO in another domain operating at a different data rate. The FIFO buffers data before transfer to a sink for further processing or storage. A source side counter tracks space available in the FIFO. In disclosed examples, the initial counter value corresponds to FIFO depth. The counter decrements in response to a data ready signal from the source domain, without delay. The counter increments in response to signaling from the sink domain of a read of data off the FIFO. Hence, incrementing is subject to the signaling latency between domains. The source may send one more beat of data when the counter indicates the FIFO is full. The last beat of data is continuously sent from the source until it is indicated that a FIFO position became available; effectively providing one more FIFO position.

    摘要翻译: 来自以第一数据速率运行的源域的数据被传送到以不同数据速率工作的另一个域中的FIFO。 FIFO在传输到宿之前缓冲数据以进一步处理或存储。 源端计数器跟踪FIFO中可用的空间。 在公开的示例中,初始计数器值对应于FIFO深度。 响应于来自源域的数据就绪信号,计数器无延迟地递减。 响应于来自接收器域的信令从FIFO读取数据,计数器递增。 因此,增量受到域之间的信令等待时间的限制。 当计数器指示FIFO已满时,源可能再发送一次数据。 数据的最后一次节拍从源头连续发送到指示FIFO位置可用为止; 有效提供一个FIFO位置。

    Latency insensitive FIFO signaling protocol
    4.
    发明申请
    Latency insensitive FIFO signaling protocol 有权
    延迟不敏感的FIFO信令协议

    公开(公告)号:US20060259669A1

    公开(公告)日:2006-11-16

    申请号:US11128135

    申请日:2005-05-11

    IPC分类号: G06F13/38

    摘要: Data from a source domain operating at a first data rate is transferred to a FIFO in another domain operating at a different data rate. The FIFO buffers data before transfer to a sink for further processing or storage. A source side counter tracks space available in the FIFO. In disclosed examples, the initial counter value corresponds to FIFO depth. The counter decrements in response to a data ready signal from the source domain, without delay. The counter increments in response to signaling from the sink domain of a read of data off the FIFO. Hence, incrementing is subject to the signaling latency between domains. The source may send one more beat of data when the counter indicates the FIFO is full. The last beat of data is continuously sent from the source until it is indicated that a FIFO position became available; effectively providing one more FIFO position.

    摘要翻译: 来自以第一数据速率运行的源域的数据被传送到以不同数据速率工作的另一个域中的FIFO。 FIFO在传输到宿之前缓冲数据以进一步处理或存储。 源端计数器跟踪FIFO中可用的空间。 在公开的示例中,初始计数器值对应于FIFO深度。 响应于来自源域的数据就绪信号,计数器无延迟地递减。 响应于来自接收器域的信令从FIFO读取数据,计数器递增。 因此,增量受到域之间的信令等待时间的限制。 当计数器指示FIFO已满时,源可能再发送一次数据。 数据的最后一次节拍从源头连续发送到指示FIFO位置可用为止; 有效提供一个FIFO位置。

    Method and apparatus for managing cache partitioning using a dynamic boundary
    5.
    发明授权
    Method and apparatus for managing cache partitioning using a dynamic boundary 有权
    使用动态边界管理缓存分区的方法和装置

    公开(公告)号:US07650466B2

    公开(公告)日:2010-01-19

    申请号:US11233575

    申请日:2005-09-21

    IPC分类号: G06F12/00

    CPC分类号: G06F12/126

    摘要: A method of managing cache partitions provides a first pointer for higher priority writes and a second pointer for lower priority writes, and uses the first pointer to delimit the lower priority writes. For example, locked writes have greater priority than unlocked writes, and a first pointer may be used for locked writes, and a second pointer may be used for unlocked writes. The first pointer is advanced responsive to making locked writes, and its advancement thus defines a locked region and an unlocked region. The second pointer is advanced responsive to making unlocked writes. The second pointer also is advanced (or retreated) as needed to prevent it from pointing to locations already traversed by the first pointer. Thus, the pointer delimits the unlocked region and allows the locked region to grow at the expense of the unlocked region.

    摘要翻译: 管理高速缓存分区的方法提供用于较高优先级写入的第一指针和用于较低优先级写入的第二指针,并且使用第一指针来划分较低优先级的写入。 例如,锁定的写入具有比解锁的写入更高的优先级,并且第一指针可以用于锁定的写入,并且第二指针可以用于解锁的写入。 响应于锁定写入,第一指针是高级的,并且其进步因此定义了锁定区域和解锁区域。 响应于解锁写入,第二个指针是高级的。 第二个指针也根据需要进行高级(或撤销),以防止它指向已经被第一个指针所遍历的位置。 因此,指针限定未锁定区域,并允许锁定区域以解锁区域为代价而增长。

    Power efficient instruction prefetch mechanism
    6.
    发明授权
    Power efficient instruction prefetch mechanism 有权
    高效的指令预取机制

    公开(公告)号:US08661229B2

    公开(公告)日:2014-02-25

    申请号:US12434804

    申请日:2009-05-04

    IPC分类号: G06F9/30 G06F9/40 G06F15/00

    CPC分类号: G06F9/3844 G06F9/3804

    摘要: A processor includes a conditional branch instruction prediction mechanism that generates weighted branch prediction values. For weakly weighted predictions, which tend to be less accurate than strongly weighted predictions, the power associating with speculatively filling and subsequently flushing the cache is saved by halting instruction prefetching. Instruction fetching continues when the branch condition is evaluated in the pipeline and the actual next address is known. Alternatively, prefetching may continue out of a cache. To avoid displacing good cache data with instructions prefetched based on a mispredicted branch, prefetching may be halted in response to a weakly weighted prediction in the event of a cache miss.

    摘要翻译: 处理器包括产生加权分支预测值的条件分支指令预测机制。 对于弱加权预测,其倾向于比强加权预测不太准确,通过停止指令预取来节省与推测性填充和随后刷新高速缓存的功率相关联的功率。 当流水线中评估分支条件并且已知实际的下一个地址时,指令获取继续。 或者,可以从高速缓存中继续预取。 为了避免基于错误预测的分支预取的指令移位良好的高速缓存数据,预取可以响应于在高速缓存未命中的情况下的弱加权预测而停止。

    Power Efficient Instruction Prefetch Mechanism
    7.
    发明申请
    Power Efficient Instruction Prefetch Mechanism 有权
    高效率指令预取机制

    公开(公告)号:US20090210663A1

    公开(公告)日:2009-08-20

    申请号:US12434804

    申请日:2009-05-04

    IPC分类号: G06F9/38

    CPC分类号: G06F9/3844 G06F9/3804

    摘要: A processor includes a conditional branch instruction prediction mechanism that generates weighted branch prediction values. For weakly weighted predictions, which tend to be less accurate than strongly weighted predictions, the power associating with speculatively filling and subsequently flushing the cache is saved by halting instruction prefetching. Instruction fetching continues when the branch condition is evaluated in the pipeline and the actual next address is known. Alternatively, prefetching may continue out of a cache. To avoid displacing good cache data with instructions prefetched based on a mispredicted branch, prefetching may be halted in response to a weakly weighted prediction in the event of a cache miss.

    摘要翻译: 处理器包括产生加权分支预测值的条件分支指令预测机制。 对于弱加权预测,其倾向于比强加权预测不太准确,通过停止指令预取来节省与推测性填充和随后刷新高速缓存的功率相关联的功率。 当流水线中评估分支条件并且已知实际的下一个地址时,指令获取继续。 或者,可以从高速缓存中继续预取。 为了避免基于错误预测的分支预取的指令移位良好的高速缓存数据,预取可以响应于在高速缓存未命中的情况下的弱加权预测而停止。

    TLB lock indicator
    8.
    发明授权
    TLB lock indicator 有权
    TLB锁定指示灯

    公开(公告)号:US07426626B2

    公开(公告)日:2008-09-16

    申请号:US11210526

    申请日:2005-08-23

    IPC分类号: G06F12/00

    摘要: A processor includes a hierarchical Translation Lookaside Buffer (TLB) comprising a Level-1 TLB and a small, high-speed Level-0 TLB. Entries in the L0 TLB replicate entries in the L1 TLB. The processor first accesses the L0 TLB in an address translation, and access the L1 TLB if a virtual address misses in the L0 TLB. When the virtual address hits in the L1 TLB, the virtual address, physical address, and page attributes are written to the L0 TLB, replacing an existing entry if the L0 TLB is full. The entry may be locked against replacement in the L0 TLB in response to an L0 Lock (L0L) indicator in the L1 TLB entry. Similarly, in a hardware-managed L1 TLB, entries may be locked against replacement in response to an L1 Lock (L1L) indicator in the corresponding page table entry.

    摘要翻译: 处理器包括包括Level-1 TLB和小的高速Level-0 TLB的分级翻译后备缓冲器(TLB)。 L0 TLB中的条目复制L1 TLB中的条目。 处理器首先在地址转换中访问L0 TLB,如果在L0 TLB中虚拟地址丢失,则访问L1 TLB。 当虚拟地址在L1 TLB中时,虚拟地址,物理地址和页面属性被写入L0 TLB,如果L0 TLB已满,则替换现有的条目。 响应于L1 TLB条目中的L0锁定(L0L)指示灯,该条目可能被锁定在L0 TLB中。 类似地,在硬件管理的L1 TLB中,可以响应于相应页表条目中的L1锁定(L1L)指示符来锁定条目以替代。

    Cache locking without interference from normal allocations
    9.
    发明授权
    Cache locking without interference from normal allocations 有权
    缓存锁定,不受正常分配的干扰

    公开(公告)号:US08527713B2

    公开(公告)日:2013-09-03

    申请号:US11343765

    申请日:2006-01-31

    摘要: A Block Normal Cache Allocation (BNCA) mode is defined for a processor. In BNCA mode, cache entries may only be allocated by predetermined instructions. Normal memory access instructions (for example, as part of interrupt code) may execute and will retrieve data from main memory in the event of a cache miss; however, these instructions are not allowed to allocate entries in the cache. Only the predetermined instructions (for example, those used to establish locked cache entries) may allocate entries in the cache. When the locked entries are established, the processor exits BNCA mode, and any memory access instruction may allocate cache entries. BNCA mode may be indicated by setting a bit in a configuration register.

    摘要翻译: 为处理器定义块正常缓存分配(BNCA)模式。 在BNCA模式中,缓存条目只能通过预定指令分配。 正常存储器访问指令(例如,作为中断代码的一部分)可以执行并且将在高速缓存未命中的情况下从主存储器检索数据; 但是,这些指令不允许在缓存中分配条目。 只有预定的指令(例如,用于建立锁定的高速缓存条目的指令)可以在高速缓存中分配条目。 当锁定条目建立时,处理器退出BNCA模式,任何存储器访问指令都可以分配高速缓存条目。 可以通过在配置寄存器中设置位来指示BNCA模式。

    Translation lookaside buffer manipulation
    10.
    发明授权
    Translation lookaside buffer manipulation 有权
    翻译后备缓冲操作

    公开(公告)号:US07721067B2

    公开(公告)日:2010-05-18

    申请号:US11336264

    申请日:2006-01-20

    IPC分类号: G06F9/34

    CPC分类号: G06F9/3861 G06F12/1027

    摘要: A processor having a multistage pipeline includes a TLB and a TLB controller. In response to a TLB miss signal, the TLB controller initiates a TLB reload, requesting address translation information from either a memory or a higher-level TLB, and placing that information into the TLB. The processor flushes the instruction having the missing virtual address, and refetches the instruction, resulting in re-insertion of the instruction at an initial stage of the pipeline above the TLB access point. The initiation of the TLB reload, and the flush/refetch of the instruction, are performed substantially in parallel, and without immediately stalling the pipeline. The refetched instruction is held at a point in the pipeline above the TLB access point until the TLB reload is complete, so that the refetched instruction generates a “hit” in the TLB upon its next access.

    摘要翻译: 具有多级流水线的处理器包括TLB和TLB控制器。 响应于TLB未命中信号,TLB控制器启动TLB重新加载,从存储器或更高级TLB请求地址转换信息,并将该信息放入TLB。 处理器刷新具有缺失的虚拟地址的指令,并且重新指示该指令,从而导致在TLB接入点上方的管线的初始阶段重新插入该指令。 TLB重新启动的启动以及指令的刷新/刷新基本上并行执行,并且不会立即停止管道。 重写指令在TLB接入点上方的管道中保持一段时间,直到TLB重新加载完成,这样,重写指令在下一次访问时就会在TLB中产生一个“命中”。