Cache memory, processing unit, data processing system and method for assuming a selected invalid coherency state based upon a request source
    11.
    发明申请
    Cache memory, processing unit, data processing system and method for assuming a selected invalid coherency state based upon a request source 有权
    高速缓冲存储器,处理单元,数据处理系统和方法,用于基于请求源假设所选择的无效一致性状态

    公开(公告)号:US20060236037A1

    公开(公告)日:2006-10-19

    申请号:US11109085

    申请日:2005-04-19

    IPC分类号: G06F13/28

    摘要: At a first cache memory affiliated with a first processor core, an exclusive memory access operation is received via an interconnect fabric coupling the first cache memory to second and third cache memories respectively affiliated with second and third processor cores. The exclusive memory access operation specifies a target address. In response to receipt of the exclusive memory access operation, the first cache memory detects presence or absence of a source indication indicating that the exclusive memory access operation originated from the second cache memory to which the first cache memory is coupled by a private communication network to which the third cache memory is not coupled. In response to detecting presence of the source indication, a coherency state field of the first cache memory that is associated with the target address is updated to a first data-invalid state. In response to detecting absence of the source indication, the coherency state field of the first cache memory is updated to a different second data-invalid state.

    摘要翻译: 在与第一处理器核心相关联的第一高速缓冲存储器处,通过将第一高速缓冲存储器耦合到分别隶属于第二和第三处理器核的第二和第三高速缓冲存储器的互连结构接收独占存储器存取操作。 独占内存访问操作指定目标地址。 响应于独占存储器访问操作的接收,第一高速缓存存储器检测是否存在指示来自第一高速缓存存储器的专用存储器访问操作的源指示由第一高速缓冲存储器通过专用通信网络耦合到 第三缓存存储器未被耦合。 响应于检测到源指示的存在,与目标地址相关联的第一高速缓冲存储器的一致性状态字段被更新为第一数据无效状态。 响应于检测到不存在源指示,将第一高速缓冲存储器的一致性状态字段更新为不同的第二数据无效状态。

    Bandwidth of a cache directory by slicing the cache directory into two smaller cache directories and replicating snooping logic for each sliced cache directory
    12.
    发明申请
    Bandwidth of a cache directory by slicing the cache directory into two smaller cache directories and replicating snooping logic for each sliced cache directory 有权
    缓存目录的带宽通过将缓存目录分成两个较小的缓存目录,并为每个切片缓存目录复制侦听逻辑

    公开(公告)号:US20060184747A1

    公开(公告)日:2006-08-17

    申请号:US11056721

    申请日:2005-02-11

    IPC分类号: G06F13/28 G06F12/00

    CPC分类号: G06F12/0831 G06F12/0851

    摘要: A cache, system and method for improving the snoop bandwidth of a cache directory. A cache directory may be sliced into two smaller cache directories each with its own snooping logic. By having two cache directories that can be accessed simultaneously, the bandwidth can be essentially doubled. Furthermore, a “frequency matcher” may shift the cycle speed to a lower speed upon receiving snoop addresses from the interconnect thereby slowing down the rate at which requests are transmitted to the dispatch pipelines. Each dispatch pipeline is coupled to a sliced cache directory and is configured to search the cache directory to determine if data at the received addresses is stored in the cache memory. As a result of slowing down the rate at which requests are transmitted to the dispatch pipelines and accessing the two sliced cache directories simultaneously, the bandwidth or throughput of the cache directory may be improved.

    摘要翻译: 用于提高缓存目录的窥探带宽的缓存,系统和方法。 高速缓存目录可以分成两个较小的缓存目录,每个具有自己的侦听逻辑。 通过具有可以同时访问的两个缓存目录,带宽可以基本上加倍。 此外,当从互连接收到窥探地址时,“频率匹配器”可以将周期速度转移到较低速度,从而将请求发送到调度管线的速率减慢。 每个调度流水线被耦合到一个切片缓存目录,并被配置为搜索该高速缓存目录以确定该接收到的地址上的数据是否被存储在该高速缓冲存储器中。 作为将请求发送到调度管线的速度变慢并且同时访问两个分片缓存目录的结果,可以提高缓存目录的带宽或吞吐量。

    Data processing system, method and interconnect fabric that protect ownership transfer with a protection window extension
    13.
    发明申请
    Data processing system, method and interconnect fabric that protect ownership transfer with a protection window extension 审中-公开
    数据处理系统,方法和互连结构,通过保护窗口扩展来保护所有权转移

    公开(公告)号:US20060179253A1

    公开(公告)日:2006-08-10

    申请号:US11054841

    申请日:2005-02-10

    IPC分类号: G06F13/28

    摘要: A data processing system includes a memory system, a plurality of masters that issue requests for access to memory blocks within the memory system, a plurality of snoopers that provide partial responses to requests by the masters, and response logic that generates combined responses for the requests in response to the partial responses provided by the plurality of snoopers. The plurality masters includes a winning master that issues a request for a particular memory block, and the plurality of snoopers includes a protecting snooper that, in response to receipt of the request, provides a partial response and protects a transfer of coherency ownership of the particular memory block to the winning master until expiration of a protection window extension following receipt from the response logic of a combined response for the request.

    摘要翻译: 数据处理系统包括存储器系统,发出对存储器系统内的存储器块的访问请求的多个主器件,提供对由主器件发出的请求的部分响应的多个监视器以及产生针对请求的组合响应的响应逻辑 响应于由多个窥探者提供的部分响应。 多个主人包括发出对特定存储块的请求的获胜主,并且多个窥探者包括保护窥探者,其响应于该请求的接收而提供部分响应并保护特定存储块的一致性所有权的转移 记忆块到获胜主机,直到从请求的组合响应的响应逻辑接收到保护窗口扩展之后到期。

    Method for completing full cacheline stores with address-only bus operations
    16.
    发明申请
    Method for completing full cacheline stores with address-only bus operations 有权
    完成具有仅地址总线操作的完整缓存线存储的方法

    公开(公告)号:US20050251623A1

    公开(公告)日:2005-11-10

    申请号:US10825189

    申请日:2004-04-15

    IPC分类号: G06F12/00 G06F12/08

    CPC分类号: G06F12/0897 G06F12/0804

    摘要: A method and processor system that substantially eliminates data bus operations when completing updates of an entire cache line with a full store queue entry. The store queue within a processor chip is designed with a series of AND gates connecting individual bits of the byte enable bits of a corresponding entry. The AND output is fed to the STQ controller and signals when the entry is full. When full entries are selected for dispatch to the RC machines, the RC machine is signaled that the entry updates the entire cache line. The RC machine obtains write permission to the line, and then the RC machine overwrites the entire cache line. Because the entire cache line is overwritten, the data of the cache line is not retrieved when the request for the cache line misses at the cache or when data goes state before write permission is obtained by the RC machine.

    摘要翻译: 一种方法和处理器系统,其在完成具有完整存储队列条目的整个高速缓存行的更新时基本上消除数据总线操作。 处理器芯片内的存储队列设计有连接相应条目的字节使能位的各个位的一系列与门。 AND输出被馈送到STQ控制器,并在条目已满时发出信号。 当选择完整条目以发送到RC机器时,RC机器发出信号,表示该条目更新整个高速缓存行。 RC机器获得线路的写入权限,然后RC机器覆盖整个高速缓存行。 由于整个高速缓存线被覆盖,当缓存线的请求在高速缓存中丢失时或在RC机器获得写入许可之前数据进入状态时,不会检索高速缓存行的数据。

    Data processing system, cache system and method for scrubbing a domain indication in response to execution of program code
    17.
    发明申请
    Data processing system, cache system and method for scrubbing a domain indication in response to execution of program code 有权
    数据处理系统,缓存系统和用于响应于程序代码的执行来擦除域指示的方法

    公开(公告)号:US20060271741A1

    公开(公告)日:2006-11-30

    申请号:US11136642

    申请日:2005-05-24

    IPC分类号: G06F13/28

    CPC分类号: G06F12/0831 G06F12/0813

    摘要: In response to execution of program code, a control register within scrubbing logic in a local coherency domain is initialized with at least a target address of a target memory block. In response to the initialization, the scrubbing logic issues to at least one cache hierarchy in a remote coherency domain a domain indication scrubbing request targeting a target memory block that may be cached by the at least one cache hierarchy. In response to receipt of a coherency response indicating that the target memory block is not cached in the remote coherency domain, a domain indication in the local coherency domain is updated to indicate that the target memory block is cached, if at all, only within the local coherency domain.

    摘要翻译: 响应于程序代码的执行,用至少目标存储器块的目标地址初始化局部一致性域内的擦除逻辑中的控制寄存器。 响应于初始化,擦除逻辑向远程一致性域中的至少一个高速缓存层级发出针对可由所述至少一个高速缓存层级缓存的目标存储器块的域指示擦除请求。 响应于接收到指示目标存储器块未被缓存在远程一致性域中的一致性响应,本地一致性域中的域指示被更新以指示目标存储器块被缓存,如果完全只在 局部一致性域。

    Processor, data processing system and method for synchronzing access to data in shared memory
    18.
    发明申请
    Processor, data processing system and method for synchronzing access to data in shared memory 有权
    处理器,数据处理系统和方法,用于同步共享存储器中数据的访问

    公开(公告)号:US20060085603A1

    公开(公告)日:2006-04-20

    申请号:US10965113

    申请日:2004-10-14

    IPC分类号: G06F12/00

    摘要: A processing unit for a multiprocessor data processing system includes a store-through upper level cache, an instruction sequencing unit that fetches instructions for execution, at least one instruction execution unit that executes a store-conditional instruction to determine a store target address, a store queue that, following execution of the store-conditional instruction, buffers a corresponding store operation, sequencer logic associated with the store queue. The sequencer logic, responsive to receipt of a latency indication indicating that resolution of the store-conditional operation as passing or failing is subject to significant latency, invalidates, prior to resolution of the store-conditional operation, a cache line in the store-through upper level cache to which a load-reserve operation previously bound.

    摘要翻译: 一种用于多处理器数据处理系统的处理单元,包括存储器上级缓存器,取指令执行指令排序单元,执行存储条件指令以确定存储目标地址的至少一个指令执行单元,存储器 在存储条件指令的执行之后,缓存与存储队列相关联的对应存储操作,定序器逻辑的队列。 定序器逻辑响应于指示存储条件操作的解析作为传递或失败的等待时间指示受到重大等待时间的影响,在存储条件操作的解析之前无效,存储器中的高速缓存行 加载预备操作先前绑定到的高级缓存。

    Reducing Number of Rejected Snoop Requests By Extending Time to Respond to Snoop Request
    20.
    发明申请
    Reducing Number of Rejected Snoop Requests By Extending Time to Respond to Snoop Request 失效
    通过延长响应Snoop请求的时间来减少被拒绝的侦听请求数

    公开(公告)号:US20070294486A1

    公开(公告)日:2007-12-20

    申请号:US11847941

    申请日:2007-08-30

    IPC分类号: G06F12/00

    CPC分类号: G06F12/0831

    摘要: A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.

    摘要翻译: 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。