Aggregate Symmetric Multiprocessor System
    131.
    发明申请
    Aggregate Symmetric Multiprocessor System 失效
    聚合对称多处理器系统

    公开(公告)号:US20110153936A1

    公开(公告)日:2011-06-23

    申请号:US12643716

    申请日:2009-12-21

    申请人: William J. Starke

    发明人: William J. Starke

    IPC分类号: G06F15/80 G06F12/08 G06F9/02

    摘要: An aggregate symmetric multiprocessor (SMP) data processing system includes a first SMP computer including at least first and second processing units and a first system memory pool and a second SMP computer including at least third and fourth processing units and second and third system memory pools. The second system memory pool is a restricted access memory pool inaccessible to the fourth processing unit and accessible to at least the second and third processing units, and the third system memory pool is accessible to both the third and fourth processing units. An interconnect couples the second processing unit in the first SMP computer for load-store coherent, ordered access to the second system memory pool in the second SMP computer, such that the second processing unit in the first SMP computer and the second system memory pool in the second SMP computer form a synthetic third SMP computer.

    摘要翻译: 聚合对称多处理器(SMP)数据处理系统包括包括至少第一和第二处理单元的第一SMP计算机和包括至少第三和第四处理单元以及第二和第三系统存储器池的第一系统存储器池和第二SMP计算机。 第二系统存储器池是第四处理单元不可访问的受限访问存储器池,并且可由至少第二和第三处理单元访问,并且第三系统存储池可由第三处理单元和第四处理单元访问。 互连耦合第一SMP计算机中的第二处理单元,用于对第二SMP计算机中的第二系统存储池进行加载存储一致的有序访问,使得第一SMP计算机中的第二处理单元和第二系统存储器池 第二台SMP计算机形成了合成的第三台SMP计算机。

    Updating Partial Cache Lines in a Data Processing System
    132.
    发明申请
    Updating Partial Cache Lines in a Data Processing System 有权
    更新数据处理系统中的部分缓存行

    公开(公告)号:US20100268884A1

    公开(公告)日:2010-10-21

    申请号:US12424434

    申请日:2009-04-15

    IPC分类号: G06F12/08 G06F12/00

    摘要: A processing unit for a data processing system includes a processor core having one or more execution units for processing instructions and a register file for storing data accessed in processing of the instructions. The processing unit also includes a multi-level cache hierarchy coupled to and supporting the processor core. The multi-level cache hierarchy includes at least one upper level of cache memory having a lower access latency and at least one lower level of cache memory having a higher access latency. The lower level of cache memory, responsive to receipt of a memory access request that hits only a partial cache line in the lower level cache memory, sources the partial cache line to the at least one upper level cache memory to service the memory access request. The at least one upper level cache memory services the memory access request without caching the partial cache line.

    摘要翻译: 用于数据处理系统的处理单元包括具有一个或多个用于处理指令的执行单元的处理器核心和用于存储在指令处理中访问的数据的寄存器文件。 处理单元还包括耦合到并支持处理器核的多级高速缓存层级。 多级高速缓存层级包括具有较低访问延迟的至少一个高级缓存存储器和具有较高访问延迟的至少一个较低级别的高速缓存存储器。 响应于仅接收低级高速缓冲存储器中的部分高速缓存行的存储器访问请求的响应,较低级别的高速缓存存储器将部分高速缓存行源送到至少一个上级高速缓冲存储器以服务存储器访问请求。 至少一个上级缓存存储器服务于存储器访问请求,而不缓存部分高速缓存行。

    Reducing number of rejected snoop requests by extending time to respond to snoop request
    133.
    发明授权
    Reducing number of rejected snoop requests by extending time to respond to snoop request 失效
    通过延长响应窥探请求的时间来减少被拒绝的窥探请求数

    公开(公告)号:US07818511B2

    公开(公告)日:2010-10-19

    申请号:US11847941

    申请日:2007-08-30

    IPC分类号: G06F12/00

    CPC分类号: G06F12/0831

    摘要: A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.

    摘要翻译: 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。

    Mode-Based Castout Destination Selection
    134.
    发明申请
    Mode-Based Castout Destination Selection 失效
    基于模式的Castout目的地选择

    公开(公告)号:US20100262783A1

    公开(公告)日:2010-10-14

    申请号:US12420933

    申请日:2009-04-09

    IPC分类号: G06F12/08

    CPC分类号: G06F12/0811 G06F12/12

    摘要: In response to a data request of a first of a plurality of processing units, the first processing unit selects a victim cache line to be castout from the lower level cache of the first processing unit and determines whether a mode is set. If not, the first processing unit issues on the interconnect fabric an LCO command identifying the victim cache line and indicating that a lower level cache is the intended destination. If the mode is set, the first processing unit issues a castout command with an alternative intended destination. In response to a coherence response to the LCO command indicating success of the LCO command, the first processing unit removes the victim cache line from its lower level cache, and the victim cache line is held elsewhere in the data processing system. The mode can be set to inhibit castouts to system memory, for example, for testing.

    摘要翻译: 响应于多个处理单元中的第一处理单元的数据请求,第一处理单元从第一处理单元的较低级高速缓存中选择要丢弃的牺牲高速缓存行,并且确定是否设置了模式。 如果不是,则第一处理单元在互连结构上发出识别受害者高速缓存行的LCO命令,并指示较低级别的高速缓存是预期的目的地。 如果模式被设置,则第一处理单元发出具有替代预定目的地的停顿命令。 响应于指示LCO命令成功的LCO命令的一致性响应,第一处理单元从其较低级高速缓存中去除受害者高速缓存行,并且将受害者高速缓存行保持在数据处理系统的其他地方。 该模式可以设置为抑制系统内存的丢弃,例如进行测试。

    Updating an invalid coherency state in response to snooping an operation
    135.
    发明授权
    Updating an invalid coherency state in response to snooping an operation 失效
    更新无效的一致性状态以响应窥探操作

    公开(公告)号:US07743218B2

    公开(公告)日:2010-06-22

    申请号:US12190766

    申请日:2008-08-13

    IPC分类号: G06F12/00

    CPC分类号: G06F12/0831 G06F2212/507

    摘要: A cache coherent data processing system includes at least first and second coherency domains. In a first cache memory within the first coherency domain of the data processing system, a coherency state field associated with a storage location and an address tag is set to a first data-invalid coherency state that indicates that the address tag is valid and that the storage location does not contain valid data. In response to snooping an exclusive access operation, the exclusive access request specifying a target address matching the address tag and indicating a relative domain location of a requestor that initiated the exclusive access operation, the first cache memory updates the coherency state field from the first data-invalid coherency state to a second data-invalid coherency state that indicates that the address tag is valid, that the storage location does not contain valid data, and whether a target memory block associated with the address tag is cached within the first coherency domain upon successful completion of the exclusive access operation based upon the relative location of the requestor.

    摘要翻译: 缓存相干数据处理系统至少包括第一和第二相干域。 在数据处理系统的第一相关域内的第一高速缓冲存储器中,将与存储位置和地址标签相关联的一致性状态字段设置为指示地址标签有效的第一数据无效一致性状态, 存储位置不包含有效数据。 响应于窥探独占访问操作,所述专用访问请求指定与所述地址标签相匹配的目标地址并且指示发起所述独占访问操作的请求者的相对域位置,所述第一高速缓存存储器从所述第一数据更新所述一致性状态字段 - 无效的一致性状态到指示地址标签有效的第二数据无效一致性状态,存储位置不包含有效数据,以及与地址标签相关联的目标存储器块是否被缓存在第一相关域内 基于请求者的相对位置成功完成独占访问操作。

    Victim Cache Prefetching
    136.
    发明申请
    Victim Cache Prefetching 失效
    受害者缓存预取

    公开(公告)号:US20100100683A1

    公开(公告)日:2010-04-22

    申请号:US12256064

    申请日:2008-10-22

    IPC分类号: G06F12/08

    摘要: A processing unit for a multiprocessor data processing system includes a processor core and a cache hierarchy coupled to the processor core to provide low latency data access. The cache hierarchy includes an upper level cache coupled to the processor core and a lower level victim cache coupled to the upper level cache. In response to a prefetch request of the processor core that misses in the upper level cache, the lower level victim cache determines whether the prefetch request misses in the directory of the lower level victim cache and, if so, allocates a state machine in the lower level victim cache that services the prefetch request by issuing the prefetch request to at least one other processing unit of the multiprocessor data processing system.

    摘要翻译: 用于多处理器数据处理系统的处理单元包括处理器核心和耦合到处理器核心的高速缓存层级以提供低延迟数据访问。 高速缓存层级包括耦合到处理器核心的高级缓存和耦合到高级缓存的较低级别的牺牲缓存。 响应于在高级缓存中丢失的处理器核心的预取请求,较低级别的受害者缓存确定预取请求是否丢失在较低级别的受害者缓存的目录中,并且如果是,则在下级缓存中分配状态机 通过向多处理器数据处理系统的至少一个其他处理单元发出预取请求来服务于预取请求。

    Victim Cache Replacement
    137.
    发明申请
    Victim Cache Replacement 有权
    受害者缓存替换

    公开(公告)号:US20100023695A1

    公开(公告)日:2010-01-28

    申请号:US12177912

    申请日:2008-07-23

    IPC分类号: G06F12/08

    摘要: A data processing system includes a processor core having an associated upper level cache and a lower level victim cache. In response to a memory access request of the processor core, the lower level cache victim determines whether the memory access request hits or misses in the directory of the lower level victim cache, and the upper level cache determines whether a castout from the upper level cache is to be performed and selects a victim coherency granule for eviction from the upper level cache. In response to determining that a castout from the upper level cache is to be performed, the upper level cache evicts the selected victim coherency granule. In the eviction, the upper level cache reads out the victim coherency granule from the data array of the upper level cache only in response to an indication that the memory access request misses in the directory of the lower level victim cache.

    摘要翻译: 数据处理系统包括具有相关联的高级缓存和较低级别的受害缓存的处理器核心。 响应于处理器核心的存储器访问请求,下级缓存受害者确定存储器访问请求是否在较低级别的受害者高速缓存的目录中命中或丢失,并且上级缓存确定来自上级缓存的丢弃 将被执行,并从上级缓存中选择被驱逐的受害者一致性粒子。 响应于确定要执行来自上级高速缓存的停顿,上级高速缓存驱逐所选择的受害者一致性粒子。 在逐出时,高级缓存器只有在响应于存储器访问请求在较低级别的受害者缓存的目录中丢失的指示时才从高级缓存的数据阵列读出受害者一致性粒子。

    Reducing number of rejected snoop requests by extending time to respond to snoop request
    139.
    发明授权
    Reducing number of rejected snoop requests by extending time to respond to snoop request 失效
    通过延长响应窥探请求的时间来减少被拒绝的窥探请求数

    公开(公告)号:US07484046B2

    公开(公告)日:2009-01-27

    申请号:US11950717

    申请日:2007-12-05

    IPC分类号: G06F12/00

    CPC分类号: G06F12/0831

    摘要: A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.

    摘要翻译: 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。

    Data processing system, cache system and method for passively scrubbing a domain indication
    140.
    发明授权
    Data processing system, cache system and method for passively scrubbing a domain indication 失效
    数据处理系统,缓存系统和被动清理域指示的方法

    公开(公告)号:US07478201B2

    公开(公告)日:2009-01-13

    申请号:US11136652

    申请日:2005-05-24

    IPC分类号: G06F12/00

    CPC分类号: G06F12/0831

    摘要: Scrubbing logic in a local coherency domain issues a domain query request to at least one cache hierarchy in a remote coherency domain. The domain query request is a non-destructive probe of a coherency state associated with a target memory block by the at least one cache hierarchy. A coherency response to the domain query request is received. In response to the coherency response indicating that the target memory block is not cached in the remote coherency domain, a domain indication in the local coherency domain is updated to indicate that the target memory block is cached, if at all, only within the local coherency domain.

    摘要翻译: 本地一致性域中的擦除逻辑向远程一致性域中的至少一个缓存层次结构发出域查询请求。 域查询请求是由至少一个高速缓存层次结构与目标存储器块相关联的一致性状态的非破坏性探测。 接收到域查询请求的一致性响应。 响应于指示目标存储器块未被缓存在远程一致性域中的相关性响应,本地一致性域中的域指示被更新以指示目标存储器块被缓存,如果完全只在本地一致性内 域。