Selective cache-to-cache lateral castouts
    72.
    发明授权
    Selective cache-to-cache lateral castouts 有权
    选择性高速缓存到缓存横向转义

    公开(公告)号:US09189403B2

    公开(公告)日:2015-11-17

    申请号:US12650018

    申请日:2009-12-30

    IPC分类号: G06F12/00 G06F12/08 G06F12/12

    CPC分类号: G06F12/0811 G06F12/12

    摘要: A data processing system includes first and second processing units and a system memory. The first processing unit has first upper and first lower level caches, and the second processing unit has second upper and lower level caches. In response to a data request, a victim cache line to be castout from the first lower level cache is selected, and the first lower level cache selects between performing a lateral castout (LCO) of the victim cache line to the second lower level cache and a castout of the victim cache line to the system memory based upon a confidence indicator associated with the victim cache line. In response to selecting an LCO, the first processing unit issues an LCO command on the interconnect fabric and removes the victim cache line from the first lower level cache, and the second lower level cache holds the victim cache line.

    摘要翻译: 数据处理系统包括第一和第二处理单元和系统存储器。 第一处理单元具有第一上层和第一下层高速缓存,第二处理单元具有第二上层和下层高速缓存。 响应于数据请求,选择要从第一较低级高速缓存丢弃的受害者高速缓存行,并且第一较低级高速缓存选择在执行到第二低级高速缓存的受害者高速缓存行的横向流出(LCO) 基于与受害者高速缓存行相关联的置信指示,将受害者缓存行的丢弃发送到系统存储器。 响应于选择LCO,第一处理单元在互连结构上发布LCO命令,并从第一低级缓存中移除受害者高速缓存行,并且第二下级缓存保存受害缓存行。

    Binding a process to a special purpose processing element having characteristics of a processor
    73.
    发明授权
    Binding a process to a special purpose processing element having characteristics of a processor 有权
    将过程绑定到具有处理器特征的专用处理元件

    公开(公告)号:US08893126B2

    公开(公告)日:2014-11-18

    申请号:US12024220

    申请日:2008-02-01

    IPC分类号: G06F9/00 G06F13/12

    CPC分类号: G06F13/12

    摘要: A heterogeneous processing element model is provided where I/O devices look and act like processors. In order to be treated like a processor, an I/O processing element, or other special purpose processing element, must follow some rules and have some characteristics of a processor, such as address translation, security, interrupt handling, and exception processing, for example. The heterogeneous processing element model puts special purpose processing elements on the same playing field as processors, from a programming perspective, operating system perspective, and power perspective. The operating system can get work to a security engine, for example, in the same way it does to a processor.

    摘要翻译: 提供异构处理元件模型,其中I / O设备看起来像处理器一样操作。 为了像处理器一样处理I / O处理元件或其他专用处理元件,必须遵循一些规则并具有处理器的某些特性,例如地址转换,安全性,中断处理和异常处理,用于 例。 异构处理元素模型将特殊处理元素与编程角度,操作系统角度和功能视角相结合,将处理器与处理器相同。 操作系统可以使用安全引擎,例如,与处理器相同。

    Data cache block deallocate requests in a multi-level cache hierarchy
    74.
    发明授权
    Data cache block deallocate requests in a multi-level cache hierarchy 有权
    数据缓存块在多级缓存层次结构中释放请求

    公开(公告)号:US08874852B2

    公开(公告)日:2014-10-28

    申请号:US13433048

    申请日:2012-03-28

    IPC分类号: G06F12/08

    摘要: In response to executing a deallocate instruction, a deallocation request specifying a target address of a target cache line is sent from a processor core to a lower level cache. In response, a determination is made if the target address hits in the lower level cache. If so, the target cache line is retained in a data array of the lower level cache, and a replacement order field of the lower level cache is updated such that the target cache line is more likely to be evicted in response to a subsequent cache miss in a congruence class including the target cache line. In response to the subsequent cache miss, the target cache line is cast out to the lower level cache with an indication that the target cache line was a target of a previous deallocation request of the processor core.

    摘要翻译: 响应于执行取消分配指令,将指定目标高速缓存行的目标地址的解除分配请求从处理器核发送到较低级高速缓存。 作为响应,确定目标地址是否在较低级别高速缓存中。 如果是这样,则目标高速缓存行被保留在较低级高速缓存的数据阵列中,并且更新较低级高速缓存的替换顺序字段,使得目标高速缓存行更可能响应于后续高速缓存未命中而被驱逐 在包含目标缓存行的同余类中。 响应于随后的高速缓存未命中,目标高速缓存行被推出到较低级缓存,指示目标高速缓存行是处理器核心的先前释放请求的目标。

    Coordinated writeback of dirty cachelines
    75.
    发明授权
    Coordinated writeback of dirty cachelines 有权
    脏缓存行的协调回写

    公开(公告)号:US08838901B2

    公开(公告)日:2014-09-16

    申请号:US12775510

    申请日:2010-05-07

    IPC分类号: G06F12/00 G06F12/08

    摘要: A data processing system includes a processor core and a cache memory hierarchy coupled to the processor core. The cache memory hierarchy includes at least one upper level cache and a lowest level cache. A memory controller is coupled to the lowest level cache and to a system memory and includes a physical write queue from which the memory controller writes data to the system memory. The memory controller initiates accesses to the lowest level cache to place into the physical write queue selected cachelines having spatial locality with data present in the physical write queue.

    摘要翻译: 数据处理系统包括处理器核心和耦合到处理器核心的高速缓存存储器层级。 高速缓存存储器层级包括至少一个上级高速缓存和最低级高速缓存。 存储器控制器耦合到最低级缓存和系统存储器,并且包括物理写队列,存储器控制器从该物理写队列将数据写入系统存储器。 存储器控制器启动对最低级高速缓存的访问以放置到物理写入队列中,所选择的高速缓存线具有空间局部性,其中数据存在于物理写入队列中。

    Aggregate data processing system having multiple overlapping synthetic computers
    76.
    发明授权
    Aggregate data processing system having multiple overlapping synthetic computers 有权
    具有多个重叠合成计算机的综合数据处理系统

    公开(公告)号:US08656128B2

    公开(公告)日:2014-02-18

    申请号:US13599856

    申请日:2012-08-30

    IPC分类号: G06F12/00

    摘要: A first SMP computer has first and second processing units and a first system memory pool, a second SMP computer has third and fourth processing units and a second system memory pool, and a third SMP computer has at least fifth and sixth processing units and third, fourth and fifth system memory pools. The fourth system memory pool is inaccessible to the third, fourth and sixth processing units and accessible to at least the second and fifth processing units, and the fifth system memory pool is inaccessible to the first, second and sixth processing units and accessible to at least the fourth and fifth processing units. A first interconnect couples the second processing unit for load-store coherent, ordered access to the fourth system memory pool, and a second interconnect couples the fourth processing unit for load-store coherent, ordered access to the fifth system memory pool.

    摘要翻译: 第一SMP计算机具有第一和第二处理单元和第一系统存储器池,第二SMP计算机具有第三和第四处理单元和第二系统存储器池,并且第三SMP计算机具有至少第五和第六处理单元,第三SMP计算机具有至少第五和第六处理单元, 第四和第五系统内存池。 第四系统存储器池对于第三,第四和第六处理单元是不可访问的,并且可访问至少第二和第五处理单元,并且第五系统存储器池对于第一,第二和第六处理单元是不可访问的,并且至少可访问 第四和第五处理单元。 第一互连耦合第二处理单元,用于对第四系统存储池进行加载存储相关的有序访问,并且第二互连耦合第四处理单元,用于加载存储相关的有序访问到第五系统存储池。

    DATA CACHE BLOCK DEALLOCATE REQUESTS
    77.
    发明申请
    DATA CACHE BLOCK DEALLOCATE REQUESTS 有权
    数据缓存块解析请求

    公开(公告)号:US20130262777A1

    公开(公告)日:2013-10-03

    申请号:US13433022

    申请日:2012-03-28

    IPC分类号: G06F12/12

    摘要: A data processing system includes a processor core supported by upper and lower level caches. In response to executing a deallocate instruction in the processor core, a deallocation request is sent from the processor core to the lower level cache, the deallocation request specifying a target address associated with a target cache line. In response to receipt of the deallocation request at the lower level cache, a determination is made if the target address hits in the lower level cache. In response to determining that the target address hits in the lower level cache, the target cache line is retained in a data array of the lower level cache and a replacement order field in a directory of the lower level cache is updated such that the target cache line is more likely to be evicted from the lower level cache in response to a subsequent cache miss.

    摘要翻译: 数据处理系统包括由上层和下层高速缓存支持的处理器核心。 响应于在处理器核心中执行取消分配指令,从处理器核心向下级高速缓存发送解除分配请求,所述释放请求指定与目标高速缓存行相关联的目标地址。 响应于在较低级别高速缓存处接收到解除分配请求,确定目标地址是否在较低级别高速缓存中。 为了响应于确定目标地址在较低级别高速缓存中的命中,目标高速缓存行被保留在较低级别高速缓存的数据阵列中,并且更新下级高速缓存的目录中的替换顺序字段,使得目标高速缓存 线路可能会响应于后续的高速缓存未命中而从较低级别的缓存中逐出。

    Memory coherence directory supporting remotely sourced requests of nodal scope
    78.
    发明授权
    Memory coherence directory supporting remotely sourced requests of nodal scope 失效
    内存一致性目录支持远程请求节点范围

    公开(公告)号:US08504779B2

    公开(公告)日:2013-08-06

    申请号:US13445010

    申请日:2012-04-12

    IPC分类号: G06F13/00 G06F13/28

    CPC分类号: G06F12/0817

    摘要: A data processing system includes at least a first through third processing nodes coupled by an interconnect fabric. The first processing node includes a master, a plurality of snoopers capable of participating in interconnect operations, and a node interface that receives a request of the master and transmits the request of the master to the second processing unit with a nodal scope of transmission limited to the second processing node. The second processing node includes a node interface having a directory. The node interface of the second processing node permits the request to proceed with the nodal scope of transmission if the directory does not indicate that a target memory block of the request is cached other than in the second processing node and prevents the request from succeeding if the directory indicates that the target memory block of the request is cached other than in the second processing node.

    摘要翻译: 数据处理系统至少包括通过互连结构耦合的第一至第三处理节点。 第一处理节点包括主机,能够参与互连操作的多个侦听器,以及接收主机请求的节点接口,并将主机的请求传送到第二处理单元,传送范围限于 第二处理节点。 第二处理节点包括具有目录的节点接口。 第二处理节点的节点接口允许请求继续进行节点传输范围,如果该目录没有指示该请求的目标存储器块不是在第二处理节点中被缓存,并且如果该请求成功 目录指示除第二处理节点之外的请求的目标存储块被缓存。

    Handling castout cache lines in a victim cache
    79.
    发明授权
    Handling castout cache lines in a victim cache 失效
    在受害者缓存中处理castout缓存行

    公开(公告)号:US08499124B2

    公开(公告)日:2013-07-30

    申请号:US12336048

    申请日:2008-12-16

    IPC分类号: G06F12/00

    摘要: A victim cache memory includes a cache array, a cache directory of contents of the cache array, and a cache controller that controls operation of the victim cache memory. The cache controller, responsive to receiving a castout command identifying a victim cache line castout from another cache memory, causes the victim cache line to be held in the cache array. If the other cache memory is a higher level cache in the cache hierarchy of the processor core, the cache controller marks the victim cache line in the cache directory so that it is less likely to be evicted by a replacement policy of the victim cache, and otherwise, marks the victim cache line in the cache directory so that it is more likely to be evicted by the replacement policy of the victim cache.

    摘要翻译: 受害者缓存存储器包括缓存阵列,高速缓存阵列的内容的高速缓存目录以及控制受害者缓存存储器的操作的高速缓存控制器。 高速缓存控制器响应于从另一高速缓冲存储器接收识别受害者高速缓存线路突发的丢弃命令,使受害者高速缓存行保持在高速缓存阵列中。 如果其他高速缓冲存储器是处理器核心的高速缓存层级中的较高级缓存,则高速缓存控制器将高速缓存目录中的受害者高速缓存行标记为不太可能被受害缓存的替换策略驱逐, 否则,将缓存目录中的受害者缓存行标记为更有可能被受害者缓存的替换策略驱逐。

    Host fabric interface (HFI) to perform global shared memory (GSM) operations
    80.
    发明授权
    Host fabric interface (HFI) to perform global shared memory (GSM) operations 失效
    主机结构接口(HFI)执行全局共享内存(GSM)操作

    公开(公告)号:US08484307B2

    公开(公告)日:2013-07-09

    申请号:US12024397

    申请日:2008-02-01

    CPC分类号: G06F12/109 G06F9/544

    摘要: A data processing system enables global shared memory (GSM) operations across multiple nodes with a distributed EA-to-RA mapping of physical memory. Each node has a host fabric interface (HFI), which includes HFI windows that are assigned to at most one locally-executing task of a parallel job. The tasks perform parallel job execution, but map only a portion of the effective addresses (EAs) of the global address space to the local, real memory of the task's respective node. The HFI window tags all outgoing GSM operations (of the local task) with the job ID, and embeds the target node and HFI window IDs of the node at which the EA is memory mapped. The HFI window also enables processing of received GSM operations with valid EAs that are homed to the local real memory of the receiving node, while preventing processing of other received operations without a valid EA-to-RA local mapping.

    摘要翻译: 数据处理系统通过物理内存的分布式EA-to-RA映射实现跨多个节点的全局共享存储(GSM)操作。 每个节点都有一个主机结构接口(HFI),它包括分配给并行作业最多一个本地执行任务的HFI窗口。 任务执行并行作业执行,但将全局地址空间的有效地址(EA)的一部分映射到任务相应节点的本地实际存储器。 HFI窗口使用作业ID对所有传出的GSM操作(本地任务)进行标记,并嵌入EA被映射到的节点的目标节点和HFI窗口ID。 HFI窗口还能够利用归属于接收节点的本地实际存储器的有效EA来处理接收的GSM操作,同时防止在没有有效的EA到RA本地映射的情况下处理其他接收到的操作。