Virtual Barrier Synchronization Cache
    11.
    发明申请
    Virtual Barrier Synchronization Cache 失效
    虚拟障碍同步缓存

    公开(公告)号:US20100257317A1

    公开(公告)日:2010-10-07

    申请号:US12419364

    申请日:2009-04-07

    IPC分类号: G06F12/08 G06F12/00

    CPC分类号: G06F12/0811 G06F9/522

    摘要: A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region of the system memory. Each of the plurality of processing units includes a processor core and a cache memory including a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory and a cache controller. The cache controller, responsive to a store request from the processor core to update a particular VBSR line, performs a non-blocking update of the cache array in each other of the plurality of processing units contemporaneously holding a copy of the particular VBSR line by transmitting a VBSR update command on the interconnect fabric.

    摘要翻译: 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问系统内存的虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括从系统存储器的虚拟屏障同步区域缓存VBSR行的缓存阵列和高速缓存控制器。 高速缓存控制器响应于来自处理器核心的存储请求来更新特定VBSR线路,通过发送来同时保存特定VBSR线路的副本的多个处理单元中的彼此之间的高速缓存阵列的非阻塞更新 互连结构上的VBSR更新命令。

    Empirically based dynamic control of transmission of victim cache lateral castouts
    12.
    发明授权
    Empirically based dynamic control of transmission of victim cache lateral castouts 有权
    基于经验的动态控制受害者缓存横向传播的传输

    公开(公告)号:US08347036B2

    公开(公告)日:2013-01-01

    申请号:US12421180

    申请日:2009-04-09

    IPC分类号: G06F12/08

    摘要: In response to a data request, a victim cache line is selected for castout from a lower level cache, and a target lower level cache of one of the plurality of processing units is selected. A determination is made whether the selected target lower level cache has provided more than a threshold number of retry responses to lateral castout (LCO) commands of the first lower level cache, and if so, a different target lower level cache is selected. The first processing unit thereafter issues a LCO command on the interconnect fabric. The LCO command identifies the victim cache line to be castout and indicates that the target lower level cache is an intended destination of the victim cache line. In response to a successful coherence response to the LCO command, the victim cache line is removed from the first lower level cache and held in the second lower level cache.

    摘要翻译: 响应于数据请求,选择从较低级别高速缓冲存储器进行丢弃的受害者高速缓存行,并且选择多个处理单元之一的目标下级高速缓存。 确定所选择的目标下层高速缓存是否为第一较低级别高速缓存的横向转移(LCO)命令提供了超过阈值数量的重试响应,如果是,则选择不同的目标低级高速缓存。 此后,第一处理单元在互连结构上发出LCO命令。 LCO命令标识要丢弃的受害者缓存行,并指示目标下级缓存是受害缓存行的预期目标。 响应于对LCO命令的成功的一致性响应,从第一低级缓存中移除受害者高速缓存行并保存在第二较低级缓存中。

    Load request scheduling in a cache hierarchy
    13.
    发明授权
    Load request scheduling in a cache hierarchy 有权
    在缓存层次结构中加载请求调度

    公开(公告)号:US08521982B2

    公开(公告)日:2013-08-27

    申请号:US12424207

    申请日:2009-04-15

    IPC分类号: G06F12/00

    摘要: A system and method for tracking core load requests and providing arbitration and ordering of requests. When a core interface unit (CIU) receives a load operation from the processor core, a new entry in allocated in a queue of the CIU. In response to allocating the new entry in the queue, the CIU detects contention between the load request and another memory access request. In response to detecting contention, the load request may be suspended until the contention is resolved. Received load requests may be stored in the queue and tracked using a least recently used (LRU) mechanism. The load request may then be processed when the load request resides in a least recently used entry in the load request queue. CIU may also suspend issuing an instruction unless a read claim (RC) machine is available. In another embodiment, CIU may issue stored load requests in a specific priority order.

    摘要翻译: 用于跟踪核心负载请求并提供仲裁和请求排序的系统和方法。 当核心接口单元(CIU)从处理器核心接收到加载操作时,分配在CIU队列中的新条目。 响应于在队列中分配新条目,CIU检测加载请求和另一个存储器访问请求之间的争用。 响应于检测到争用,负载请求可以被暂停,直到争用被解决。 接收到的加载请求可以存储在队列中,并使用最近最少使用的(LRU)机制进行跟踪。 然后可以在加载请求驻留在加载请求队列中最近最少使用的条目中时处理加载请求。 除非读取权利要求(RC)机器可用,否则CIU也可以暂停发出指令。 在另一个实施例中,CIU可以以特定优先级顺序发布存储的加载请求。

    Data processing system, processor and method of data processing that reduce store queue entry utilization for synchronizing operations
    14.
    发明授权
    Data processing system, processor and method of data processing that reduce store queue entry utilization for synchronizing operations 失效
    数据处理系统,处理器和数据处理方法,减少存储队列入口利用率,用于同步操作

    公开(公告)号:US07454580B2

    公开(公告)日:2008-11-18

    申请号:US11380020

    申请日:2006-04-25

    摘要: A data processing system includes a processor core and a memory subsystem. The memory subsystem includes a store queue having a plurality of entries, where each entry includes an address field for holding the target address of store operation, a data field for holding data for the store operation, and a virtual sync field indicating a presence or absence of a synchronizing operation associated with the entry. The memory subsystem further includes a store queue controller that, responsive to receipt at the memory subsystem of a sequence of operations including a synchronizing operation and a particular store operation, places a target address and data of the particular store operation within the address field and data field, respectively, of an entry in the store queue and sets the virtual sync field of the entry to represent the synchronizing operation, such that a number of store queue entries utilized is reduced.

    摘要翻译: 数据处理系统包括处理器核心和存储器子系统。 存储器子系统包括具有多个条目的存储队列,其中每个条目包括用于保存存储操作的目标地址的地址字段,用于保存用于存储操作的数据的数据字段和指示存在或不存在的虚拟同步字段 与该条目相关联的同步操作。 存储器子系统还包括存储队列控制器,其响应于在存储器子系统处的接收包括同步操作和特定存储操作的一系列操作,将特定存储操作的目标地址和数据放置在地址字段和数据中 字段,并且设置条目的虚拟同步字段以表示同步操作,使得减少使用的存储队列条目的数量。