Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors
    1.
    发明申请
    Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors 有权
    芯片多处理器中亲和力引导的投机辅助线程的方法和装置

    公开(公告)号:US20050027941A1

    公开(公告)日:2005-02-03

    申请号:US10632431

    申请日:2003-07-31

    IPC分类号: G06F9/38 G06F12/08 G06F12/00

    摘要: Apparatus, system and methods are provided for performing speculative data prefetching in a chip multiprocessor (CMP). Data is prefetched by a helper thread that runs on one core of the CMP while a main program runs concurrently on another core of the CMP. Data prefetched by the helper thread is provided to the helper core. For one embodiment, the data prefetched by the helper thread is pushed to the main core. It may or may not be provided to the helper core as well. A push of prefetched data to the main core may occur during a broadcast of the data to all cores of an affinity group. For at least one other embodiment, the data prefetched by a helper thread is provided, upon request from the main core, to the main core from the helper core's local cache.

    摘要翻译: 提供了用于在芯片多处理器(CMP)中执行推测性数据预取的装置,系统和方法。 数据由在CMP的一个核心上运行的辅助线程预取,而主程序在CMP的另一个核心上同时运行。 由辅助线程预取的数据被提供给辅助核心。 对于一个实施例,由辅助线程预取的数据被推送到主核心。 它也可以也可以不被提供给辅助核心。 在将数据广播到亲和组的所有核心的过程中,可能会将预取数据推送到主核心。 对于至少另一个实施例,根据主核心的请求,从辅助核心的本地高速缓存提供由辅助线程预取的数据到主核心。

    Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors
    4.
    发明授权
    Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors 有权
    芯片多处理器中亲和力引导的投机辅助线程的方法和装置

    公开(公告)号:US07844801B2

    公开(公告)日:2010-11-30

    申请号:US10632431

    申请日:2003-07-31

    IPC分类号: G06F7/38 G06F9/00 G06F9/44

    摘要: Apparatus, system and methods are provided for performing speculative data prefetching in a chip multiprocessor (CMP). Data is prefetched by a helper thread that runs on one core of the CMP while a main program runs concurrently on another core of the CMP. Data prefetched by the helper thread is provided to the helper core. For one embodiment, the data prefetched by the helper thread is pushed to the main core. It may or may not be provided to the helper core as well. A push of prefetched data to the main core may occur during a broadcast of the data to all cores of an affinity group. For at least one other embodiment, the data prefetched by a helper thread is provided, upon request from the main core, to the main core from the helper core's local cache.

    摘要翻译: 提供了用于在芯片多处理器(CMP)中执行推测性数据预取的装置,系统和方法。 数据由在CMP的一个核心上运行的辅助线程预取,而主程序在CMP的另一个核心上同时运行。 由辅助线程预取的数据被提供给辅助核心。 对于一个实施例,由辅助线程预取的数据被推送到主核心。 它也可以也可以不被提供给辅助核心。 在将数据广播到亲和组的所有核心的过程中,可能会将预取数据推送到主核心。 对于至少另一个实施例,根据主核心的请求,从辅助核心的本地高速缓存提供由辅助线程预取的数据到主核心。

    METHOD AND APPARATUS FOR AFFINITY-GUIDED SPECULATIVE HELPER THREADS IN CHIP MULTIPROCESSORS
    5.
    发明申请
    METHOD AND APPARATUS FOR AFFINITY-GUIDED SPECULATIVE HELPER THREADS IN CHIP MULTIPROCESSORS 有权
    芯片多路由器中辅助引导的辅助线路的方法和装置

    公开(公告)号:US20110035555A1

    公开(公告)日:2011-02-10

    申请号:US12909774

    申请日:2010-10-21

    摘要: Apparatus, system and methods are provided for performing speculative data prefetching in a chip multiprocessor (CMP). Data is prefetched by a helper thread that runs on one core of the CMP while a main program runs concurrently on another core of the CMP. Data prefetched by the helper thread is provided to the helper core. For one embodiment, the data prefetched by the helper thread is pushed to the main core. It may or may not be provided to the helper core as well. A push of prefetched data to the main core may occur during a broadcast of the data to all cores of an affinity group. For at least one other embodiment, the data prefetched by a helper thread is provided, upon request from the main core, to the main core from the helper core's local cache.

    摘要翻译: 提供了用于在芯片多处理器(CMP)中执行推测性数据预取的装置,系统和方法。 数据由在CMP的一个核心上运行的辅助线程预取,而主程序在CMP的另一个核心上同时运行。 由辅助线程预取的数据被提供给辅助核心。 对于一个实施例,由辅助线程预取的数据被推送到主核心。 它也可以也可以不被提供给辅助核心。 在将数据广播到亲和组的所有核心的过程中,可能会将预取数据推送到主核心。 对于至少另一个实施例,根据主核心的请求,从辅助核心的本地高速缓存提供由辅助线程预取的数据到主核心。

    Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors
    6.
    发明授权
    Method and apparatus for affinity-guided speculative helper threads in chip multiprocessors 有权
    芯片多处理器中亲和力引导的投机辅助线程的方法和装置

    公开(公告)号:US08078831B2

    公开(公告)日:2011-12-13

    申请号:US12909774

    申请日:2010-10-21

    IPC分类号: G06F15/00 G06F15/76

    摘要: Apparatus, system and methods are provided for performing speculative data prefetching in a chip multiprocessor (CMP). Data is prefetched by a helper thread that runs on one core of the CMP while a main program runs concurrently on another core of the CMP. Data prefetched by the helper thread is provided to the helper core. For one embodiment, the data prefetched by the helper thread is pushed to the main core. It may or may not be provided to the helper core as well. A push of prefetched data to the main core may occur during a broadcast of the data to all cores of an affinity group. For at least one other embodiment, the data prefetched by a helper thread is provided, upon request from the main core, to the main core from the helper core's local cache.

    摘要翻译: 提供了用于在芯片多处理器(CMP)中执行推测性数据预取的装置,系统和方法。 数据由在CMP的一个核心上运行的辅助线程预取,而主程序在CMP的另一个核心上同时运行。 由辅助线程预取的数据被提供给辅助核心。 对于一个实施例,由辅助线程预取的数据被推送到主核心。 它也可以也可以不被提供给辅助核心。 在将数据广播到亲和组的所有核心的过程中,可能会将预取数据推送到主核心。 对于至少另一个实施例,根据主核心的请求,从辅助核心的本地高速缓存提供由辅助线程预取的数据到主核心。

    UNPACKING PACKED DATA IN MULTIPLE LANES
    10.
    发明申请
    UNPACKING PACKED DATA IN MULTIPLE LANES 有权
    在多个土地上打包包装数据

    公开(公告)号:US20100332794A1

    公开(公告)日:2010-12-30

    申请号:US12494667

    申请日:2009-06-30

    IPC分类号: G06F15/76

    摘要: Receiving an instruction indicating first and second operands. Each of the operands having packed data elements that correspond in respective positions. A first subset of the data elements of the first operand and a first subset of the data elements of the second operand each corresponding to a first lane. A second subset of the data elements of the first operand and a second subset of the data elements of the second operand each corresponding to a second lane. Storing result, in response to instruction, including: (1) in first lane, only lowest order data elements from first subset of first operand interleaved with corresponding lowest order data elements from first subset of second operand; and (2) in second lane, only highest order data elements from second subset of first operand interleaved with corresponding highest order data elements from second subset of second operand.

    摘要翻译: 接收指示第一和第二操作数的指令。 每个操作数具有在相应位置对应的数据元素的打包。 第一操作数的数据元素的第一子集和第二操作数的数据元素的第一子集,每个对应于第一通道。 所述第一操作数的数据元素的第二子集和所述第二操作数的数据元素的第二子集分别对应于第二通道。 存储结果,响应于指令,包括:(1)在第一通道中,仅来自第一操作数的第一子集的最低顺序数据元素与来自第二操作数的第一子集的相应最低数据元素交织; 和(2)在第二通道中,仅来自第一操作数的第二子集的最高阶数据元素与来自第二操作数的第二子集的对应最高阶数据元素交织。