Generating and Issuing Global Shared Memory Operations Via a Send FIFO
    11.
    发明申请
    Generating and Issuing Global Shared Memory Operations Via a Send FIFO 有权
    通过发送FIFO生成和发出全局共享内存操作

    公开(公告)号:US20090199195A1

    公开(公告)日:2009-08-06

    申请号:US12024664

    申请日:2008-02-01

    CPC classification number: G06F9/544

    Abstract: A method for issuing global shared memory (GSM) operations from an originating task on a first node coupled to a network fabric of a distributed network via a host fabric interface (HFI). The originating task generates a GSM command within an effective address (EA) space. The task then places the GSM command within a send FIFO. The send FIFO is a portion of real memory having real addresses (RA) that are memory mapped to EAs of a globally executing job. The originating task maintains a local EA-to-RA mapping of only a portion of the real address space of the globally executing job. The task enables the HFI to retrieve the GSM command from the send FIFO into an HFI window allocated to the originating task. The HFI window generates a corresponding GSM packet containing GSM operations and/or data, and the HFI window issues the GSM packet to the network fabric.

    Abstract translation: 一种用于通过主机结构接口(HFI)从耦合到分布式网络的网络结构的第一节点上的始发任务发出全局共享存储器(GSM)操作的方法。 始发任务在有效地址(EA)空间内生成GSM命令。 然后任务将GSM命令放在发送FIFO中。 发送FIFO是具有存储器映射到全局执行作业的EA的实际地址(RA)的实际存储器的一部分。 始发任务维护仅全局执行作业的实际地址空间的一部分的本地EA到RA映射。 该任务使HFI能够将GSM命令从发送FIFO检索到分配给始发任务的HFI窗口中。 HFI窗口产生包含GSM操作和/或数据的相应的GSM分组,并且HFI窗口向网络结构发出GSM分组。

    COMPLETION OF ASYNCHRONOUS MEMORY MOVE IN THE PRESENCE OF A BARRIER OPERATION
    12.
    发明申请
    COMPLETION OF ASYNCHRONOUS MEMORY MOVE IN THE PRESENCE OF A BARRIER OPERATION 失效
    在障碍物操作中完成异步记忆移动

    公开(公告)号:US20090198963A1

    公开(公告)日:2009-08-06

    申请号:US12024513

    申请日:2008-02-01

    Abstract: A method within a data processing system by which a processor executes an asynchronous memory move (AMM) store (ST) instruction to complete a corresponding AMM operation in parallel with an ongoing (not yet completed), previously issued barrier operation. The processor receives the AMM ST instruction after executing the barrier operation (or SYNC instruction) and before the completion of the barrier operation or SYNC on the system fabric. The processor continues executing the AMM ST instruction, which performs a move in virtual address space and then triggers the generation of the AMM operation. The AMM operation proceeds while the barrier operation continues, independent of the processor. The processor stops further execution of all other memory access requests, excluding AMM ST instructions that are received after the barrier operation, but before completion of the barrier operation.

    Abstract translation: 数据处理系统中的方法,通过该方法,处理器执行异步存储器移动(AMM)存储(ST)指令以与正在进行的(未完成)先前发布的屏障操作并行地完成对应的AMM操作。 执行屏障操作(或SYNC指令)后,在系统结构上完成屏障操作或SYNC之前,处理器接收AMM ST指令。 处理器继续执行AMM ST指令,其在虚拟地址空间中执行移动,然后触发AMM操作的生成。 无障碍操作继续进行,与处理器无关,AMM操作继续进行。 处理器停止所有其他存储器访问请求的进一步执行,排除在屏障操作之后但在屏障操作完成之前接收的AMM ST指令。

    Interface method, system, and program product for facilitating layering of a data communications protocol over an active message layer protocol
    13.
    发明授权
    Interface method, system, and program product for facilitating layering of a data communications protocol over an active message layer protocol 失效
    接口方法,系统和程序产品,用于促进通过活动消息层协议分层数据通信协议

    公开(公告)号:US07536468B2

    公开(公告)日:2009-05-19

    申请号:US10875471

    申请日:2004-06-24

    CPC classification number: H04L51/18 H04L69/22

    Abstract: A protocol interface is provided for an active message protocol of a computing environment and a client process employing the active message protocol. The protocol interface includes an interface to a header handler function associated with the client process. The interface to the header handler function has parameters to be passed by and a parameter to be returned to the active message protocol when processing a message received through the active message protocol. The parameters to be passed include current message state information and current message type information for the received message. These parameters facilitate message-specific decisions by the header handler function about processing data of the message by the active message protocol. The parameter to be returned to the active message protocol instructs the active message protocol how to process the received message other than just where to store the message.

    Abstract translation: 为计算环境的活动消息协议和采用活动消息协议的客户端进程提供协议接口。 协议接口包括与客户端进程相关联的报头处理函数的接口。 在处理通过活动消息协议接收的消息时,头处理程序函数的接口具有要传递的参数和要返回到活动消息协议的参数。 要传递的参数包括接收到的消息的当前消息状态信息和当前消息类型信息。 这些参数有助于报头处理函数关于通过活动消息协议处理消息的数据的消息特定决定。 要返回到活动消息协议的参数指示活动消息协议如何处理接收到的消息,而不仅仅是存储消息的位置。

    Method and system for efficiently transferring a self-defined non-contiguous message in a one-sided communication model
    14.
    发明授权
    Method and system for efficiently transferring a self-defined non-contiguous message in a one-sided communication model 失效
    用于在单面通信模型中有效传送自定义非连续消息的方法和系统

    公开(公告)号:US07454491B2

    公开(公告)日:2008-11-18

    申请号:US10965597

    申请日:2004-10-14

    CPC classification number: H04L67/10 H04L69/06 H04L69/328

    Abstract: A method and system for transferring noncontiguous messages group including assembling a set of data into a series of transmission packets, packaging a description of the layout of the transmission packets into description packets and then places each description packet into a local buffer while maintaining a count of the number of description packets, transfers each description packet into a transmit buffer for transmission to at least one receiving node, identifies the data packets, and forwards each data packet to the transmit buffer for transmission to the at least one receiving node. The receiving node receives the transmission packets, identifies each packet as a description packet or data packet, places the description packets in a local buffer for storage until the description is complete, places each description packet into a user data buffer, stores data packets in a local queue until the description is complete, then transfers the data packets to the user buffer.

    Abstract translation: 一种用于传送不连续消息组的方法和系统,包括将一组数据组合成一系列传输分组,将传输分组的布局的描述打包成描述分组,然后将每个描述分组放置到本地缓冲器中,同时保持计数 描述分组的数量将每个描述分组传送到用于发送到至少一个接收节点的发送缓冲器,识别数据分组,并将每个数据分组转发到发送缓冲器以传输到至少一个接收节点。 接收节点接收传输分组,将每个分组标识为描述分组或数据分组,将描述分组置于本地缓冲区中进行存储,直到描述完成,将每个描述分组放入用户数据缓冲区,将数据分组存储在 本地队列直到描述完成,然后将数据包传送到用户缓冲区。

    Load balancing of adapters on a multi-adapter node
    16.
    发明授权
    Load balancing of adapters on a multi-adapter node 有权
    多适配器节点上适配器的负载平衡

    公开(公告)号:US09430297B2

    公开(公告)日:2016-08-30

    申请号:US12334624

    申请日:2008-12-15

    CPC classification number: G06F9/5083 G06F9/541

    Abstract: Load balancing of adapters on a multi-adapter node of a communications environment. A task executing on the node selects an adapter resource unit to be used as its primary port for communications. The selection is based on the task's identifier, and facilitates a balancing of the load among the adapter resource units. Using the task's identifier, an index is generated that is used to select a particular adapter resource unit from a list of adapter resource units assigned to the task. The generation of the index is efficient and predictable.

    Abstract translation: 通信环境的多适配器节点上的适配器的负载平衡。 在节点上执行的任务选择要用作其通信主要端口的适配器资源单元。 该选择基于任务的标识符,并且有助于平衡适配器资源单元之间的负载。 使用任务的标识符,生成用于从分配给任务的适配器资源单元的列表中选择特定适配器资源单元的索引。 指数的生成是有效和可预测的。

    Establishing a group of endpoints in a parallel computer
    17.
    发明授权
    Establishing a group of endpoints in a parallel computer 有权
    在并行计算机中建立一组端点

    公开(公告)号:US09250948B2

    公开(公告)日:2016-02-02

    申请号:US13231326

    申请日:2011-09-13

    CPC classification number: G06F9/46 G06F9/546

    Abstract: A parallel computer executes a number of tasks, each task includes a number of endpoints and the endpoints are configured to support collective operations. In such a parallel computer, establishing a group of endpoints receiving a user specification of a set of endpoints included in a global collection of endpoints, where the user specification defines the set in accordance with a predefined virtual representation of the endpoints, the predefined virtual representation is a data structure setting forth an organization of tasks and endpoints included in the global collection of endpoints and the user specification defines the set of endpoints without a user specification of a particular endpoint; and defining a group of endpoints in dependence upon the predefined virtual representation of the endpoints and the user specification.

    Abstract translation: 并行计算机执行多个任务,每个任务包括多个端点,并且端点被配置为支持集合操作。 在这样的并行计算机中,建立一组端点,其接收包括在端点的全局集合中的一组端点的用户指定,其中用户规范根据端点的预定义虚拟表示来定义集合,预定义的虚拟表示 是阐述包括在全局端点集合中的任务和端点的组织的数据结构,并且用户规范定义了一组端点,而没有特定端点的用户指定; 并且根据端点的预定义虚拟表示和用户规范来定义一组端点。

    Preparing parallel tasks to use a synchronization register

    公开(公告)号:US09092272B2

    公开(公告)日:2015-07-28

    申请号:US13314477

    申请日:2011-12-08

    CPC classification number: G06F9/52 G06F9/522

    Abstract: A job may be divided into multiple tasks that may execute in parallel on one or more compute nodes. The tasks executing on the same compute node may be coordinated using barrier synchronization. However, to perform barrier synchronization, the tasks use (or attach) to a barrier synchronization register which establishes a common checkpoint for each of the tasks. A leader task may use a shared memory region to publish to follower tasks the location of the barrier synchronization register—i.e., a barrier synchronization register ID. The follower tasks may then monitor the shared memory to determine the barrier synchronization register ID. The leader task may also use a count to ensure all the tasks attach to the BSR. This advantageously avoids any task-to-task communication which may reduce overhead and improve performance.

    PREPARING PARALLEL TASKS TO USE A SYNCHRONIZATION REGISTER
    20.
    发明申请
    PREPARING PARALLEL TASKS TO USE A SYNCHRONIZATION REGISTER 有权
    准备使用并发寄存器的并行任务

    公开(公告)号:US20130152101A1

    公开(公告)日:2013-06-13

    申请号:US13314477

    申请日:2011-12-08

    CPC classification number: G06F9/52 G06F9/522

    Abstract: A job may be divided into multiple tasks that may execute in parallel on one or more compute nodes. The tasks executing on the same compute node may be coordinated using barrier synchronization. However, to perform barrier synchronization, the tasks use (or attach) to a barrier synchronization register which establishes a common checkpoint for each of the tasks. A leader task may use a shared memory region to publish to follower tasks the location of the barrier synchronization register—i.e., a barrier synchronization register ID. The follower tasks may then monitor the shared memory to determine the barrier synchronization register ID. The leader task may also use a count to ensure all the tasks attach to the BSR. This advantageously avoids any task-to-task communication which may reduce overhead and improve performance.

    Abstract translation: 作业可以被划分为可以在一个或多个计算节点上并行执行的多个任务。 可以使用屏障同步来协调在相同计算节点上执行的任务。 然而,为了执行屏障同步,任务使用(或附加)到屏障同步寄存器,其为每个任务建立公共检查点。 领导任务可以使用共享存储器区域向跟随者任务发布屏障同步寄存器的位置,即屏障同步寄存器ID。 随后的任务可以监视共享存储器以确定屏障同步寄存器ID。 领导任务还可以使用计数来确保所有任务附加到BSR。 这有利地避免任何任务到任务的通信,这可以减少开销并提高性能。

Patent Agency Ranking