In-node Aggregation and Disaggregation of MPI Alltoall and Alltoallv Collectives

    公开(公告)号:US20170255501A1

    公开(公告)日:2017-09-07

    申请号:US15446004

    申请日:2017-03-01

    Abstract: An MPI collective operation carried out in a fabric of network elements by transmitting MPI messages from all the initiator processes in an initiator node to designated ones of the responder processes in respective responder nodes. Respective payloads of the MPI messages are combined in a network interface device of the initiator node to form an aggregated MPI message. The aggregated MPI message is transmitted through the fabric to network interface devices of responder nodes, disaggregating the aggregated MPI message into individual messages, and distributing the individual messages to the designated responder node processes.

    MAINTAINING DATA CONFIDENTIALITY IN SHARED COMPUTING ENVIRONMENTS

    公开(公告)号:US20240202315A1

    公开(公告)日:2024-06-20

    申请号:US18084964

    申请日:2022-12-20

    CPC classification number: G06F21/53 G06F2221/033

    Abstract: The technology disclosed herein enables selective clearing of memory regions upon a context switch. An example method includes the operations of: receiving a memory access request referencing a memory region; determining an identifier of a current execution context associated with the memory region; determining an identifier of a previous execution context specified by metadata associated with the memory region; responsive to determining that the identifier of the current execution context does not match the identifier of the previous execution context, updating the metadata associated with the memory region to store the identifier of the current execution context; clearing at least a part of the memory region; and processing the memory access request with respect to the memory region.

    CONFIDENTIAL OFFLOADING OF PERSISTENT STORAGE OPERATIONS IN CONFIDENTIAL COMPUTING ENVIRONMENTS

    公开(公告)号:US20230273808A1

    公开(公告)日:2023-08-31

    申请号:US18104086

    申请日:2023-01-31

    Abstract: The technology disclosed herein enables a Trusted Execution Environment (TEE) to be extended to an auxiliary device that handles persistently storing data in a security enhanced manner. Extending the trusted computing base to the auxiliary device may involve establishing an auxiliary TEE in the auxiliary device and a trusted communication link between the primary and auxiliary TEEs. The primary TEE may include the computing resources of the primary devices (e.g., CPU and host memory) and the auxiliary TEE may include the computing resources of the auxiliary devices (e.g., hardware accelerators and auxiliary memory). The trusted communication link may enable the auxiliary TEE to access data of the primary TEE that is otherwise inaccessible to all software executing external to the primary TEE (e.g., host operating system and hypervisor). The auxiliary device may use the auxiliary TEE to process the data to avoid compromising the security enhancements provided by the primary TEE.

    Delayed responder-not-ready negative acknowledgement

    公开(公告)号:US20190289065A1

    公开(公告)日:2019-09-19

    申请号:US15924293

    申请日:2018-03-19

    Abstract: A method for communication includes receiving in a computer system a request from a peer computer system. Upon finding that the computer system is currently not ready to process the request, a Negative Acknowledgement (NAK) message is sent from the computer system to the peer computer system, at a sending time that is derived from a time at which the computer system is ready to process the request.

    Efficient and flexible flow inspector

    公开(公告)号:US12231401B2

    公开(公告)日:2025-02-18

    申请号:US17714207

    申请日:2022-04-06

    Abstract: In one embodiment, a data communication device includes a network interface controller to process packets received from at least one of a host device for sending over a network, and at least one remote device over the network, at least one processor to execute computer instructions to receive a configuration, and extract filtering rules from the configuration, and at least one hardware accelerator to receive the filtering rules from the at least one processor, and filter the packets based on the rules so that some of the packets are dropped and some of the packets are forwarded to the at least one processor to send data based on the forwarded packets to another device.

Patent Agency Ranking