Cache management
    21.
    发明授权
    Cache management 有权
    缓存管理

    公开(公告)号:US09390010B2

    公开(公告)日:2016-07-12

    申请号:US13715526

    申请日:2012-12-14

    CPC classification number: G06F12/0804 G06F12/0888 Y02D10/13

    Abstract: The present disclosure provides techniques for cache management. A data block may be received from an IO interface. After receiving the data block, the occupancy level of a cache memory may be determined. The data block may be directed to a main memory if the occupancy level exceeds a threshold. The data block may be directed to a cache memory if the occupancy level is below a threshold.

    Abstract translation: 本公开提供了用于高速缓存管理的技术。 可以从IO接口接收数据块。 在接收到数据块之后,可以确定高速缓冲存储器的占用水平。 如果占用率超过阈值,则数据块可以被引导到主存储器。 如果占用水平低于阈值,则数据块可以被引导到高速缓冲存储器。

    Technologies for dynamic batch size management

    公开(公告)号:US11362968B2

    公开(公告)日:2022-06-14

    申请号:US15640258

    申请日:2017-06-30

    Abstract: Technologies for dynamically managing a batch size of packets include a network device. The network device is to receive, into a queue, packets from a remote node to be processed by the network device, determine a throughput provided by the network device while the packets are processed, determine whether the determined throughput satisfies a predefined condition, and adjust a batch size of packets in response to a determination that the determined throughput satisfies a predefined condition. The batch size is indicative of a threshold number of queued packets required to be present in the queue before the queued packets in the queue can be processed by the network device.

    Technologies for offloading data object replication and service function chain management

    公开(公告)号:US11082515B2

    公开(公告)日:2021-08-03

    申请号:US14866891

    申请日:2015-09-26

    Abstract: Technologies for offloading data object replication and service function chain management include a switch communicatively coupled to one or more computing nodes capable of executing virtual machines and storing data objects. The switch is configured to determine metadata of a service function chain, transmit a network packet to a service function of the service function chain being executed by one or more of the computing nodes for processing the network packet. The switch is further configured to receive feedback from service function, update the metadata based on the feedback, and transmit the network packet to a next service function of the service function chain. Additionally or alternatively, the switch is configured to identify a plurality of computing nodes (i.e., storage nodes) at which to store a received data object, replicate the data object based on the number of storage nodes, and transmit each of the received data object and replicated data object(s) to different corresponding storage nodes. Other embodiments are described and claimed.

    Technologies for a least recently used cache replacement policy using vector instructions

    公开(公告)号:US10789176B2

    公开(公告)日:2020-09-29

    申请号:US16059147

    申请日:2018-08-09

    Abstract: Technologies for least recently used (LRU) cache replacement include a computing device with a processor with vector instruction support. The computing device retrieves a bucket of an associative cache from memory that includes multiple entries arranged from front to back. The bucket may be a 256-bit array including eight 32-bit entries. For lookups, a matching entry is located at a position in the bucket. The computing device executes a vector permutation processor instruction that moves the matching entry to the front of the bucket while preserving the order of other entries of the bucket. For insertion, an inserted entry is written at the back of the bucket. The computing device executes a vector permutation processor instruction that moves the inserted entry to the front of the bucket while preserving the order of other entries. The permuted bucket is stored to the memory. Other embodiments are described and claimed.

    Apparatus and method for prioritized quality of service processing for transactional memory

    公开(公告)号:US10719442B2

    公开(公告)日:2020-07-21

    申请号:US16126907

    申请日:2018-09-10

    Abstract: An apparatus and method for prioritizing transactional memory regions. For example, one embodiment of a processor comprises: a plurality of cores to execute threads comprising sequences of instructions, at least some of the instructions specifying a transactional memory region; a cache of each core to store a plurality of cache lines; transactional memory circuitry of each core to manage execution of the transactional memory (TM) regions based on priorities associated with each of the TM regions; and wherein the transactional memory circuitry, upon detecting a conflict between a first TM region having a first priority value and a second TM region having a second priority value, is to determine which of the first TM region or the second TM region is permitted to continue executing and which is to be aborted based, at least in part, on the first and second priority values.

Patent Agency Ranking