Abstract:
Techniques for implementing a hardware-based cache controller in, e.g., a tiered memory computer system are provided. In one set of embodiments, the cache controller can flexibly operate in a number of different modes that aid the OS/hypervisor of the computer system in managing and optimizing its use of the system's memory tiers. In another set of embodiments, the cache controller can implement a hardware architecture that enables it to significantly reduce the probability of tag collisions, decouple cache capacity management from cache lookup and allocation, and handle multiple concurrent cache transactions.
Abstract:
A device (120) is disclosed. An interface (305) may connect the device (120) to a processor (110). The interface (305) may support a first protocol. The device (120) may include a first storage (315) for a data and a second storage (320) for the data. The second storage (320) may support a second protocol different from the first protocol. A controller (325) may be connected to the interface (305) and the first storage (315). A bridge (330) may be connected to the interface (305), the first storage (315), and the second storage (320). The bridge (330) may include a filter (335) configured to coordinate a data transfer between the first storage (315) and the second storage (320).
Abstract:
Embodiments of this application disclose a translation lookaside buffer maintenance method and a related device. The method is applied to an electronic device, the electronic device includes a plurality of physical central processing units CPUs, a first process is run on the electronic device, the first process currently includes M first threads, the M first threads are currently being respectively run on M physical CPUs of the plurality of physical CPUs, and M is an integer greater than or equal to 1. The method includes: determining a physical CPU range S1 currently corresponding to the first process, where the physical CPU range S1 includes the M physical CPUs on which the first threads in the first process are currently being run; and updating, based on page table information maintained by the first process, translation lookaside buffer TLB information maintained by all physical CPUs in the physical CPU range S 1. According to embodiments of this application, a TLB maintenance delay can be reduced.
Abstract:
In various embodiments, methods and systems, for implementing modular data operations, are provided. A data access request, associated with data, is received at a data access component. The data access component selectively implements modular data operations functionality based on configuration settings. A translation table associated with a working set is accessed, based on the configuration settings of the data access component, to determine a location for executing the data access request. The data access request is executed using the cache store or a backing store associated with the working set. The data access request is executed using the location that is determined using the translation table of the working set. The data access request is executed using the cache store when the data is cached in the cache store and the data access requested is executed based on the backing store when the data is un-cached in the cache store.
Abstract:
An instruction in an instruction cache line having a first portion that is cacheable, a second portion that is from a page that is non-cacheable, and crosses a cache line is prevented from executing from the instruction cache. An attribute associated with the non-cacheable second portion is tracked separately from the attributes of the rest of the instructions in the cache line. If the page crossing instruction is reached for execution, the page crossing instruction and instructions following are flushed and a non-cacheable request is made to memory for at least the second portion. Once the second portion is received, the whole page crossing instruction is reconstructed from the first portion saved in the previous fetch group. The page crossing instruction or portion thereof is returned with the proper attribute for a non-cached fetched instruction and the reconstructed instruction can be executed without being cached.
Abstract:
A host device includes a controller configured to receive input/output (TO) access information associated with an IO workload, the IO access information identifying at least one of a read action and a write action associated with the IO workload over a period of time. Based upon the received IO access information associated with the storage element, the controller is configured to derive a predicted cache access ratio associated with the IO workload and relating a predicted number of cache accesses associated with the IO workload with at least one of a total number of read actions and a total number of write actions associated with the IO workload. When the predicted cache access ratio reaches a threshold cache access ratio value, the controller is configured to identify the IO workload as an IO workload candidate for caching by the host device.
Abstract:
A cache memory apparatus according to the present invention includes a cache memory that caches an instruction code corresponding to a fetch address and a cache control circuit that controls the instruction code to be cached in the cache memory. The cache control circuit caches an instruction code corresponding to a subroutine when the fetch address indicates a branch into the subroutine and disables the instruction code to be cached when the number of the instruction codes to be cached exceeds a previously set maximum number.
Abstract:
Provided is a processor including a plurality of devices. The processor includes a source processing device configured to identify data to request from another device, and a destination processing device configured to, in response to a request for the identified data from the source processing device using credit-based flow control, transmit the identified data to the source processing device using the credit-based flow control. The source processing device includes a credit buffer used for the credit-based flow control, the credit buffer being allocable to include a cache region configured to cache the transmitted identified data received by the source processing device.
Abstract:
Examples of the invention present an optimised method of managing a cache, where data items, such as video clips, are moved up in rank in a cache whenever it is requested, swapping rank with the data item above it. New data items not present in the cache are obtained and stored in the cache with a rank below the lowest ranked data item. If the cache is full, then the data item at the lowest rank is replaced by the newly requested data item. Thus data items that receive increasing numbers of requests are moved up the cache and thus have more protection from deletion.