摘要:
A method for memory bandwidth compression comprising analyzing a texture surface to identify one or more areas of the texture surface that are fetchable with lower memory bandwidth consumption as compared to other areas of the texture surface, adding metadata to a metadata surface associated with the texture surface based on the analysis, the metadata indicating the one or more areas of the texture surface that are fetchable with lower memory bandwidth consumption as compared to other areas of the texture surface, and fetching the texture surface in accordance with the metadata.
摘要:
A mechanism is described for facilitating general purpose input/output data capture and neural cache system for autonomous machines. A method of embodiments, as described herein, includes capturing, by an image capturing device, one or more images of one or more objects, where the one or more images represent input data associated with a neural network. The method may further include determining accuracy of first output results generated by a default neural caching system by comparing the first output results with second output results predicted by a custom neural caching system. The method may further include outputting, based on the accuracy, a final output results including at least one of the first output results or the second output results.
摘要:
Systems and methods may provide for identifying a tile associated with an image and ordering an entirety of the tile into a linear stream of pages associated with a frame buffer. Additionally, the linear stream of pages may be allocated to a cache. In one example, the linear stream of pages is allocated to the cache in accordance with a fixed set selection policy of the cache.
摘要:
A system and method for efficiently allocating data in a memory hierarchy. A system includes a memory controller for controlling accesses to a memory and a display controller for processing video frame data. The memory controller includes a cache capable of storing data read from the memory. A given video frame may be processed by the display controller and presented on a respective display screen. During processing, control logic within the display controller sends multiple memory access requests to the memory controller with cache hint information. For the frame data, the cache hint information may alternate between (i) indicating to store frame data read in response to respective requests in the memory cache and (ii) indicating to not store the frame data read in response to respective requests in the memory cache.
摘要:
Methods, systems, and devices are disclosed for preparing to prefetch memory in 3D graphical shader programs. Based on the memory addresses of texels from a texture map that were previously read, a memory address of a to-be-read texel is estimated by using a first-order derivative of the memory address with respect to pixel distance. The estimated memory address is then prefetched for use when the pixel corresponding to the texel is rendered. Whether or not to perform the estimation of the texel memory address can be determined by whether the corresponding pixels are on the same, flat facet on a 3D model. Multiple directions on the facet can be used for memory address estimation as well as higher-order derivatives.
摘要:
Methods and apparatus relating to techniques to improve/optimize latency and bandwidth efficiency for read modify write operations when a read operation is requested to a partially modified write only cacheline are described. In an embodiment, a first cache stores data from one or more cachelines of a second cache in response to a read hit write only operation (e.g., instead of sending the data to main memory). Write accumulate logic merges the stored data with one or more write operations. Other embodiments are also disclosed and claimed.
摘要:
While prefetching data for a second fiber, a hierarchical data structure is traversed using a first fiber after deferring traversal for the second fiber. Then context is switched to the second fiber, and the hierarchical data structure is traversed using second fiber while prefetching data for another fiber.
摘要:
A data caching method is performed to receive an instruction to operate based on a specific data set; determine whether the specific data set is cached in its memory; when the specific data set is not cached in the memory, determine a plurality of attributes for a plurality of data sets currently stored in the memory, determine whether these attributes satisfy data caching criteria for storing the specific data set, and furthermore, when the data caching criteria are not satisfied, select at least one of the plurality of data sets according to a data replacement rule, delete at least a portion of the selected data set from the memory, and download the specific data set from a remote source; operate the specific data set according to the user instruction; and store at least a portion of the specific data set in the memory.
摘要:
A method for storing information may include determining whether a received data object fits inside a particular one of a plurality of free blocks in a memory bitmap. Each of the plurality of free blocks may include a column of the memory bitmap with a top margin, a bottom margin, and a predetermined width. If the received data object fits, the received data object may be stored in the particular one of the plurality of free blocks, starting at the top margin of the particular one of the plurality of free blocks. The particular one of the plurality of data blocks may be resized by moving the top margin to start below the stored received data object. The determining may include, for each of the plurality of free blocks, a height of the received data object may be compared with a height of each of the free data blocks.
摘要:
Methods and apparatus for constructing objects within a cache system thereby allowing the cache system to respond to requested objects that are not initially available within the cache system. One embodiment of the invention caches image files, where the images are divided into components and stored in a format that allows identification and access to the components. The cache system determines that an object, such as an image file, is missing from the cache memory, locates sufficient components from the cache memory and/or external storage, and constructs the object from the located components.