CACHING OF ADAPTIVELY SIZED CACHE TILES IN A UNIFIED L2 CACHE WITH SURFACE COMPRESSION
    91.
    发明申请
    CACHING OF ADAPTIVELY SIZED CACHE TILES IN A UNIFIED L2 CACHE WITH SURFACE COMPRESSION 有权
    在具有表面压缩的统一L2缓存中缓存适应尺寸的高速缓存

    公开(公告)号:US20140118379A1

    公开(公告)日:2014-05-01

    申请号:US14012308

    申请日:2013-08-28

    Abstract: One embodiment of the present invention includes techniques for adaptively sizing cache tiles in a graphics system. A device driver associated with a graphics system sets a cache tile size associated with a cache tile to a first size. The detects a change from a first render target configuration that includes a first set of render targets to a second render target configuration that includes a second set of render targets. The device driver sets the cache tile size to a second size based on the second render target configuration. One advantage of the disclosed approach is that the cache tile size is adaptively sized, resulting in fewer cache tiles for less complex render target configurations. Adaptively sizing cache tiles leads to more efficient processor utilization and reduced power requirements. In addition, a unified L2 cache tile allows dynamic partitioning of cache memory between cache tile data and other data.

    Abstract translation: 本发明的一个实施例包括用于在图形系统中自动调整高速缓存块的尺寸的技术。 与图形系统相关联的设备驱动程序将与高速缓存平铺相关联的高速缓存平铺大小设置为第一大小。 检测从包括第一组渲染目标的第一渲染目标配置到包括第二组渲染目标的第二渲染目标配置的改变。 设备驱动程序基于第二渲染目标配置将高速缓存磁贴大小设置为第二大小。 所公开方法的一个优点是缓存片大小自适应地大小,导致用于较不复杂的渲染目标配置的较少高速缓存片。 自动调整高速缓存片段的大小可以提高处理器利用率并降低功耗。 另外,一个统一的L2高速缓存块允许缓存区块数据和其他数据之间的高速缓冲存储器的动态分区。

    Fast, dynamic cache packing
    92.
    发明授权
    Fast, dynamic cache packing 有权
    快速,动态的高速缓存包装

    公开(公告)号:US08510531B1

    公开(公告)日:2013-08-13

    申请号:US13627320

    申请日:2012-09-26

    Abstract: A method for storing information may include determining whether a received data object fits inside a particular one of a plurality of free blocks in a memory bitmap. Each of the plurality of free blocks may include a column of the memory bitmap with a top margin, a bottom margin, and a predetermined width. If the received data object fits, the received data object may be stored in the particular one of the plurality of free blocks, starting at the top margin of the particular one of the plurality of free blocks. The particular one of the plurality of data blocks may be resized by moving the top margin to start below the stored received data object. The determining may include, for each of the plurality of free blocks, a height of the received data object may be compared with a height of each of the free data blocks.

    Abstract translation: 用于存储信息的方法可以包括确定接收到的数据对象是否适合存储器位图中的多个空闲块中的特定一个。 多个空闲块中的每一个可以包括具有顶边距,底边距和预定宽度的存储位图的列。 如果接收的数据对象适合,则从多个空闲块的特定一个的顶部边缘开始,接收的数据对象可以存储在多个空闲块中的特定的一个中。 多个数据块中的特定一个可以通过移动顶部余量来重新调整大小,从而开始于存储的接收数据对象之下。 对于多个空闲块中的每一个,确定可以包括将接收到的数据对象的高度与每个空闲数据块的高度进行比较。

    Matrix transpose hardware acceleration

    公开(公告)号:US12125124B1

    公开(公告)日:2024-10-22

    申请号:US18118251

    申请日:2023-03-07

    Inventor: Kun Xu Ron Diamant

    Abstract: In one example, an apparatus comprises: a buffer memory; and a memory access circuit configured to: fetch, from a first memory, a set of first groups of data elements of a first matrix, each first group of data elements being stored at consecutive memory addresses at the first memory; based on a first configuration, store the set of first groups of data elements at consecutive memory addresses or at non-consecutive memory addresses at the buffer memory; based on a second configuration that defines a memory address offset, fetch a set of second groups of the data elements from the buffer memory, each second group of the data elements being stored at consecutive memory addresses of the buffer memory, each second group being separated by the memory address offset in the buffer memory; and store each fetched second group at consecutive addresses of a destination memory to form a second matrix.

Patent Agency Ranking