METHODS AND APPARATUS TO REDUCE BANK PRESSURE USING AGGRESSIVE WRITE MERGING

    公开(公告)号:US20200371921A1

    公开(公告)日:2020-11-26

    申请号:US16882264

    申请日:2020-05-22

    Abstract: Methods, apparatus, systems and articles of manufacture to reduce bank pressure using aggressive write merging are disclosed. An example apparatus includes a first cache storage; a second cache storage; a store queue coupled to at least one of the first cache storage and the second cache storage and operable to: receive a first memory operation; process the first memory operation for storing the first set of data in at least one of the first cache storage and the second cache storage; receive a second memory operation; and prior to storing the first set of data in the at least one of the first cache storage and the second cache storage, merge the first memory operation and the second memory operation.

    Protection of memories, datapath and pipeline registers, and other storage elements by distributed delayed detection and correction of soft errors
    23.
    发明授权
    Protection of memories, datapath and pipeline registers, and other storage elements by distributed delayed detection and correction of soft errors 有权
    通过分布式延迟检测和软错误校正保护存储器,数据通路和流水线寄存器以及其他存储元件

    公开(公告)号:US09557936B2

    公开(公告)日:2017-01-31

    申请号:US14587234

    申请日:2014-12-31

    CPC classification number: G06F3/0673 G06F3/0619 G06F3/064 G06F11/1048

    Abstract: This invention is data processing apparatus and method. Data is protecting from corruption using an error correction code by generating an error correction code corresponding to the data. In this invention the data and the corresponding error correction code are carried forward to another set of registers without regenerating the error correction code or using the error correction code for error detection or correction. Only later are error correction detection and correction actions taken. The differing data/error correction code registers may be in differing pipeline phases in the data processing apparatus. This invention forwards the error correction code with the data through the entire datapath that carries the data. This invention provides error protection to the whole datapath without requiring extensive hardware or additional time.

    Abstract translation: 本发明是数据处理装置和方法。 通过产生对应于该数据的纠错码,使用纠错码来防止数据损坏。 在本发明中,将数据和相应的纠错码转发到另一组寄存器,而不用再生纠错码或使用纠错码进行错误检测或校正。 只有以后才采取纠错检测和纠正措施。 在数据处理装置中不同的数据/纠错码寄存器可能处于不同的流水线相位。 本发明通过携带数据的整个数据路径转发具有数据的纠错码。 本发明为整个数据路径提供错误保护,而不需要大量硬件或额外的时间。

    ZERO CYCLE CLOCK INVALIDATE OPERATION
    24.
    发明申请
    ZERO CYCLE CLOCK INVALIDATE OPERATION 审中-公开
    零周期无效操作

    公开(公告)号:US20160026569A1

    公开(公告)日:2016-01-28

    申请号:US14875801

    申请日:2015-10-06

    Abstract: A method to eliminate the delay of a block invalidate operation in a multi CPU environment by overlapping the block invalidate operation with normal CPU accesses, thus making the delay transparent. A range check is performed on each CPU access while a block invalidate operation is in progress, and an access that maps to within the address range of the block invalidate operation will be treated as a cache miss to ensure that the requesting CPU will receive valid data.

    Abstract translation: 通过将块无效操作与正常CPU访问重叠来消除多CPU环境中的块无效操作的延迟的方法,从而使得延迟变得透明。 在块无效操作正在进行时,对每个CPU访问执行范围检查,并且映射到块无效操作的地址范围内的访问将被视为缓存未命中,以确保请求的CPU将接收有效数据 。

    ALIASED MODE FOR CACHE CONTROLLER
    29.
    发明申请

    公开(公告)号:US20250021481A1

    公开(公告)日:2025-01-16

    申请号:US18797945

    申请日:2024-08-08

    Abstract: An apparatus includes first CPU and second CPU cores, a L1 cache subsystem coupled to the first CPU core and comprising a L1 controller, and a L2 cache subsystem coupled to the L1 cache subsystem and to the second CPU core. The L2 cache subsystem includes a L2 memory and a L2 controller configured to operate in an aliased mode in response to a value in a memory map control register being asserted. In the aliased mode, the L2 controller receives a first request from the first CPU core directed to a virtual address in the L2 memory, receives a second request from the second CPU core directed to the virtual address in the L2 memory, directs the first request to a physical address A in the L2 memory, and directs the second request to a physical address B in the L2 memory.

Patent Agency Ranking