MEMORY LOAD TO LOAD FUSING
    1.
    发明申请

    公开(公告)号:US20180267800A1

    公开(公告)日:2018-09-20

    申请号:US15615811

    申请日:2017-06-06

    Abstract: A system and a method to cascade execution of instructions in a load-store unit (LSU) of a central processing unit (CPU) to reduce latency associated with the instructions. First data stored in a cache is read by the LSU in response a first memory load instruction of two immediately consecutive memory load instructions. Alignment, sign extension and/or endian operations are performed on the first data read from the cache in response to the first memory load instruction, and, in parallel, a memory-load address-forwarded result is selected based on a corrected alignment of the first data read in response to the first memory load instruction to provide a next address for a second of the two immediately consecutive memory load instructions. Second data stored in the cache is read by the LSU in response to the second memory load instruction based on the selected memory-load address-forwarded result.

    MEMORY LOAD TO LOAD FUSING
    2.
    发明申请

    公开(公告)号:US20190278603A1

    公开(公告)日:2019-09-12

    申请号:US16421463

    申请日:2019-05-23

    Abstract: A system and a method to cascade execution of instructions in a load-store unit (LSU) of a central processing unit (CPU) to reduce latency associated with the instructions. First data stored in a cache is read by the LSU in response a first memory load instruction of two immediately consecutive memory load instructions. Alignment, sign extension and/or endian operations are performed on the first data read from the cache in response to the first memory load instruction, and, in parallel, a memory-load address-forwarded result is selected based on a corrected alignment of the first data read in response to the first memory load instruction to provide a next address for a second of the two immediately consecutive memory load instructions. Second data stored in the cache is read by the LSU in response to the second memory load instruction based on the selected memory-load address-forwarded result.

    MEMORY LOAD AND ARITHMETIC LOAD UNIT (ALU) FUSING

    公开(公告)号:US20180267775A1

    公开(公告)日:2018-09-20

    申请号:US15612963

    申请日:2017-06-02

    CPC classification number: G06F7/485 G06F7/38 G06F7/50

    Abstract: According to one general aspect, a load unit may include a load circuit configured to load at least one piece of data from a memory. The load unit may include an alignment circuit configured to align the data to generate an aligned data. The load unit may also include a mathematical operation execution circuit configured to generate a resultant of a predetermined mathematical operation with the at least one piece of data as an operand. Wherein the load unit is configured to, if an active instruction is associated with the predetermined mathematical operation, bypass the alignment circuit and input the piece of data directly to the mathematical operation execution circuit.

    METHOD AND APPARATUS FOR HANDLING PROCESSOR READ-AFTER-WRITE HAZARDS WITH CACHE MISSES
    4.
    发明申请
    METHOD AND APPARATUS FOR HANDLING PROCESSOR READ-AFTER-WRITE HAZARDS WITH CACHE MISSES 有权
    处理器处理器读取后写入危险的方法和设备与缓存错误

    公开(公告)号:US20150186285A1

    公开(公告)日:2015-07-02

    申请号:US14307444

    申请日:2014-06-17

    Inventor: Paul E. KITCHIN

    CPC classification number: G06F12/0875 G06F9/3834 G06F9/3836 G06F12/0831

    Abstract: According to one general aspect, an apparatus may include an instruction fetch unit, an execution unit, and a cache resynchronization predictor, as described above. The instruction fetch unit may be configured to issue a first memory read operation to a memory address, and a first memory write operation to the memory address, wherein the first memory read operation is stored at an instruction address. The execution unit may be configured to execute the first memory read operation, wherein the execution of the first memory read operation causes a resynchronization exception. The cache resynchronization predictor may be configured to associate the instruction address with a resynchronization exception, and determine if a memory read operation stored at the instruction address comprises a resynchronization predicted store.

    Abstract translation: 根据一个一般方面,如上所述,装置可以包括指令获取单元,执行单元和高速缓存再同步预测器。 指令提取单元可以被配置为向存储器地址发出第一存储器读取操作以及对存储器地址的第一存储器写入操作,其中第一存储器读取操作存储在指令地址处。 执行单元可以被配置为执行第一存储器读取操作,其中执行第一存储器读取操作引起重新同步异常。 缓存重新同步预测器可以被配置为将指令地址与重新同步异常相关联,并且确定存储在指令地址处的存储器读取操作是否包括重新同步预测存储。

    REGISTER RENAMING OF A SHAREABLE INSTRUCTION OPERAND CACHE

    公开(公告)号:US20200225954A1

    公开(公告)日:2020-07-16

    申请号:US16294916

    申请日:2019-03-06

    Abstract: A system and a method are disclosed to process instructions in an execution unit (EU) that includes an operand cache (OC). The OC stores a copy of at least one frequently used operand stored in a physical register file (PRF). The EU may process instructions using operands obtained from the PRF or from the OC. In the first mode, an OC renaming unit (OC-REN) indicates to the EU to process instructions using operands obtained from the OC if processing the instructions using operands obtained from the OC uses less power than using operands obtained from the PRF. In the second mode, the OC-REN indicates to the EU to process the instructions using operands obtained from the PRF if processing the instructions using operands obtained from the PRF uses less power than using operands obtained from the OC.

    DECOUPLING L2 BTB FROM L2 CACHE TO ACCELERATE SEARCH FOR MISS AFTER MISS
    6.
    发明申请
    DECOUPLING L2 BTB FROM L2 CACHE TO ACCELERATE SEARCH FOR MISS AFTER MISS 审中-公开
    从L2缓存中解码L2 BTB,以便在MISS之后加速MISS的搜索

    公开(公告)号:US20150268961A1

    公开(公告)日:2015-09-24

    申请号:US14463638

    申请日:2014-08-19

    Abstract: According to one general aspect, a method may include requesting, from a second tier of a cache memory system, a first instruction stored at a first memory address. The method may also include requesting, from a second tier of a branch target buffer system, a branch record associated with the first memory address. The method may also include receiving the branch record before receiving the first instruction. The method may also include pre-fetching, in response to receiving the branch record and before receiving the first instruction, a non-sequential instruction stored at a non-sequential memory address.

    Abstract translation: 根据一个一般方面,一种方法可以包括从高速缓冲存储器系统的第二层请求存储在第一存储器地址处的第一指令。 该方法还可以包括从分支目标缓冲器系统的第二层请求与第一存储器地址相关联的分支记录。 该方法还可以包括在接收第一指令之前接收分支记录。 该方法还可以包括:响应于接收到分支记录并在接收到第一指令之前预取存储在非顺序存储器地址处的非顺序指令。

Patent Agency Ranking