METHOD AND APPARATUS WITH TRANSFORMER MODEL TRAINING

    公开(公告)号:US20240232581A9

    公开(公告)日:2024-07-11

    申请号:US18450839

    申请日:2023-08-16

    IPC分类号: G06N3/0455

    CPC分类号: G06N3/0455

    摘要: A device including processors configured to execute instructions and memories storing the instructions, which when executed by the processors configure the processors to perform an operation for training a transformer model having a plurality of encoders and a plurality of decoders by configuring the processors to identify the batches of training data into a plurality of micro-batches, select layer pairs for the plurality of micro-batches, assemble a processing order of the layer pairs, determining resource information to be allocated to the layer pairs, and allocate resources to the layer pairs based on the determined resource information to be allocated to the layer pairs, dependent con the processing order of the layer pairs.

    DEVICE AND METHOD WITH BATCH NORMALIZATION
    2.
    发明公开

    公开(公告)号:US20240184630A1

    公开(公告)日:2024-06-06

    申请号:US18526603

    申请日:2023-12-01

    IPC分类号: G06F9/50 G06F5/01 G06F15/80

    摘要: A device and method with batch normalization are provided. An accelerator includes: core modules, each core module including a respective plurality of cores configured to perform a first convolution operation using feature map data and a weight; local reduction operation modules adjacent to the respective core modules, each including a respective plurality of local reduction operators configured to perform a first local operation that obtains first local statistical values of the corresponding core module; a global reduction operation module configured to perform a first global operation that generates first global statistical values of the core module based on the first local statistical values of the core modules; and a normalization operation module configured to perform a first normalization operation on the feature map data based on the first global statistical values.

    MEMORY DEVICE, MEMORY MODULE INCLUDING THE SAME, AND MEMORY SYSTEM INCLUDING THE SAME
    3.
    发明申请
    MEMORY DEVICE, MEMORY MODULE INCLUDING THE SAME, AND MEMORY SYSTEM INCLUDING THE SAME 有权
    存储器件,包括其的存储器模块和包括其的存储器系统

    公开(公告)号:US20160240242A1

    公开(公告)日:2016-08-18

    申请号:US14734101

    申请日:2015-06-09

    摘要: A memory device includes a first memory cell, a second memory cell, a precharge circuit, a sense amplifier, a switch circuit, and a controller. The first memory cell is connected to a first bit line, the second memory cell is connected to a second bit line, and the precharge circuit connected between the first bit line and the second bit line. The sense amplifier includes a first input terminal and a second input terminal. The switch circuit is connected to the first bit line and the first input terminal and to the second bit line and the second input terminal and is configured to control a connection between the first bit line and the first input terminal and a connection between the second bit line and the second input terminal in response to a switch signal. The controller is configured to generate the switch signal in response to a command.

    摘要翻译: 存储器件包括第一存储单元,第二存储单元,预充电电路,读出放大器,开关电路和控制器。 第一存储单元连接到第一位线,第二存储单元连接到第二位线,并且预充电电路连接在第一位线和第二位线之间。 读出放大器包括第一输入端和第二输入端。 开关电路连接到第一位线和第一输入端以及第二位线和第二输入端,并且被配置为控制第一位线和第一输入端之间的连接以及第二位之间的连接 线路和第二输入端子响应于开关信号。 控制器被配置为响应于命令产生开关信号。

    METHOD AND APPARATUS WITH TRANSFORMER MODEL TRAINING

    公开(公告)号:US20240135147A1

    公开(公告)日:2024-04-25

    申请号:US18450839

    申请日:2023-08-15

    IPC分类号: G06N3/0455

    CPC分类号: G06N3/0455

    摘要: A device including processors configured to execute instructions and memories storing the instructions, which when executed by the processors configure the processors to perform an operation for training a transformer model having a plurality of encoders and a plurality of decoders by configuring the processors to identify the batches of training data into a plurality of micro-batches, select layer pairs for the plurality of micro-batches, assemble a processing order of the layer pairs, determining resource information to be allocated to the layer pairs, and allocate resources to the layer pairs based on the determined resource information to be allocated to the layer pairs, dependent con the processing order of the layer pairs.

    DEVICE AND METHOD WITH TRANSFORMER MODEL IMPLEMENTATION

    公开(公告)号:US20230138659A1

    公开(公告)日:2023-05-04

    申请号:US17887145

    申请日:2022-08-12

    IPC分类号: G06N3/04

    摘要: A device and method with transformer model implementation are provided. The electronic device includes a processor configured to perform an inference by implementing a transformer model including a plurality of encoders and a plurality of decoders, and a memory configured to store instructions to be executed by the processor. Each of the encoders and the decoders includes an attention block that determines an attention value. The processor is configured to perform a first sub-softmax tile-wise operation in the attention block, perform a reduction operation to determine an adjustment factor based on a resulting value of the first sub-softmax operation, and perform a second sub-softmax tile-wise operation based on a resulting value of the reduction operation.

    ACCELERATOR AND ELECTRONIC DEVICE INCLUDING THE SAME

    公开(公告)号:US20220365891A1

    公开(公告)日:2022-11-17

    申请号:US17876116

    申请日:2022-07-28

    IPC分类号: G06F13/16 G06F15/80

    摘要: An accelerator includes: a memory configured to store input data; a plurality of shift buffers each configured to shift input data received sequentially from the memory in each cycle, and in response to input data being stored in each of internal elements of the shift buffer, output the stored input data to a processing element (PE) array; a plurality of backup buffers each configured to store input data received sequentially from the memory and transfer the stored input data to one of the shift buffers; and the PE array configured to perform an operation on input data received from one or more of the shift buffers and on a corresponding kernel.