METHOD AND APPARATUS WITH NEURAL NETWORK CONVOLUTION OPERATIONS

    公开(公告)号:US20230169340A1

    公开(公告)日:2023-06-01

    申请号:US18089696

    申请日:2022-12-28

    CPC classification number: G06N3/08 G06F7/74 G06N20/00

    Abstract: A processor-implemented method of performing convolution operations in a neural network includes generating a plurality of first sub-bit groups and a plurality of second sub-bit groups, respectively from at least one pixel value of an input feature map and at least one predetermined weight, performing a convolution operation on a first pair that includes a first sub-bit group including a most significant bit (MSB) of the at least one pixel value and a second sub-bit group including an MSB of the at least one predetermined weight, based on the plurality of second sub-bit groups, obtaining a maximum value of a sum of results for convolution operations of remaining pairs excepting the first pair, and based on a result of the convolution operation on the first pair and the maximum value, determining whether to perform the convolution operations of the remaining pairs.

    NEURAL NETWORK OPERATION METHOD AND DEVICE

    公开(公告)号:US20220269950A1

    公开(公告)日:2022-08-25

    申请号:US17397082

    申请日:2021-08-09

    Inventor: Sehwan LEE

    Abstract: A neural network operation device includes an input feature map buffer to store an input feature map, a weight buffer to store a weight, an operator including an adder tree unit to perform an operation between the input feature map and the weight by a unit of a reference bit length, and a controller to map the input feature map and the weight to the operator to provide one or both of a mixed precision operation and data parallelism.

    COMPUTING METHOD AND DEVICE WITH DATA SHARING

    公开(公告)号:US20220164289A1

    公开(公告)日:2022-05-26

    申请号:US17317339

    申请日:2021-05-11

    Abstract: A computing method and device with data sharing re provided. The method includes loading, by a loader, input data of an input feature map stored in a memory in loading units according to a loading order, storing, by a buffer controller, the loaded input data in a reuse buffer of an address rotationally allocated according to the loading order, and transmitting, by each of a plurality of senders, to an executer respective input data corresponding to each output data of respective convolution operations among the input data stored in the reuse buffer, wherein portions of the transmitted respective input data overlap other.

    ELECTRONIC DEVICE AND METHOD FOR INFERENCE BINARY AND TERNARY NEURAL NETWORKS

    公开(公告)号:US20210150313A1

    公开(公告)日:2021-05-20

    申请号:US17098589

    申请日:2020-11-16

    Abstract: A method for computing an inner product on a binary data, a ternary data, a non-binary data, and a non-ternary data using an electronic device. The method includes calculating the inner product on a ternary data, designing a fused bitwise data path to support the inner product calculation on the binary data and the ternary data, designing a FPL data path to calculate an inner product between one of the non-binary data and the non-ternary data and one of the binary data and the ternary data, and distributing the inner product calculation for the binary data and the ternary data and the inner product between one of the non-binary data and the non-ternary data and one of the binary data and the ternary data in the fused bitwise data path and the FPL data path.

    METHOD AND APPARATUS WITH CONVOLUTION NEURAL NETWORK PROCESSING

    公开(公告)号:US20200250842A1

    公开(公告)日:2020-08-06

    申请号:US16564215

    申请日:2019-09-09

    Inventor: Sehwan LEE

    Abstract: A neural network apparatus includes one or more processors comprising: a controller configured to determine a shared operand to be shared in parallelized operations as being either one of a pixel value among pixel values of an input feature map and a weight value among weight values of a kernel, based on either one or both of a feature of the input feature map and a feature of the kernel; and one or more processing units configured to perform the parallelized operations based on the determined shared operand.

    NEURAL NETWORK METHOD AND APPARATUS
    29.
    发明公开

    公开(公告)号:US20240112030A1

    公开(公告)日:2024-04-04

    申请号:US18529620

    申请日:2023-12-05

    CPC classification number: G06N3/08 G06N3/0495

    Abstract: A neural network method and apparatus is provided. A processor-implemented neural network method includes a processor and a memory storing information, including stored predetermined precision parameters of a layer of a n neural network, about the layer, the method includes obtaining information about the layer in the memory indicative of the number of output classes; determining, based on the obtained information, a precision for the layer based on the number of output classes of the layer, wherein the precision is determined proportionally with respect to the obtained number of output classes; and processing new parameters, with a set precision, for the layer based on the stored parameter.

Patent Agency Ranking