METHOD AND APPARATUS OF OPERATING A NEURAL NETWORK

    公开(公告)号:US20220253692A1

    公开(公告)日:2022-08-11

    申请号:US17400353

    申请日:2021-08-12

    Abstract: Disclosed is a method and apparatus of operating a neural network. The neural network operation method includes receiving data for the neural network operation, verifying whether competition occurs between a first data traversal path corresponding to a first operation device and a second data traversal path corresponding to a second operation device, determining first operand data and second operand data from among the data using a result of the verifying and a priority between the first data traversal path and the second data traversal path, and performing the neural network operation based on the first operand data and the second operand data.

    Z-FIRST REFERENCE NEURAL PROCESSING UNIT FOR MAPPING WINOGRAD CONVOLUTION AND A METHOD THEREOF

    公开(公告)号:US20210357734A1

    公开(公告)日:2021-11-18

    申请号:US17239892

    申请日:2021-04-26

    Abstract: A z-first reference neural processing unit (NPU) for mapping Winograd Convolution is disclosed where the NPU includes memory banks configured to store input feature maps (IFMs) in a z-first data storage layout, each of the memory banks being configured to store the IFMs in one of a direct convolution (DConv) mode or a Winograd convolution (WgConv) mode, a reconfigurable IFM distributor configured to receive the IFMs from the memory banks, a parallel reconfigurable Winograd forward transform module configured to receive the IFMs from the reconfigurable IFM distributor and to transform the IFMs in a Winograd domain to transformed IFMs in the WgConv mode, multiply and accumulate (MAC) units configured to perform dot product operations on one of IFMs in the DConv mode and the transformed IFMs in the WgConv mode to obtain intermediate output feature maps (OFMs), and a reconfigurable OFM adder and Winograd inverse transform module configured to generate one of an OFM from the intermediate OFMs in the DConv mode and OFMs from the intermediate OFMs in the WgConv.

    METHOD AND APPARATUS WITH NEURAL NETWORK PERFORMING DECONVOLUTION

    公开(公告)号:US20210117791A1

    公开(公告)日:2021-04-22

    申请号:US17112041

    申请日:2020-12-04

    Abstract: A neural network apparatus configured to perform a deconvolution operation includes a memory configured to store a first kernel; and a processor configured to: obtain, from the memory, the first kernel; calculate a second kernel by adjusting an arrangement of matrix elements comprised in the first kernel; generate sub-kernels by dividing the second kernel; perform a convolution operation between an input feature map and the sub-kernels using a convolution operator; and generate an output feature map, as a deconvolution of the input feature map, by merging results of the convolution operation.

    NEURAL NETWORK METHOD AND APPARATUS
    37.
    发明申请

    公开(公告)号:US20200285887A1

    公开(公告)日:2020-09-10

    申请号:US16884232

    申请日:2020-05-27

    Abstract: A processor-implemented neural network method includes: obtaining, from a memory, data of an input feature map and kernels having a binary-weight, wherein the kernels are to be processed in a layer of a neural network; decomposing each of the kernels into a first type sub-kernel reconstructed with weights of a same sign, and a second type sub-kernel for correcting a difference between a respective kernel, among the kernels, and the first type sub-kernel; performing a convolution operation by using the input feature map and the first type sub-kernels and the second type sub-kernels decomposed from each of the kernels; and obtaining an output feature map by combining results of the convolution operation.

    METHOD AND SYSTEM OF PERFORMING CONVOLUTION IN NEURAL NETWORKS WITH VARIABLE DILATION RATE

    公开(公告)号:US20200218936A1

    公开(公告)日:2020-07-09

    申请号:US16733314

    申请日:2020-01-03

    Abstract: A method of performing convolution in a neural network with variable dilation rate is provided. The method includes receiving a size of a first kernel and a dilation rate, determining at least one of size of one or more disintegrated kernels based on the size of the first kernel, a baseline architecture of a memory and the dilation rate, determining an address of one or more blocks of an input image based on the dilation rate, and one or more parameters associated with a size of the input image and the memory. Thereafter, the one or more blocks of the input image and the one or more disintegrated kernels are fetched from the memory, and an output image is obtained based on convolution of each of the one or more disintegrated kernels and the one or more blocks of the input image.

    APPARATUS AND METHOD WITH NEURAL NETWORK
    39.
    发明申请

    公开(公告)号:US20190138891A1

    公开(公告)日:2019-05-09

    申请号:US16106902

    申请日:2018-08-21

    Abstract: A neural network apparatus includes a plurality of node buffers connected to a node lane and configured to store input node data by a predetermined bit size; a plurality of weight buffers connected to a weight lane and configured to store weights; and one or more processors configured to: generate first and second split data by splitting the input node data by the predetermined bit size, store the first and second split data in the node buffers, output the first split data to an operation circuit for a neural network operation on an index-by-index basis, shift the second split data, and output the second split data to the operation circuit on the index-by-index basis.

Patent Agency Ranking