Abstract:
A neural network accelerator in which processing elements are configured in a systolic array structure includes a memory to store a plurality of feature data including first and second feature data and a plurality of kernel data including first and second kernel data, a first processing element to perform an operation based on the first feature data and the first kernel data and output the first feature data, a selection circuit to select one of the first feature data and the second feature data, based on a control signal, and output the selected feature data, a second processing element to perform an operation based on the selected feature data and one of the first and the second kernel data, and a controller to generate the control signal, based on a neural network characteristic associated with the plurality of feature data and kernel data.
Abstract:
In the present invention, by providing an apparatus for processing a convolutional neural network (CNN), including a weight memory configured to store a first weight group of a first layer, a feature map memory configured to store an input feature map where the first weight group is to be applied, an address generator configured to determine a second position spaced from a first position of a first input pixel of the input feature map based on a size of the first weight group, and determine a plurality of adjacent pixels adjacent to the second position; and a processor configured to apply the first weight group to the plurality of adjacent pixels to obtain a first output pixel corresponding to the first position, a memory space may be efficiently used by saving the memory space.
Abstract:
A reorganizable neural network computing device is provided. The computing device includes a data processing array unit including a plurality of operators disposed at locations corresponding to a row and a column. One or more chaining paths which transfer the first input data from the operator of the first row of the data processing array to the operator of the second row are optionally formed. The plurality of first data input processors of the computing device transfer the first input data for a layer of the neural network to the operators along rows of the data processing array unit, and the plurality of second data input processors of the computing device transfer the second input data to the operators along the columns of the data processing array.
Abstract:
Provided is an operation method of a convolution circuit. The method includes receiving input feature maps, generating output feature maps corresponding to the respective input feature maps through convolution operations for performing parallel processing with a kernel unit, and outputting the output feature maps to an external memory.
Abstract:
Provided is a neural network accelerator which performs a calculation of a neural network provided with layers, the neural network accelerator including a kernel memory configured to store kernel data related to a filter, a feature map memory configured to store feature map data which are outputs of the layers, and a Processing Element (PE) array including PEs arranged along first and second directions, wherein each of the PEs performs a calculation using the feature map data transmitted in the first direction from the feature map memory and the kernel data transmitted in the second direction from the kernel memory, and transmits a calculation result to the feature map memory in a third direction opposite to the first direction.
Abstract:
There are provided a method and apparatus for controlling an electronic device. The apparatus for controlling an electronic device includes a marker recognition unit configured to recognize the marker of the electronic device and a control unit configured to perform communication with the electronic device based on the marker and control the electronic device using a Graphic User Interface (GUI) program received from the electronic device.
Abstract:
Disclosed is a neural network accelerator including a maximum value determiner outputting a maximum value based on a first magnitude component corresponding to first input data and a second magnitude component corresponding to second input data, a sign determiner outputting a sign component corresponding to the maximum value among a first sign component corresponding to the first input data and a second sign component corresponding to the second input data, as an output sign component, an offset operator quantizing a difference between the first magnitude component and the second magnitude component and outputting an output offset based on the first sign component, the second sign component, and the quantization result, and a magnitude operator calculating an output magnitude component of an output data based on the maximum value and the output offset. Each of the first input data and the second input data is data on a logarithm domain.
Abstract:
Provided is a method and system of transmitting and receiving a three-dimensional (3D) broadcasting service, and more particularly, to a method and system for transmitting and receiving a reference image and a 3D auxiliary image for the 3D broadcasting service through a digital broadcasting network in real-time or in non-real time. According to the present invention, a smooth broadcasting service may be provided to terminals receiving a two-dimensional (2D) or 3D broadcast.