METHOD AND APPARATUS WITH NEURAL NETWORK TRAINING

    公开(公告)号:US20240232632A1

    公开(公告)日:2024-07-11

    申请号:US18343303

    申请日:2023-06-28

    CPC classification number: G06N3/084 G06N3/045

    Abstract: A processor-implemented method may include generating respective first neural network differential data by differentiating a respective output of each layer of a first neural network with respect to input data provided to the first neural network that estimates output data from the input data, by a forward propagation process of the first neural network, generating, using a second neural network, an output differential value of the output data with respect to the input data using the respective first neural network differential data, and training the first neural network and the second neural network based on ground truth data of the output data and ground truth data of the output differential value.

    METHOD AND APPARATUS WITH NEURAL NETWORK PARAMETER QUANTIZATION

    公开(公告)号:US20230017432A1

    公开(公告)日:2023-01-19

    申请号:US17950342

    申请日:2022-09-22

    Abstract: A processor-implemented neural network method includes: determining a respective probability density function (PDF) of normalizing a statistical distribution of parameter values, for each channel of each of a plurality of feature maps of a pre-trained neural network; determining, for each channel, a corresponding first quantization range for performing quantization of corresponding parameter values, based on a quantization error and a quantization noise of the respective determined PDF; determining, for each channel, a corresponding second quantization range, based on a signal-to-quantization noise ratio (SQNR) of the respective determined PDF; correcting, for each channel, the corresponding first quantization range based on the corresponding second quantization range; and generating a quantized neural network, based on the corrected first quantization range corresponding for each channel.

    METHOD AND APPARATUS WITH DISTRIBUTED TRAINING OF NEURAL NETWORK

    公开(公告)号:US20230169333A1

    公开(公告)日:2023-06-01

    申请号:US17862881

    申请日:2022-07-12

    CPC classification number: G06N3/08

    Abstract: Disclosed are a training method and apparatus for distributed training of a neural network, the training apparatus including processors configured to perform distributed training, wherein each of the processors is further configured to perform a forward direction operation for layers of the neural network, determine a loss of the neural network based on the forward direction operation, determine a local gradient for each layer of the neural network by performing a backward direction operation for the layers of the neural network based on the loss, determine whether to perform gradient clipping for a local gradient determined for a previous layer, in response to determining a local gradient for a current layer through the backward direction operation, determine an aggregated gradient based on the backward direction operation and the gradient clipping performed by each of the processors, and update parameters of the neural network based on the aggregated gradient.

    METHOD AND APPARATUS WITH DATA LOADING

    公开(公告)号:US20230140239A1

    公开(公告)日:2023-05-04

    申请号:US17868361

    申请日:2022-07-19

    Abstract: A processor-implemented method with data loading includes: dividing a training data set into a plurality of subsets based on sizes of a plurality of data files included in the training data set; loading, from each of the plurality of subsets, a portion of data files in the subset to a plurality of processors based on a proportion of a number of data files of the plurality of subsets in the subset and a batch size of distributed training; and reallocating, based on sizes of data files loaded to processors in a same group among the plurality of processors, the loaded data files to the processors in the same group.

    METHOD AND APPARATUS FOR QUANTIZING PARAMETERS OF NEURAL NETWORK

    公开(公告)号:US20220092384A1

    公开(公告)日:2022-03-24

    申请号:US17192048

    申请日:2021-03-04

    Abstract: A method of quantizing parameters of a neural network includes acquiring a parameter of a floating-point format used in a process of inferring by the neural network, quantizing, based on statistics of a weight included in the parameter, the weight into a fixed-point format, determining, based on statistics of an activation of one or more layers configuring the neural network included in the parameter, a dynamic range of the activation, and quantizing, based on statistics of input data of the neural network, the input data into a fixed-point format.

    METHOD AND APPARATUS WITH NEURAL NETWORK PARAMETER QUANTIZATION

    公开(公告)号:US20210201117A1

    公开(公告)日:2021-07-01

    申请号:US16909095

    申请日:2020-06-23

    Abstract: A processor-implemented neural network method includes: determining a respective probability density function (PDF) of normalizing a statistical distribution of parameter values, for each channel of each of a plurality of feature maps of a pre-trained neural network; determining, for each channel, a corresponding first quantization range for performing quantization of corresponding parameter values, based on a quantization error and a quantization noise of the respective determined PDF; determining, for each channel, a corresponding second quantization range, based on a signal-to-quantization noise ratio (SQNR) of the respective determined PDF; correcting, for each channel, the corresponding first quantization range based on the corresponding second quantization range; and generating a quantized neural network, based on the corrected first quantization range corresponding for each channel.

    METHOD AND DEVICE WITH INFERENCE-BASED DIFFERENTIAL CONSIDERATION

    公开(公告)号:US20230306262A1

    公开(公告)日:2023-09-28

    申请号:US18187030

    申请日:2023-03-21

    CPC classification number: G06N3/08 G06N3/0464

    Abstract: A processor-implemented method is provided. The method includes, for each layer of a plurality of layers of a neural network for an input data provided to the neural network, obtain activation data of a corresponding layer of the plurality of layers, resulting from an inference operation of the corresponding layer; generate differential data of the activation data of the corresponding layer with respect to input data; and generate differential data of output data of the neural network with respect to the input data, based on the generated differential data of each layer.

Patent Agency Ranking