- 专利标题: QUANTIZING NEURAL NETWORKS USING SHIFTING AND SCALING
-
申请号: US16596177申请日: 2019-10-08
-
公开(公告)号: US20210034982A1公开(公告)日: 2021-02-04
- 发明人: Eric A. Sather , Steven L. Teig
- 申请人: Perceive Corporation
- 申请人地址: US CA San Jose
- 专利权人: Perceive Corporation
- 当前专利权人: Perceive Corporation
- 当前专利权人地址: US CA San Jose
- 主分类号: G06N3/08
- IPC分类号: G06N3/08 ; G06N3/04 ; G06N3/063 ; G06N5/04 ; G06N20/00 ; G06F7/483
摘要:
Some embodiments of the invention provide a novel method for training a quantized machine-trained network. Some embodiments provide a method of scaling a feature map of a pre-trained floating-point neural network in order to match the range of output values provided by quantized activations in a quantized neural network. A quantization function is modified, in some embodiments, to be differentiable to fix the mismatch between the loss function computed in forward propagation and the loss gradient used in backward propagation. Variational information bottleneck, in some embodiments, is incorporated to train the network to be insensitive to multiplicative noise applied to each channel. In some embodiments, channels that finish training with large noise, for example, exceeding 100%, are pruned.
公开/授权文献
- US11847568B2 Quantizing neural networks using shifting and scaling 公开/授权日:2023-12-19
信息查询