Abstract:
A method and system for constructing a convolutional neural network (CNN) model are herein disclosed. The method includes regularizing spatial domain weights, providing quantization of the spatial domain weights, pruning small or zero weights in a spatial domain, fine-tuning a quantization codebook, compressing a quantization output from the quantization codebook, and decompressing the spatial domain weights and using either sparse spatial domain convolution and sparse Winograd convolution after pruning Winograd-domain weights.
Abstract:
A system and method for characterizing an interference demodulation reference signal (DMRS) in a piece of user equipment (UE), e.g., a mobile device. The UE determines whether the serving signal is transmitted in a DMRS-based transmission mode; if it is, the UE cancels the serving DMRS from the received signal; otherwise the UE cancels the serving data signal from the received signal. The remaining signal is then analyzed for the amount of power it has in each of four interference DMRS candidates, and hypothesis testing is performed to determine whether interference DMRS is present in the signal, and, if so, to determine the rank of the interference DMRS, and the port and scrambling identity of each of the interference DMRS layers.
Abstract:
A system and method for characterizing an interference demodulation reference signal (DMRS) in a piece of user equipment (UE), e.g., a mobile device. The UE determines whether the serving signal is transmitted in a DMRS-based transmission mode; if it is, the UE cancels the serving DMRS from the received signal; otherwise the UE cancels the serving data signal from the received signal. The remaining signal is then analyzed for the amount of power it has in each of four interference DMRS candidates, and hypothesis testing is performed to determine whether interference DMRS is present in the signal, and, if so, to determine the rank of the interference DMRS, and the port and scrambling identity of each of the interference DMRS layers.
Abstract:
A system and method for removing bias from a frequency estimate. A simulation is used to predict, for various values of the signal to noise ratio, a bias in a raw frequency estimate produced by a frequency estimation algorithm. A straight line is fit to simulated frequency offset estimates as a function of true frequency offset, and the reciprocal of the slope of the line is stored, as a multiplicative bias removal term, in a lookup table, for the simulated signal to noise ratio. In operation, the raw frequency estimate is multiplied by a multiplicative bias removal term, obtained from the lookup table, to form a corrected frequency offset estimate.
Abstract:
A method and apparatus for variable rate compression with a conditional autoencoder is herein provided. According to one embodiment, a method includes training a conditional autoencoder using a Lagrange multiplier and training a neural network that includes the conditional autoencoder with mixed quantization bin sizes.
Abstract:
Apparatuses and methods of manufacturing same, systems, and methods for performing network parameter quantization in deep neural networks are described. In one aspect, diagonals of a second-order partial derivative matrix (a Hessian matrix) of a loss function of network parameters of a neural network are determined and then used to weight (Hessian-weighting) the network parameters as part of quantizing the network parameters. In another aspect, the neural network is trained using first and second moment estimates of gradients of the network parameters and then the second moment estimates are used to weight the network parameters as part of quantizing the network parameters. In yet another aspect, network parameter quantization is performed by using an entropy-constrained scalar quantization (ECSQ) iterative algorithm. In yet another aspect, network parameter quantization is performed by quantizing the network parameters of all layers of a deep neural network together at once.
Abstract:
A method and apparatus for variable rate compression with a conditional autoencoder is herein provided. According to one embodiment, a method includes training a conditional autoencoder using a Lagrange multiplier and training a neural network that includes the conditional autoencoder with mixed quantization bin sizes.