Abstract:
Disclosed herein is convolutional neural network (CNN) system for generating a classification for an input image. According to an embodiment, the CNN system comprises a sequence of neural network layers configured to: derive a feature map based on at least the input image; puncture at least one selection among the feature map and a kernel by setting the value of one or more elements of a row of the at least one selection to zero according to a pattern and cyclic shifting the pattern by a predetermined interval per row to set the value of one or more elements of the rest of the rows of the at least one selection according to the cyclic shifted pattern; convolve the feature map with the kernel to generate a first convolved output; and generate the classification for the input image based on at least the first convolved output.
Abstract:
A electronic system includes: a support chip configured to receive an input code stream; a circular Viterbi mechanism, coupled to the support chip, configured to: generate a final path metric for the input code stream, store intermediate path metrics at the repetition depth, generate a repetition path metric for the input code stream, and calculate a soft correlation metric based on the final path metric, the repetition path metric, and the intermediate path metrics.
Abstract:
An apparatus and a method. The apparatus includes a receiver to receive a polar codeword of length mj; a processor configured to determine a decoding node tree structure with mj leaf nodes for the received codeword, and receive i indicating a level at which parallelism of order m is applied to the decoding node tree structure, wherein i indicates levels of the decoding node tree structure, and wherein the mj leaf nodes are at level j; and m successive cancellation list decoders (SCLDs) applied to each child node of each node in the decoding node tree structure at level i−1, wherein each of the m SCLDs executes in parallel to determine log likelihood ratios (LLRs) for a codeword of length mj−i, and wherein each of the m SCLDs uses LLRs of an associated parent node without using a hard decision or a soft reliability estimate of any other node of the other m SCLDs.
Abstract:
A computing system includes: a communication unit configured to: determine a relaxed coding profile including a polar-processing range for processing content data over a bit channel; process the content data based on a total polarization level being within the polar-processing range, the polar-processing range for controlling a polar processing mechanism or a portion therein corresponding to the bit channel for the content data; and an inter-device interface, coupled to the communication unit, configured to communicate the content data.
Abstract:
A electronic system includes: a support chip configured to receive an input code stream; a circular Viterbi mechanism, coupled to the support chip, configured to: generate a final path metric for the input code stream, store intermediate path metrics at the repetition depth, generate a repetition path metric for the input code stream, and calculate a soft correlation metric based on the final path metric, the repetition path metric, and the intermediate path metrics.
Abstract:
A concatenated encoder is provided that includes an outer encoder, a symbol interleaver and a polar inner encoder. The outer encoder is configured to encode a data stream using an outer code to generate outer codewords. The symbol interleaver is configured to interleave symbols of the outer codewords and generate a binary stream. The polar inner encoder is configured to encode the binary stream using a polar inner code to generate an encoded stream. A concatenated decoder is provided that includes a polar inner decoder, a symbol de-interleaver and an outer decoder. The polar inner decoder is configured to decode an encoded stream using a polar inner code to generate a binary stream. The symbol de-interleaver is configured to de-interleave symbols in the binary stream to generate outer codewords. The outer decoder is configured to decode the outer codewords using an outer code to generate a decoded stream.
Abstract:
Methods and devices are provided for performing federated learning. A global model is distributed from a server to a plurality of client devices. At each of the plurality of client devices: model inversion is performed on the global model to generate synthetic data; the global model is on an augmented dataset of collected data and the synthetic data to generate a respective client model; and the respective client model is transmitted to the server. At the server: client models are received from the plurality of client devices, where each client model is received from a respective client device of the plurality of client devices; model inversion is performed on each client model to generate a synthetic dataset; the client models are averaged to generate an averaged model; and the averaged model is trained using the synthetic dataset to generate an updated model.
Abstract:
Apparatuses and methods of manufacturing same, systems, and methods for performing network parameter quantization in deep neural networks are described. In one aspect, multi-dimensional vectors representing network parameters are constructed from a trained neural network model. The multi-dimensional vectors are quantized to obtain shared quantized vectors as cluster centers, which are fine-tuned. The fine-tuned and shared quantized vectors/cluster centers are then encoded. Decoding reverses the process.
Abstract:
A system and a method are disclosed for tuning parameters of a large language model. The method comprises identifying first weights of a machine learning (ML) model. Second weights are received from a client device. The second weights may be based on updating, by the client device, the first weights. An update matrix may be generated based on the second weights. The update matrix may be decomposed into first decomposition matrices. Singular values that satisfy a criterion may be identified based on the first decomposition matrices. Singular vectors may be identified based on the singular values. Second decomposition matrices may be identified based on the singular vectors. Updates may be received from the client device of third weights associated with the second decomposition matrices. An updated ML model may be generated based on the updates of the third weights. An inference may be generated based on the updated ML model.
Abstract:
A method and system for providing Gaussian weighted self-attention for speech enhancement are herein provided. According to one embodiment, the method includes receiving an input noise signal, generating a score matrix based on the received input noise signal, and applying a Gaussian weighted function to the generated score matrix.