SYSTEM AND METHOD FOR TRAINING A SPARSE NEURAL NETWORK WHILST MAINTAINING SPARSITY

    公开(公告)号:US20230124177A1

    公开(公告)日:2023-04-20

    申请号:US17914035

    申请日:2021-06-04

    Abstract: A computer-implemented method of training a neural network. The method comprises repeatedly determining a forward-pass set of network parameters by selecting a first sub-set of parameters of the neural network and setting all other parameters of the forward-pass set of network parameters to zero. The method then processes a training data item using the neural network in accordance with the forward-pass set of network parameters to generate a neural network output, determines a value of an objective function from the neural network output and the training data item, selects a second sub-set of network parameters, determines a backward-pass set of network parameters comprising the first and second sub-sets of parameters, and updates parameters corresponding to the backward-pass set of network parameters using a gradient estimate determined from the value of the objective function.

    SCALABLE AND COMPRESSIVE NEURAL NETWORK DATA STORAGE SYSTEM

    公开(公告)号:US20250053780A1

    公开(公告)日:2025-02-13

    申请号:US18662972

    申请日:2024-05-13

    Abstract: A system for compressed data storage using a neural network. The system comprises a memory comprising a plurality of memory locations configured to store data; a query neural network configured to process a representation of an input data item to generate a query; an immutable key data store comprising key data for indexing the plurality of memory locations; an addressing system configured to process the key data and the query to generate a weighting associated with the plurality of memory locations; a memory read system configured to generate output memory data from the memory based upon the generated weighting associated with the plurality of memory locations and the data stored at the plurality of memory locations; and a memory write system configured to write received write data to the memory based upon the generated weighting associated with the plurality of memory locations.

    Gated attention neural networks
    15.
    发明授权

    公开(公告)号:US12033055B2

    公开(公告)日:2024-07-09

    申请号:US17763984

    申请日:2020-09-07

    CPC classification number: G06N3/044 G06N3/048 G06N3/08

    Abstract: A system including an attention neural network that is configured to receive an input sequence and to process the input sequence to generate an output is described. The attention neural network includes: an attention block configured to receive a query input, a key input, and a value input that are derived from an attention block input. The attention block includes an attention neural network layer configured to: receive an attention layer input derived from the query input, the key input, and the value input, and apply an attention mechanism to the query input, the key input, and the value input to generate an attention layer output for the attention neural network layer; and a gating neural network layer configured to apply a gating mechanism to the attention block input and the attention layer output of the attention neural network layer to generate a gated attention output.

    SCALABLE AND COMPRESSIVE NEURAL NETWORK DATA STORAGE SYSTEM

    公开(公告)号:US20210150314A1

    公开(公告)日:2021-05-20

    申请号:US17102318

    申请日:2020-11-23

    Abstract: A system for compressed data storage using a neural network. The system comprises a memory comprising a plurality of memory locations configured to store data; a query neural network configured to process a representation of an input data item to generate a query; an immutable key data store comprising key data for indexing the plurality of memory locations; an addressing system configured to process the key data and the query to generate a weighting associated with the plurality of memory locations; a memory read system configured to generate output memory data from the memory based upon the generated weighting associated with the plurality of memory locations and the data stored at the plurality of memory locations; and a memory write system configured to write received write data to the memory based upon the generated weighting associated with the plurality of memory locations.

    Neural Networks with Relational Memory

    公开(公告)号:US20210081795A1

    公开(公告)日:2021-03-18

    申请号:US17107621

    申请日:2020-11-30

    Abstract: A system including one or more computers and one or more storage devices storing instructions that when executed by the one or more computers cause the one or more computers to implement a memory and memory-based neural network is described. The memory is configured to store a respective memory vector at each of a plurality of memory locations in the memory. The memory-based neural network is configured to: at each of a plurality of time steps: receive an input; determine an update to the memory, wherein determining the update comprising applying an attention mechanism over the memory vectors in the memory and the received input; update the memory using the determined update to the memory; and generate an output for the current time step using the updated memory.

    Neural networks with relational memory

    公开(公告)号:US10853725B2

    公开(公告)日:2020-12-01

    申请号:US16415954

    申请日:2019-05-17

    Abstract: A system including one or more computers and one or more storage devices storing instructions that when executed by the one or more computers cause the one or more computers to implement a memory and memory-based neural network is described. The memory is configured to store a respective memory vector at each of a plurality of memory locations in the memory. The memory-based neural network is configured to: at each of a plurality of time steps: receive an input; determine an update to the memory, wherein determining the update comprising applying an attention mechanism over the memory vectors in the memory and the received input; update the memory using the determined update to the memory; and generate an output for the current time step using the updated memory.

Patent Agency Ranking