PROPAGATING ATTENTION INFORMATION IN EFFICIENT MACHINE LEARNING MODELS

    公开(公告)号:US20240160896A1

    公开(公告)日:2024-05-16

    申请号:US18335685

    申请日:2023-06-15

    CPC classification number: G06N3/0455

    Abstract: Certain aspects of the present disclosure provide techniques and apparatus for improved attention-based machine learning. A first attention propagation output is generated using a first transformer block of a plurality of transformer blocks, this generation including processing input data for the first transformer block using a first self-attention sub-block of the first transformer block. The first attention propagation output is propagated to a second transformer block of the plurality of transformer blocks. An output for the second transformer block is generated, this generation including generating output features for the second transformer block based on the first attention propagation output.

    MULTIDIMENSIONAL SPACE DECOMPOSITION FOR TRANSFORMER NEURAL NETWORKS

    公开(公告)号:US20240330662A1

    公开(公告)日:2024-10-03

    申请号:US18193234

    申请日:2023-03-30

    CPC classification number: G06N3/0499

    Abstract: Certain aspects of the present disclosure provide techniques and apparatus for processing multidimensional content using neural networks. An example method generally includes decomposing a multidimensional input into a plurality of two-dimensional subspaces, wherein the plurality of two-dimensional subspaces share a common dimension. A first attention matrix is generated based on a projection of tokens in a first two-dimensional subspace of the plurality of two-dimensional subspaces via an attention block of a transformer neural network, and a second attention matrix is generated based on a projection of tokens in a second two-dimensional subspace of the plurality of two-dimensional subspaces via the attention block of the transformer neural network. An output of the transformer neural network is generated based on a combination of the first attention matrix and the second attention matrix.

    EFFICIENT SELF-ATTENTION FOR VIDEO PROCESSING

    公开(公告)号:US20220301311A1

    公开(公告)日:2022-09-22

    申请号:US17696797

    申请日:2022-03-16

    Abstract: A processor-implemented method for processing a video includes receiving the video as an input at an artificial neural network (ANN). The video includes a sequence of frames. A set of features of a current frame of the video and a prior frame of the video are extracted. The set of features including a set of support features for a set of pixels of the prior frame to be aligned with a set of reference features of the current frame. A similarity between a support feature for each pixel in the set of pixels of the set of support features of the prior frame and a corresponding reference feature of the current frame is computed. An attention map is generated based on the similarity. An output including a reconstruction of the current frame is generated based on the attention map.

Patent Agency Ranking