Invention Grant
- Patent Title: Attention neural networks with sparse attention mechanisms
-
Application No.: US17341193Application Date: 2021-06-07
-
Publication No.: US11238332B2Publication Date: 2022-02-01
- Inventor: Joshua Timothy Ainslie , Santiago Ontañón , Philip Pham , Manzil Zaheer , Guru Guruganesh , Kumar Avinava Dubey , Amr Ahmed
- Applicant: Google LLC
- Applicant Address: US CA Mountain View
- Assignee: Google LLC
- Current Assignee: Google LLC
- Current Assignee Address: US CA Mountain View
- Agency: Fish & Richardson P.C.
- Main IPC: G06N3/04
- IPC: G06N3/04 ; G06N3/08 ; G06N3/063 ; G06N20/00

Abstract:
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing network inputs using an attention neural network that has one or more sparse attention sub-layers. Each sparse attention sub-layer is configured to apply a sparse attention mechanism that attends differently for input positions that are in a first proper subset of the input positions in the input to the sub-layer than for positions that are not in the first proper subset.
Public/Granted literature
- US20210383191A1 ATTENTION NEURAL NETWORKS WITH SPARSE ATTENTION MECHANISMS Public/Granted day:2021-12-09
Information query