Invention Grant
- Patent Title: Parallel decoding using transformer models
-
Application No.: US16682611Application Date: 2019-11-13
-
Publication No.: US11681954B2Publication Date: 2023-06-20
- Inventor: Noam M. Shazeer , Jakob D. Uszkoreit , Mitchell Thomas Stern
- Applicant: Google LLC
- Applicant Address: US CA Mountain View
- Assignee: Google LLC
- Current Assignee: Google LLC
- Current Assignee Address: US CA Mountain View
- Agency: Fish & Richardson P.C.
- Main IPC: G06N20/20
- IPC: G06N20/20 ; G06N20/00 ; G06N7/00 ; G06F18/21 ; G06F18/22 ; G06V10/764 ; G06V10/82

Abstract:
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for performing parallel generation of output from an autoregressive sequence to sequence model. In one aspect, a blockwise parallel decoding method takes advantage of the fact that some architectures can score sequences in sublinear time. By generating predictions for multiple time steps at once then backing off to a longest prefix validated by the scoring model, the methods can substantially improve the speed of greedy decoding without compromising performance.
Public/Granted literature
- US20200082226A1 PARALLEL DECODING USING TRANSFORMER MODELS Public/Granted day:2020-03-12
Information query