-
1.
公开(公告)号:US20240194192A1
公开(公告)日:2024-06-13
申请号:US18078782
申请日:2022-12-09
Applicant: GOOGLE LLC
Inventor: Ehsan Amid , Rajiv Mathews , Shankar Kumar , Jared Lichtarge , Mingqing Chen , Tien-Ju Yang , Yuxin Ding
CPC classification number: G10L15/16 , G10L15/063
Abstract: Information can be distilled from a global automatic speech recognition (ASR) model to a client ASR model. Many implementations include using an RNN-T model as the ASR model, where the global ASR model includes a global encoder, a joint network, a prediction network, and where the client ASR model includes a client encoder, the joint network, and the prediction network. Various implementations include using principal component analysis (PCA) while training the global ASR model to learn a mean vector and a set of principal components corresponding to the global ASR model. Additional or alternative implementations include training the client ASR model to generate one or more predicted coefficients of the global ASR model.
-
公开(公告)号:US12260875B2
公开(公告)日:2025-03-25
申请号:US18609362
申请日:2024-03-19
Applicant: Google LLC
Inventor: Ehsan Amid , Om Dipakbhai Thakkar , Rajiv Mathews , Francoise Beaufays
IPC: G10L21/0332 , G10L15/06 , G10L15/08 , G10L21/10
Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
-
公开(公告)号:US11955134B2
公开(公告)日:2024-04-09
申请号:US17643848
申请日:2021-12-13
Applicant: Google LLC
Inventor: Ehsan Amid , Om Thakkar , Rajiv Mathews , Francoise Beaufays
IPC: G10L21/0332 , G10L15/06 , G10L15/08 , G10L21/10
CPC classification number: G10L21/0332 , G10L15/063 , G10L15/08 , G10L21/10
Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
-
公开(公告)号:US20230044078A1
公开(公告)日:2023-02-09
申请号:US17816197
申请日:2022-07-29
Applicant: Google LLC
Inventor: Abhishek Kumar , Ehsan Amid
IPC: G06N20/00
Abstract: A method includes receiving training data for a machine learning model, the training data comprising a plurality of training examples and a corresponding plurality of labels. The method further includes dividing the training data into a plurality of training batches. For each training batch of the plurality of training batches, the method additionally includes learning a weight for each training example in the training batch that minimizes a sum of weighted losses for the training batch subject to a divergence constraint, where the divergence constraint limits a divergence of the learned weights for the training batch from a reference distribution, where the divergence is determined according to a chosen divergence measure. The method also includes training the machine learning model with each training batch of the plurality of training batches using the learned weight for each training example in the training batch. The method additionally includes providing the trained machine learning model.
-
公开(公告)号:US20250005453A1
公开(公告)日:2025-01-02
申请号:US18710814
申请日:2022-12-12
Applicant: Google LLC
Inventor: Ehsan Amid , Christopher James Fifty , Manfred Klaus Warmuth , Rohan Anil
IPC: G06N20/00
Abstract: Provided is an approach for knowledge distillation based on exporting Principal Components approximations (e.g., Bregman representations) of one or more layer-wise representations of the teacher model. In particular, the present disclosure provides an extension to the original Bregman PCA formulation by incorporating a mean vector and orthonormalizing the principal directions with respect to the geometry of the local convex function around the mean. This extended formulation allows viewing the learned representation as a dense layer, thus casting the problem as learning the linear coefficients of the compressed examples, as the input to this layer, by the student network. Example empirical data indicates that example implementations of the approach improve performance when compared to typical teacher-student training using soft labels.
-
公开(公告)号:US20240233707A9
公开(公告)日:2024-07-11
申请号:US18488578
申请日:2023-10-17
Applicant: Google LLC
Inventor: Tien-Ju Yang , You-Chi Cheng , Shankar Kumar , Jared Lichtarge , Ehsan Amid , Yuxin Ding , Rajiv Mathews , Mingqing Chen
IPC: G10L15/06 , G10L15/197 , G10L15/30
CPC classification number: G10L15/063 , G10L15/197 , G10L15/30 , G10L2015/0635
Abstract: A method includes receiving distillation data including a plurality of out-of-domain training utterances. For each particular out-of-domain training utterance of the distillation data, the method includes generating a corresponding augmented out-of-domain training utterance, and generating, using a teacher ASR model trained on training data corresponding to a target domain, a pseudo-label corresponding to the corresponding augmented out-of-domain training utterance. The method also includes distilling a student ASR model from the teacher ASR model by training the student ASR model using the corresponding augmented out-of-domain training utterances paired with the corresponding pseudo-labels generated by the teacher ASR model.
-
公开(公告)号:US20240135918A1
公开(公告)日:2024-04-25
申请号:US18488578
申请日:2023-10-16
Applicant: Google LLC
Inventor: Tien-Ju Yang , You-Chi Cheng , Shankar Kumar , Jared Lichtarge , Ehsan Amid , Yuxin Ding , Rajiv Mathews , Mingqing Chen
IPC: G10L15/06 , G10L15/197 , G10L15/30
CPC classification number: G10L15/063 , G10L15/197 , G10L15/30 , G10L2015/0635
Abstract: A method includes receiving distillation data including a plurality of out-of-domain training utterances. For each particular out-of-domain training utterance of the distillation data, the method includes generating a corresponding augmented out-of-domain training utterance, and generating, using a teacher ASR model trained on training data corresponding to a target domain, a pseudo-label corresponding to the corresponding augmented out-of-domain training utterance. The method also includes distilling a student ASR model from the teacher ASR model by training the student ASR model using the corresponding augmented out-of-domain training utterances paired with the corresponding pseudo-labels generated by the teacher ASR model.
-
公开(公告)号:US20220253713A1
公开(公告)日:2022-08-11
申请号:US17666488
申请日:2022-02-07
Applicant: Google LLC
Inventor: Ehsan Amid , Manfred Klaus Warmuth , Rohan Anil
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network using local layer-wise losses.
-
公开(公告)号:US20240249193A1
公开(公告)日:2024-07-25
申请号:US18417947
申请日:2024-01-19
Applicant: Google LLC
Inventor: Jared Alexander Lichtarge , Rajiv Mathews , Rohan Anil , Ehsan Amid , Shankar Kumar
IPC: G06N20/00
CPC classification number: G06N20/00
Abstract: Generally, the present disclosure is directed to enhanced federated learning (FL) that employs a set of clients with varying amounts of computational resources (e.g., system memory, storage, and processing bandwidth). To overcome limitations of conventional FL methods that employ a set of clients with varying amounts of computational resources, the embodiments run multi-directional knowledge distillation between the server models produced by each federated averaging (FedAvg) pool, using unlabeled server data as the distillation dataset. By co-distilling the two (or more) models frequently over the course of FedAvg rounds, information is shared between the pools without sharing model parameters. This leads to increased performance and faster convergence (in fewer federated rounds).
-
公开(公告)号:US20240221772A1
公开(公告)日:2024-07-04
申请号:US18609362
申请日:2024-03-19
Applicant: Google LLC
Inventor: Ehsan Amid , Om Dipakbhai Thakkar , Rajiv Mathews , Francoise Beaufays
IPC: G10L21/0332 , G10L15/06 , G10L15/08 , G10L21/10
CPC classification number: G10L21/0332 , G10L15/063 , G10L15/08 , G10L21/10
Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
-
-
-
-
-
-
-
-
-