-
公开(公告)号:US20240249204A1
公开(公告)日:2024-07-25
申请号:US18419476
申请日:2024-01-22
Applicant: Google LLC
Inventor: Jinsung Yoon , Jiefeng Chen , Sayna Ebrahimi , Sercan Omer Arik
IPC: G06N20/20
CPC classification number: G06N20/20
Abstract: A method includes obtaining a set of unlabeled test data samples and, for each respective initial training step, determining a first average output for each unlabeled test data sample using a deep ensemble. For each round of a plurality of rounds, the method includes selecting a subset of unlabeled test data samples based on the determined first average outputs, labeling each respective unlabeled in the subset of unlabeled test data samples, fine-tuning the deep ensemble model using the subset of labeled test data samples, and determining a second average output for each unlabeled test data sample using the fine-tuned deep ensemble model. The method also includes generating, using the set of unlabeled test data samples and the determined second average outputs, a pseudo-labeled set of training data samples. The method also includes training the deep ensemble model using the pseudo-labeled set of training data samples.
-
公开(公告)号:US20230274143A1
公开(公告)日:2023-08-31
申请号:US18173985
申请日:2023-02-24
Applicant: Google LLC
Inventor: Zizhao Zhang , Zifeng Wang , Chen-Yu Lee , Ruoxi Sun , Sayna Ebrahimi , Xiaoqi Ren , Guolong Su , Vincent Perot , Tomas Pfister , Han Zhang
IPC: G06N3/08
CPC classification number: G06N3/08
Abstract: A method for rehearsal-free continual learning includes obtaining a set of training samples where training sample in the set of training samples is associated with a respective task of a plurality of different tasks. The method includes obtaining a task-invariant prompt representative of learned knowledge common to each respective task of the plurality of different tasks. The method includes, for each respective task of the plurality of different tasks, obtaining a respective task-specific prompt representative of learned knowledge specific to the respective task. The method includes, during each of one or more training iterations, for each respective training sample in the set of training samples, selecting the respective task-specific prompt representative of the respective task of the respective training sample and training a model using the task-invariant prompt and the selected respective task-specific prompt.
-
公开(公告)号:US20240428015A1
公开(公告)日:2024-12-26
申请号:US18386343
申请日:2023-11-02
Applicant: Google LLC
Inventor: Jinsung Yoon , Jiefeng Chen , Sayna Ebrahimi , Sercan Omer Arik
IPC: G06F40/40
Abstract: Aspects of the disclosure are directed to methods, systems, and computer readable media for adaptation with self-evaluation to improve selective prediction in large language models (LLMs), generally referred to as ASPIRE. ASPIRE includes training LLMs on a portion of training data from a question answering task to learn self-evaluation, e.g., learn to distinguish whether a generated answer is correct or not. ASPIRE further includes a selection score that combines a likelihood of that generated answer is correct with a self-evaluation score for selective prediction. ASPIRE demonstrates improved selective prediction performance with less computational cost.
-
公开(公告)号:US20240386321A1
公开(公告)日:2024-11-21
申请号:US18639519
申请日:2024-04-18
Applicant: Google LLC
Inventor: Sayna Ebrahimi , Yihe Dong , Tomas Pfister , Sercan Omer Arik
IPC: G06N20/00
Abstract: Aspects of the disclosure are directed to a multimodal processing system for processing both structured and un-structured data. Real-world data is not always consistent in form or content. The multimodal processing system includes model that can be trained to account for this characteristic of real-world data, by selectively masking data of different modalities during pretraining to learn outputs that are the same or comparable between the masked and un-masked inputs. The model is trained according to modality-specific masking objectives computed for each modality of data and joint modality similarity-based masking objectives for a joint representation of the data across all modalities. The system provides consistent and accurate input, even when input data may have substantial portions of data from different modalities missing. Cross-modal relationships in data are reinforced by the model as different portions of data are masked, contributing to an overall increase in model accuracy versus other approaches.
-
公开(公告)号:US20230377359A1
公开(公告)日:2023-11-23
申请号:US18199129
申请日:2023-05-18
Applicant: Google LLC
Inventor: Sayna Ebrahimi , Sercan Omer Arik , Tomas Pfister
CPC classification number: G06V30/1912 , G06V30/19147 , G06V10/70
Abstract: An aspect of the disclosed technology comprises a test-time adaptation (“TTA”) technique for visual document understanding (“VDU”) tasks that uses self-supervised learning on different modalities (e.g., text and layout) by applying masked visual language modeling (“MVLM”) along with pseudo-labeling. In accordance with an aspect of the disclosed technology, the TTA technique enables a document model to adapt to domain or distribution shifts that are detected.
-
-
-
-