-
公开(公告)号:US11868381B2
公开(公告)日:2024-01-09
申请号:US17215465
申请日:2021-03-29
Applicant: Google LLC
Inventor: Thomas Müller , Jonathan Herzig , Pawel Nowak , Julian Eisenschlos , Francesco Piccinno , Syrine Krichene
IPC: G06F16/332 , G06N3/08 , G06F40/20 , G06F40/284 , G06F40/35
CPC classification number: G06F16/3329 , G06F40/20 , G06F40/284 , G06F40/35 , G06N3/08
Abstract: Systems and methods for pre-training and fine-tuning of neural-network-based language models to reason directly over tables without generating logical forms. In some examples, a language model can be pre-trained using masked-language modeling tasks synthetically generated from tables pulled from a knowledge corpus. In some examples, the language model may be further pre-trained using pairs of counterfactual statements generated from those tables, and/or one or more statements that compare selected data from those tables. The language model may then be fine-tuned using examples that include only a question, an answer, and a table, allowing fine-tuning examples to be harvested directly from existing benchmark datasets or synthetically generated.
-
公开(公告)号:US20240386215A1
公开(公告)日:2024-11-21
申请号:US18319249
申请日:2023-05-17
Applicant: Google LLC
Inventor: Julian Martin Eisenschlos , Francesco Piccinno , Yasemin Altun , Syrine Krichene , Kenton Chiu Tsun Lee , Fangyu Liu , Mandar Joshi , Chenxi Pang , Wenhu Chen
Abstract: Provided is a one-shot solution to visual language reasoning. Example systems described herein decompose the challenge of visual language reasoning into two steps: translation of a graphical depiction of data (e.g., a plot or chart) into text; followed by reasoning over the translated text. In particular, example systems described herein can include a machine-learned visual-to-language conversion model that translates a graphical depiction of a dataset to a set of text descriptive of the dataset. The output of visual-to-language conversion model can then be directly used to prompt a language model, (e.g., a pretrained large language model (LLM)), exploiting the few-shot reasoning capabilities of the language model.
-
公开(公告)号:US20240086436A1
公开(公告)日:2024-03-14
申请号:US18513981
申请日:2023-11-20
Applicant: Google LLC
Inventor: Thomas Müller , Jonathan Herzig , Pawel Nowak , Julian Eisenschlos , Francesco Piccinno , Syrine Krichene
IPC: G06F16/332 , G06F40/20 , G06F40/284 , G06F40/35 , G06N3/08
CPC classification number: G06F16/3329 , G06F40/20 , G06F40/284 , G06F40/35 , G06N3/08
Abstract: Systems and methods for pre-training and fine-tuning of neural-network-based language models to reason directly over tables without generating logical forms. In some examples, a language model can be pre-trained using masked-language modeling tasks synthetically generated from tables pulled from a knowledge corpus. In some examples, the language model may be further pre-trained using pairs of counterfactual statements generated from those tables, and/or one or more statements that compare selected data from those tables. The language model may then be fine-tuned using examples that include only a question, an answer, and a table, allowing fine-tuning examples to be harvested directly from existing benchmark datasets or synthetically generated.
-
公开(公告)号:US20220309087A1
公开(公告)日:2022-09-29
申请号:US17215465
申请日:2021-03-29
Applicant: Google LLC
Inventor: Thomas Müller , Jonathan Herzig , Pawel Nowak , Julian Eisenschlos , Francesco Piccinno , Syrine Krichene
IPC: G06F16/332 , G06F40/20 , G06N3/08
Abstract: Systems and methods for pre-training and fine-tuning of neural-network-based language models to reason directly over tables without generating logical forms. In some examples, a language model can be pre-trained using masked-language modeling tasks synthetically generated from tables pulled from a knowledge corpus. In some examples, the language model may be further pre-trained using pairs of counterfactual statements generated from those tables, and/or one or more statements that compare selected data from those tables. The language model may then be fine-tuned using examples that include only a question, an answer, and a table, allowing fine-tuning examples to be harvested directly from existing benchmark datasets or synthetically generated.
-
-
-