-
公开(公告)号:US12112131B2
公开(公告)日:2024-10-08
申请号:US17588043
申请日:2022-01-28
Applicant: Salesforce, Inc.
Inventor: Benjamin Newman , Nazneen Rajani , Prafulla Kumar Choubey
IPC: G06F40/30 , G06F3/08 , G06F40/126 , G06F40/279 , G06N3/044
CPC classification number: G06F40/279 , G06F40/126 , G06N3/044
Abstract: Embodiments described herein provide a system and method for extracting factual information. The system transforms a query into a natural language prompt in a format of a query subject and a queried relation. The system encodes, via an embedding layer of a pre-trained language model, the natural language prompt into a first embedding. The system encodes, via the adapter model, the first embedding into a second embedding based on a probability that the second embedding returns the factual information when the second embedding is fed the first attention layer of the pre-trained language model. The system decodes, by the first attention layer of the pre-trained language mode, the second embedding into a response to the query. The system extracts the factual information from the decoded response to the query.
-
公开(公告)号:US12204847B2
公开(公告)日:2025-01-21
申请号:US17938572
申请日:2022-10-06
Applicant: Salesforce, Inc.
Inventor: Alexander R. Fabbri , Prafulla Kumar Choubey , Jesse Vig , Chien-Sheng Wu , Caiming Xiong
IPC: G06F17/00 , G06F40/166 , G06F40/284 , G06N20/00
Abstract: Embodiments described herein provide a method for text summarization. The method includes receiving a training dataset having at least an uncompressed text, a compressed text, and one or more information entities accompanying the compressed text. The method also includes generating, using a perturber model, a perturbed text with the one or more information entities being inserted into the compressed text. The method further includes training the perturber model based on a first training objective, and generating, using the trained perturber model, a perturbed summary in response to an input of a reference summary. The method further includes generating, via an editor model, a predicted summary by removing information from the perturbed summary conditioned on a source document of the reference summary, and training the editor model based on a second training objective.
-
公开(公告)号:US20240411992A1
公开(公告)日:2024-12-12
申请号:US18335898
申请日:2023-06-15
Applicant: Salesforce, Inc.
Inventor: Shiva Kumar Pentyala , Prafulla Kumar Choubey , Shashank Harinath , Sitaram Asur , Chien-Sheng Jason Wu , Zachary Alexander , Caiming Xiong
IPC: G06F40/284 , G06N3/08
Abstract: Embodiments described herein provide a training framework for generative NLP models. Specifically, the training input, e.g., in the form of a sequence of tokens representing a user-agent dialogue, may be randomly masked for a few spans, which can be one or more tokens, one or more words, one or more sentences, or one or more paragraphs. These masked spans are replaced with their embeddings generated from pre-trained large language models are then used for training the NLP model.
-
公开(公告)号:US20240411991A1
公开(公告)日:2024-12-12
申请号:US18330216
申请日:2023-06-06
Applicant: Salesforce, Inc.
Inventor: Shiva Kumar Pentyala , Prafulla Kumar Choubey , Shashank Harinath , Sitaram Asur , Chien-Sheng Jason Wu , Zachary Alexander , Caiming Xiong
IPC: G06F40/284
Abstract: Embodiments described herein provide a training framework for generative NLP models that operate on previously learnt knowledge from pretrained large language models. Specifically, to train an NLP model to generate a response to a user utterance (e.g., “resolve login issue”), document embeddings of support IT documents encoded by a pretrained LLM are fed to an NLP decoder together with a training dialogue (e.g., a dialogue between the chat agent on how to “resolve login issue”). The NLP decoder can thus be trained by a causal language modeling loss computed based on the predicted next token and the ground-truth token from the training dialogue.
-
5.
公开(公告)号:US20240070394A1
公开(公告)日:2024-02-29
申请号:US18160967
申请日:2023-01-27
Applicant: Salesforce, Inc.
Inventor: Xiangyu Peng , Chen Xing , Prafulla Kumar Choubey , Chieng-Sheng Wu
IPC: G06F40/284 , G06F40/40
CPC classification number: G06F40/284 , G06F40/40
Abstract: Embodiments described herein provide a mechanism that ensembles trainable soft prompts to transfer knowledge from source tasks under few-shot learning settings. Specifically, given a source task input from a source task training dataset, a set of soft prompts may be trained using a frozen PLM on the large-scale source task training dataset. The set of soft prompts are then prepended to a target task input, based on which the frozen pre-trained language model generates a set of logits for predicting classification of the target task input, respectively. An attention module is used to generate input-logit attention scores, which are used to compute a weighted linear combination of the logits given the attention scores. The weighted linear combination are the final logits to predict the final classification of the target task input.
-
公开(公告)号:US20230419017A1
公开(公告)日:2023-12-28
申请号:US17938572
申请日:2022-10-06
Applicant: Salesforce, Inc.
Inventor: Alexander R. Fabbri , Prafulla Kumar Choubey , Jesse Vig , Chien-Sheng Wu , Caiming Xiong
IPC: G06F40/166 , G06F40/284 , G06N20/00
CPC classification number: G06F40/166 , G06F40/284 , G06N20/00
Abstract: Embodiments described herein provide a method for text summarization. The method includes receiving a training dataset having at least an uncompressed text, a compressed text, and one or more information entities accompanying the compressed text. The method also includes generating, using a perturber model, a perturbed text with the one or more information entities being inserted into the compressed text. The method further includes training the perturber model based on a first training objective, and generating, using the trained perturber model, a perturbed summary in response to an input of a reference summary. The method further includes generating, via an editor model, a predicted summary by removing information from the perturbed summary conditioned on a source document of the reference summary, and training the editor model based on a second training objective.
-
7.
公开(公告)号:US20230376677A1
公开(公告)日:2023-11-23
申请号:US17880502
申请日:2022-08-03
Applicant: Salesforce, Inc.
Inventor: Prafulla Kumar Choubey , Alexander R. Fabbri , Jesse Vig , Chien-Sheng Wu , Wenhao Liu , Nazneen Rajani
IPC: G06F40/166 , G06N20/00
CPC classification number: G06F40/166 , G06N20/00
Abstract: Embodiments described herein provide a document summarization framework that employs an ensemble of summarization models, each of which is a modified version of a base summarization model to control hallucination. For example, a base summarization model may first be trained on a full training data set. The trained base summarization model is then fine-tuned using a first filtered subset of the training data which contains noisy data, resulting in an “anti-expert” model. The parameters of the anti-expert model are subtracted from the parameters of the trained base model to produce a final summarization model which yields robust factual performance.
-
公开(公告)号:US20230334245A1
公开(公告)日:2023-10-19
申请号:US17889166
申请日:2022-08-16
Applicant: Salesforce, Inc.
Inventor: Prafulla Kumar Choubey , Yu Bai , Nazneen Rajani , Wenhao Liu
IPC: G06F40/284 , G06F40/40
CPC classification number: G06F40/284 , G06F40/40
Abstract: Embodiments described herein provide a Conformal Predictor (CP) that reduces the number of likely target class labels CP. Specifically, the CP provides a model agnostic framework to generate a label set, instead of a single label prediction, within a pre-defined error rate. The CP employs a fast base classifier which may be used to filter out unlikely labels from the target label set, and thus restrict the number of probable target class labels while ensuring the candidate class labels set meets the pre-defined error rate.
-
-
-
-
-
-
-