-
公开(公告)号:US20250077775A1
公开(公告)日:2025-03-06
申请号:US18457794
申请日:2023-08-29
Applicant: Adobe Inc.
Inventor: Zhongfen Deng , Seunghyun Yoon , Trung Bui , Quan Tran , Franck Dernoncourt
IPC: G06F40/284 , G06F40/166 , G06N20/00
Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for generating aspect-based summaries utilizing deep learning. In particular, in one or more embodiments, the disclosed systems access a transcript comprising sentences. The disclosed systems generate, utilizing a sentence classification machine learning model, aspect labels for the sentences of the transcript. The disclosed systems organize the sentences based on the aspect labels. The disclosed systems generate, utilizing a summary machine learning model, a summary of the transcript for each aspect of the plurality of aspects from the organized sentences.
-
公开(公告)号:US20230259718A1
公开(公告)日:2023-08-17
申请号:US17651555
申请日:2022-02-17
Applicant: Adobe Inc.
Inventor: Cesa Salaam , Seunghyun Yoon , Trung Huu Bui , Franck Dernoncourt
CPC classification number: G06F40/58 , G06F40/47 , G06N3/0454 , G06N3/08
Abstract: Techniques for training a language model for code switching content are disclosed. Such techniques include, in some embodiments, generating a dataset, which includes identifying one or more portions within textual content in a first language, the identified one or more portions each including one or more of offensive content or non-offensive content; translating the identified one or more salient portions to a second language; and reintegrating the translated one or more portions into the textual content to generate code-switched textual content. In some cases, the textual content in the first language includes offensive content and non-offensive content, the identified one or more portions include the offensive content, and the translated one or more portions include a translated version of the offensive content. In some embodiments, the code-switched textual content is at least part of a synthetic dataset usable to train a language model, such as a multilingual classification model.
-
3.
公开(公告)号:US20250022459A1
公开(公告)日:2025-01-16
申请号:US18220910
申请日:2023-07-12
Applicant: Adobe Inc.
Inventor: Viet Dac Lai , Trung Bui , Seunghyun Yoon , Quan Tran , Hao Tan , Hanieh Deilamsalehy , Abel Salinas , Franck Dernoncourt
IPC: G10L15/183 , G10L15/065
Abstract: The disclosed method generates helpful training data for a language model, for example, a model implementing a punctuation restoration task, for real-world ASR texts. The method uses a reinforcement learning method using a generative AI model to generate additional data to train the language model. The method allows the generative AI model to learn from real-world ASR text to generate more effective training examples based on gradient feedback from the language model.
-
公开(公告)号:US20230267726A1
公开(公告)日:2023-08-24
申请号:US17651771
申请日:2022-02-18
Applicant: Adobe Inc.
Inventor: Seunghyun Yoon , Trung Huu Bui , Franck Dernoncourt , Hyounghun Kim , Doo Soon Kim
CPC classification number: G06V10/86 , G06V10/82 , G06V10/806 , G06V10/7715 , G06N3/088 , G06N3/0445 , G06F40/284
Abstract: Embodiments of the disclosure provide a machine learning model for generating a predicted executable command for an image. The learning model includes an interface configured to obtain an utterance indicating a request associated with the image, an utterance sub-model, a visual sub-model, an attention network, and a selection gate. The machine learning model generates a segment of the predicted executable command from weighted probabilities of each candidate token in a predetermined vocabulary determined based on the visual features, the concept features, current command features, and the utterance features extracted from the utterance or the image.
-
公开(公告)号:US20230153522A1
公开(公告)日:2023-05-18
申请号:US17455533
申请日:2021-11-18
Applicant: ADOBE INC.
Inventor: Jaemin Cho , Seunghyun Yoon , Ajinkya Gorakhnath Kale , Trung Huu Bui , Franck Dernoncourt
IPC: G06F40/253 , G06K9/62 , G06F16/583
CPC classification number: G06F40/253 , G06K9/6256 , G06K9/6262 , G06F16/583
Abstract: Systems and methods for image captioning are described. One or more aspects of the systems and methods include generating a training caption for a training image using an image captioning network; encoding the training caption using a multi-modal encoder to obtain an encoded training caption; encoding the training image using the multi-modal encoder to obtain an encoded training image; computing a reward function based on the encoded training caption and the encoded training image; and updating parameters of the image captioning network based on the reward function.
-
6.
公开(公告)号:US20230153341A1
公开(公告)日:2023-05-18
申请号:US17528901
申请日:2021-11-17
Applicant: Adobe Inc.
Inventor: Seunghyun Yoon
CPC classification number: G06F16/35 , G06K9/00469 , G06N3/04 , G06K9/6256
Abstract: An incongruent headline detection system receives a request to determine a headline incongruence score for an electronic document. The incongruent headline detection system determines the headline incongruence score for the electronic document by applying a machine learning model to the electronic document. Applying the machine learning model to the electronic document includes generating a graph representing a textual similarity between a headline of the electronic document and each of a plurality of paragraphs of the electronic document and determining the headline incongruence score using the graph. The incongruent headline detection system transmits, responsive to the request, the headline incongruence score for the electronic document.
-
公开(公告)号:US20240020337A1
公开(公告)日:2024-01-18
申请号:US17811963
申请日:2022-07-12
Applicant: ADOBE INC.
Inventor: Adyasha Maharana , Quan Hung Tran , Seunghyun Yoon , Franck Dernoncourt , Trung Huu Bui , Walter W. Chang
IPC: G06F16/738 , G10L13/08 , G06F40/284 , G06F16/783
CPC classification number: G06F16/739 , G10L13/08 , G06F40/284 , G06F16/7844
Abstract: Systems and methods for intent discovery and video summarization are described. Embodiments of the present disclosure receive a video and a transcript of the video, encode the video to obtain a sequence of video encodings, encode the transcript to obtain a sequence of text encodings, apply a visual gate to the sequence of text encodings based on the sequence of video encodings to obtain gated text encodings, and generate an intent label for the transcript based on the gated text encodings.
-
公开(公告)号:US20230419164A1
公开(公告)日:2023-12-28
申请号:US17846428
申请日:2022-06-22
Applicant: Adobe Inc.
Inventor: Khalil Mrini , Franck Dernoncourt , Seunghyun Yoon , Trung Huu Bui , Walter W. Chang , Emilia Farcas , Ndapandula T. Nakashole
IPC: G06N20/00
CPC classification number: G06N20/00
Abstract: Multitask machine-learning model training and training data augmentation techniques are described. In one example, training is performed for multiple tasks simultaneously as part of training a multitask machine-learning model using question pairs. Examples of the multiple tasks include question summarization and recognizing question entailment. Further, a loss function is described that incorporates a parameter sharing loss that is configured to adjust an amount that parameters are shared between corresponding layers trained for the first and second tasks, respectively. In an implementation, training data augmentation techniques are also employed by synthesizing question pairs, automatically and without user intervention, to improve accuracy in model training.
-
公开(公告)号:US20230418868A1
公开(公告)日:2023-12-28
申请号:US17808599
申请日:2022-06-24
Applicant: ADOBE INC.
Inventor: Yeon Seonwoo , Seunghyun Yoon , Trung Huu Bui , Franck Dernoncourt , Roger K. Brooks , Mihir Naware
IPC: G06F16/901 , G06F16/903 , G06F16/9038 , G06F16/93
CPC classification number: G06F16/9024 , G06F16/90335 , G06F16/9038 , G06F16/93
Abstract: Systems and methods for text processing are described. Embodiments of the present disclosure receive a query comprising a natural language expression; extract a plurality of mentions from the query; generate a relation vector between a pair of the plurality of mentions using a relation encoder network, wherein the relation encoder network is trained using a contrastive learning process where mention pairs from a same document are labeled as positive samples and mention pairs from different documents are labeled as negative samples; combine the plurality of mentions with the relation vector to obtain a virtual knowledge graph of the query; identify a document corresponding to the query by comparing the virtual knowledge graph of the query to a virtual knowledge graph of the document; and transmit a response to the query, wherein the response includes a reference to the document.
-
10.
公开(公告)号:US20220076693A1
公开(公告)日:2022-03-10
申请号:US17526810
申请日:2021-11-15
Applicant: Adobe Inc.
Inventor: Trung Bui , Subhadeep Dey , Seunghyun Yoon
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for determining speech emotion. In particular, a speech emotion recognition system generates an audio feature vector and a textual feature vector for a sequence of words. Further, the speech emotion recognition system utilizes a neural attention mechanism that intelligently blends together the audio feature vector and the textual feature vector to generate attention output. Using the attention output, which includes consideration of both audio and text modalities for speech corresponding to the sequence of words, the speech emotion recognition system can apply attention methods to one of the feature vectors to generate a hidden feature vector. Based on the hidden feature vector, the speech emotion recognition system can generate a speech emotion probability distribution of emotions among a group of candidate emotions, and then select one of the candidate emotions as corresponding to the sequence of words.
-
-
-
-
-
-
-
-
-