MULTI-ATTRIBUTE INVERSION FOR TEXT-TO-IMAGE SYNTHESIS

    公开(公告)号:US20250166243A1

    公开(公告)日:2025-05-22

    申请号:US18513071

    申请日:2023-11-17

    Applicant: ADOBE INC.

    Abstract: An image generation model obtains a text prompt, a first attribute token, and a second attribute token. A first set of layers of the image generation model and a first set of time-steps are identified for the first attribute token and a second set of layers of the image generation model and a second set of time-steps are identified for the second attribute token. A synthetic image is generated based on the text prompt, the first attribute token, and the second attribute token by providing the first attribute token to the first set layers of the image generation model during the first set of time-steps and providing the second attribute token to the second set of layers of the image generation model during the second set of time-steps.

    Modality adaptive information retrieval

    公开(公告)号:US12198048B2

    公开(公告)日:2025-01-14

    申请号:US17153130

    申请日:2021-01-20

    Applicant: Adobe Inc.

    Abstract: In some embodiments, a multimodal computing system receives a query and identifies, from source documents, text passages and images that are relevant to the query. The multimodal computing system accesses a multimodal question-answering model that includes a textual stream of language models and a visual stream of language models. Each of the textual stream and the visual stream contains a set of transformer-based models and each transformer-based model includes a cross-attention layer using data generated by both the textual stream and visual stream of language models as an input. The multimodal computing system identifies text relevant to the query by applying the textual stream to the text passages and computes, using the visual stream, relevance scores of the images to the query, respectively. The multimodal computing system further generates a response to the query by including the text and/or an image according to the relevance scores.

    TEXT-TO-IMAGE SYNTHESIS UTILIZING DIFFUSION MODELS WITH TEST-TIME ATTENTION SEGREGATION AND RETENTION OPTIMIZATION

    公开(公告)号:US20240428468A1

    公开(公告)日:2024-12-26

    申请号:US18337634

    申请日:2023-06-20

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that utilizes attention segregation loss and/or attention retention loss at inference time of a diffusion neural network to generate a text-conditioned image. In particular, in some embodiments, the disclosed systems utilize the attention segregation loss to reduce overlap between concepts by comparing attention maps for multiple concepts of a text query corresponding to a denoising step. Further, in some embodiments, the disclosed systems utilize the attention retention loss to improve information retention for concepts across denoising steps by comparing attention maps between different denoising steps. Accordingly, in some embodiments, by utilizing the attention segregation loss and the attention retention loss, the disclosed systems accurately maintain multiple concepts from a text query when generating a text-conditioned image.

    Generating summary content tuned to a target characteristic using a word generation model

    公开(公告)号:US11657225B2

    公开(公告)日:2023-05-23

    申请号:US17348257

    申请日:2021-06-15

    Applicant: Adobe Inc.

    CPC classification number: G06F40/284 G06N20/00

    Abstract: Systems and methods for generating a tuned summary using a word generation model. An example method includes receiving, at a decoder of the word generation model, a training data learned subspace representation of training data. The method also includes identifying tunable linguistic characteristics of the word generation model and training the decoder to output a training tuned summary of the training data learned subspace representation based on at least one of the tunable linguistic characteristics. The method further includes receiving an input text and a target characteristic token, and generating, by the trained decoder of the word generation model, each word of a tuned summary of the input text from a learned subspace representation and from feedback about preceding words of the tuned summary, wherein the tuned summary is tuned to target characteristics represented by the target characteristic token.

    Word attribution prediction from subject data

    公开(公告)号:US11580307B2

    公开(公告)日:2023-02-14

    申请号:US16825864

    申请日:2020-03-20

    Applicant: Adobe Inc.

    Abstract: A digital attribution system is described to generate predictions of word attributions from subject data, e.g., titles, subject lines of emails, and so on. To do so, an attribution score is first generated by the digital attribution system that describe an amount to which respective words in the subject data cause performance of a corresponding outcome. The attribution scores are then used by the digital attribution system to generate representations for display in a user interface for respective words in the subject data and may also be used to generate attribution recommendations of changes to be made to the subject data.

    Multi-dimensional language style transfer

    公开(公告)号:US11487971B2

    公开(公告)日:2022-11-01

    申请号:US17073258

    申请日:2020-10-16

    Applicant: Adobe Inc.

    Abstract: In some embodiments, a style transfer computing system generates a set of discriminator models corresponding to a set of styles based on a set of training datasets labeled for respective styles. The style transfer computing system further generates a style transfer language model for a target style combination that includes multiple target styles from the set of styles. The style transfer language model includes a cascaded language model and multiple discriminator models selected from the set of discriminator models. The style transfer computing system trains the style transfer language model to minimize a loss function containing a loss term for the cascaded language model and multiple loss terms for the multiple discriminator models. For a source sentence and a given target style combination, the style transfer computing system applies the style transfer language model on the source sentence to generate a target sentence in the given target style combination.

    Stylistic Text Rewriting for a Target Author

    公开(公告)号:US20210406465A1

    公开(公告)日:2021-12-30

    申请号:US17467672

    申请日:2021-09-07

    Applicant: Adobe Inc.

    Abstract: Rewriting text in the writing style of a target author is described. A stylistic rewriting system receives input text and an indication of the target author. The system trains a language model to understand the target author's writing style using a corpus of text associated with the target author. The language model may be transformer-based, and is first trained on a different corpus of text associated with a range of different authors to understand linguistic nuances of a particular language. Copies of the language model are then cascaded into an encoder-decoder framework, which is further trained using a masked language modeling objective and a noisy version of the target author corpus. After training, the encoder-decoder framework of the trained language model automatically rewrites input text in the writing style of the target author and outputs the rewritten text as stylized text.

    Generating a topic-based summary of textual content

    公开(公告)号:US10685050B2

    公开(公告)日:2020-06-16

    申请号:US15960505

    申请日:2018-04-23

    Applicant: Adobe Inc.

    Abstract: A word generation model obtains textual content and a requested topic of interest, and generates a targeted summary of the textual content tuned to the topic of interest. To do so, a topic-aware encoding model encodes the textual content with a topic label corresponding to the topic of interest to generate topic-aware encoded text. A word generation model selects a next word for the topic-based summary from the topic-aware encoded text. The word generation model is trained to generate topic-based summaries using machine learning on training data including a multitude of documents, a respective summary of each document, and a respective topic of each summary. Feedback of the selected next word is provided to the word generation model. The feedback causes the word generation model to select subsequent words for the topic-based summary based on the feedback of the next selected word.

Patent Agency Ranking