-
公开(公告)号:US20230102055A1
公开(公告)日:2023-03-30
申请号:US18058163
申请日:2022-11-22
Applicant: Adobe Inc.
Inventor: Taesung Park , Richard Zhang , Oliver Wang , Junyan Zhu , Jingwan Lu , Elya Shechtman , Alexei A. Efros
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for generating a modified digital image from extracted spatial and global codes. For example, the disclosed systems can utilize a global and spatial autoencoder to extract spatial codes and global codes from digital images. The disclosed systems can further utilize the global and spatial autoencoder to generate a modified digital image by combining extracted spatial and global codes in various ways for various applications such as style swapping, style blending, and attribute editing.
-
公开(公告)号:US20230274535A1
公开(公告)日:2023-08-31
申请号:US17680906
申请日:2022-02-25
Applicant: ADOBE INC.
Inventor: Yijun Li , Utkarsh Ojha , Richard Zhang , Jingwan Lu , Elya Shechtman , Alexei A. Efros
IPC: G06V10/774 , G06F3/04842
CPC classification number: G06V10/7747 , G06F3/04842
Abstract: An image generation system enables user input during the process of training a generative model to influence the model's ability to generate new images with desired visual features. A source generative model for a source domain is fine-tuned using training images in a target domain to provide an adapted generative model for the target domain. Interpretable factors are determined for the source generative model and the adapted generative model. A user interface is provided that enables a user to select one or more interpretable factors. The user-selected interpretable factor(s) are used to generate a user-adapted generative model, for instance, by using a loss function based on the user-selected interpretable factor(s). The user-adapted generative model can be used to create new images in the target domain.
-
3.
公开(公告)号:US20230260175A1
公开(公告)日:2023-08-17
申请号:US17650957
申请日:2022-02-14
Applicant: Adobe Inc.
Inventor: Nadav Epstein , Alexei A. Efros , Taesung Park , Richard Zhang , Elya Shechtman
CPC classification number: G06T11/60 , G06T7/90 , G06T2207/20084 , G06T2207/20212
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for generating digital images depicting photorealistic scenes utilizing a digital image collaging neural network. For example, the disclosed systems utilize a digital image collaging neural network having a particular architecture for disentangling generation of scene layouts and pixel colors for different regions of a digital image. In some cases, the disclosed systems break down the process of generating a collage digital into generating images representing different regions such as a background and a foreground to be collaged into a final result. For example, utilizing the digital image collaging neural network, the disclosed systems determine scene layouts and pixel colors for both foreground digital images and background digital images to ultimately collage the foreground and background together into a collage digital image depicting a real-world scene.
-
4.
公开(公告)号:US11625875B2
公开(公告)日:2023-04-11
申请号:US17091416
申请日:2020-11-06
Applicant: Adobe Inc.
Inventor: Taesung Park , Alexei A. Efros , Elya Shechtman , Richard Zhang , Junyan Zhu
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly generating modified digital images utilizing a novel swapping autoencoder that incorporates scene layout. In particular, the disclosed systems can receive a scene layout map that indicates or defines locations for displaying specific digital content within a digital image. In addition, the disclosed systems can utilize the scene layout map to guide combining portions of digital image latent code to generate a modified digital image with a particular textural appearance and a particular geometric structure defined by the scene layout map. Additionally, the disclosed systems can utilize a scene layout map that defines a portion of a digital image to modify by, for instance, adding new digital content to the digital image, and can generate a modified digital image depicting the new digital content.
-
5.
公开(公告)号:US20220148241A1
公开(公告)日:2022-05-12
申请号:US17091416
申请日:2020-11-06
Applicant: Adobe Inc.
Inventor: Taesung Park , Alexei A. Efros , Elya Shechtman , Richard Zhang , Junyan Zhu
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly generating modified digital images utilizing a novel swapping autoencoder that incorporates scene layout. In particular, the disclosed systems can receive a scene layout map that indicates or defines locations for displaying specific digital content within a digital image. In addition, the disclosed systems can utilize the scene layout map to guide combining portions of digital image latent code to generate a modified digital image with a particular textural appearance and a particular geometric structure defined by the scene layout map. Additionally, the disclosed systems can utilize a scene layout map that defines a portion of a digital image to modify by, for instance, adding new digital content to the digital image, and can generate a modified digital image depicting the new digital content.
-
公开(公告)号:US20250104399A1
公开(公告)日:2025-03-27
申请号:US18473603
申请日:2023-09-25
Applicant: ADOBE INC.
Inventor: Sheng-Yu Wang , Alexei A. Efros , Junyan Zhu , Richard Zhang
IPC: G06V10/77 , G06N3/0895 , G06V10/74 , G06V10/774 , G06V10/82
Abstract: Embodiments of the present disclosure perform training attribution by identifying a synthesized image and a training image, where the synthesized image was generated by an image generation model that was trained with the training image. A machine learning model computes first attribution features for the synthesized image using a first mapping layer and second attribution features for the training image using a second mapping layer that is different from the first mapping layer. Then, an attribution score is generated based on the first attribution features and the second attribution features, where the attribution score indicates a degree of influence for the training image on generating the synthesized image.
-
公开(公告)号:US20240169604A1
公开(公告)日:2024-05-23
申请号:US18057453
申请日:2022-11-21
Applicant: ADOBE INC.
Inventor: Yosef Gandelsman , Taesung Park , Richard Zhang , Elya Shechtman , Alexei A. Efros
IPC: G06T11/00 , G06F3/04842 , G06F3/04845 , G06T11/20
CPC classification number: G06T11/001 , G06F3/04842 , G06F3/04845 , G06T11/20
Abstract: Systems and methods for image generation are described. Embodiments of the present disclosure obtain user input that indicates a target color and a semantic label for a region of an image to be generated. The system also generates of obtains a noise map including noise biased towards the target color in the region indicated by the user input. A diffusion model generates the image based on the noise map and the semantic label for the region. The image can include an object in the designated region that is described by the semantic label and that has the target color.
-
公开(公告)号:US20220254071A1
公开(公告)日:2022-08-11
申请号:US17163284
申请日:2021-01-29
Applicant: Adobe Inc.
Inventor: Utkarsh Ojha , Yijun Li , Richard Zhang , Jingwan Lu , Elya Shechtman , Alexei A. Efros
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and efficiently modifying a generative adversarial neural network using few-shot adaptation to generate digital images corresponding to a target domain while maintaining diversity of a source domain and realism of the target domain. In particular, the disclosed systems utilize a generative adversarial neural network with parameters learned from a large source domain. The disclosed systems preserve relative similarities and differences between digital images in the source domain using a cross-domain distance consistency loss. In addition, the disclosed systems utilize an anchor-based strategy to encourage different levels or measures of realism over digital images generated from latent vectors in different regions of a latent space.
-
公开(公告)号:US12230014B2
公开(公告)日:2025-02-18
申请号:US17680906
申请日:2022-02-25
Applicant: ADOBE INC.
Inventor: Yijun Li , Utkarsh Ojha , Richard Zhang , Jingwan Lu , Elya Shechtman , Alexei A. Efros
IPC: G06V10/774 , G06F3/04842
Abstract: An image generation system enables user input during the process of training a generative model to influence the model's ability to generate new images with desired visual features. A source generative model for a source domain is fine-tuned using training images in a target domain to provide an adapted generative model for the target domain. Interpretable factors are determined for the source generative model and the adapted generative model. A user interface is provided that enables a user to select one or more interpretable factors. The user-selected interpretable factor(s) are used to generate a user-adapted generative model, for instance, by using a loss function based on the user-selected interpretable factor(s). The user-adapted generative model can be used to create new images in the target domain.
-
公开(公告)号:US12136151B2
公开(公告)日:2024-11-05
申请号:US17650957
申请日:2022-02-14
Applicant: Adobe Inc.
Inventor: Nadav Epstein , Alexei A. Efros , Taesung Park , Richard Zhang , Elya Shechtman
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for generating digital images depicting photorealistic scenes utilizing a digital image collaging neural network. For example, the disclosed systems utilize a digital image collaging neural network having a particular architecture for disentangling generation of scene layouts and pixel colors for different regions of a digital image. In some cases, the disclosed systems break down the process of generating a collage digital into generating images representing different regions such as a background and a foreground to be collaged into a final result. For example, utilizing the digital image collaging neural network, the disclosed systems determine scene layouts and pixel colors for both foreground digital images and background digital images to ultimately collage the foreground and background together into a collage digital image depicting a real-world scene.
-
-
-
-
-
-
-
-
-