Automated digital parameter adjustment for digital images

    公开(公告)号:US11930303B2

    公开(公告)日:2024-03-12

    申请号:US17526998

    申请日:2021-11-15

    Applicant: Adobe Inc.

    CPC classification number: H04N9/3182 G06T5/92 H04N9/73 G06T2207/20081

    Abstract: Systems and techniques for automatic digital parameter adjustment are described that leverage insights learned from an image set to automatically predict parameter values for an input item of digital visual content. To do so, the automatic digital parameter adjustment techniques described herein captures visual and contextual features of digital visual content to determine balanced visual output in a range of visual scenes and settings. The visual and contextual features of digital visual content are used to train a parameter adjustment model through machine learning techniques that captures feature patterns and interactions. The parameter adjustment model exploits these feature interactions to determine visually pleasing parameter values for an input item of digital visual content. The predicted parameter values are output, allowing further adjustment to the parameter values.

    ADAPTING GENERATIVE NEURAL NETWORKS USING A CROSS DOMAIN TRANSLATION NETWORK

    公开(公告)号:US20240037922A1

    公开(公告)日:2024-02-01

    申请号:US17815451

    申请日:2022-07-27

    Applicant: Adobe Inc.

    CPC classification number: G06V10/82 G06V10/7715 G06V10/469

    Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for adapting generative neural networks to target domains utilizing an image translation neural network. In particular, in one or more embodiments, the disclosed systems utilize an image translation neural network to translate target results to a source domain for input in target neural network adaptation. For instance, in some embodiments, the disclosed systems compare a translated target result with a source result from a pretrained source generative neural network to adjust parameters of a target generative neural network to produce results corresponding in features to source results and corresponding in style to the target domain.

    DIGITAL IMAGE INPAINTING UTILIZING A CASCADED MODULATION INPAINTING NEURAL NETWORK

    公开(公告)号:US20230360180A1

    公开(公告)日:2023-11-09

    申请号:US17661985

    申请日:2022-05-04

    Applicant: Adobe Inc.

    CPC classification number: G06T5/005 G06T3/4046 G06V10/40 G06T2207/20084

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media that generate inpainted digital images utilizing a cascaded modulation inpainting neural network. For example, the disclosed systems utilize a cascaded modulation inpainting neural network that includes cascaded modulation decoder layers. For example, in one or more decoder layers, the disclosed systems start with global code modulation that captures the global-range image structures followed by an additional modulation that refines the global predictions. Accordingly, in one or more implementations, the image inpainting system provides a mechanism to correct distorted local details. Furthermore, in one or more implementations, the image inpainting system leverages fast Fourier convolutions block within different resolution layers of the encoder architecture to expand the receptive field of the encoder and to allow the network encoder to better capture global structure.

    Diverse Image Inpainting Using Contrastive Learning

    公开(公告)号:US20230342884A1

    公开(公告)日:2023-10-26

    申请号:US17725818

    申请日:2022-04-21

    Applicant: Adobe Inc.

    Abstract: An image inpainting system is described that receives an input image that includes a masked region. From the input image, the image inpainting system generates a synthesized image that depicts an object in the masked region by selecting a first code that represents a known factor characterizing a visual appearance of the object and a second code that represents an unknown factor characterizing the visual appearance of the object apart from the known factor in latent space. The input image, the first code, and the second code are provided as input to a generative adversarial network that is trained to generate the synthesized image using contrastive losses. Different synthesized images are generated from the same input image using different combinations of first and second codes, and the synthesized images are output for display.

    Style-aware audio-driven talking head animation from a single image

    公开(公告)号:US11776188B2

    公开(公告)日:2023-10-03

    申请号:US17887685

    申请日:2022-08-15

    Applicant: Adobe Inc.

    CPC classification number: G06T13/205 G06T13/40 G06T17/20

    Abstract: Embodiments of the present invention provide systems, methods, and computer storage media for generating an animation of a talking head from an input audio signal of speech and a representation (such as a static image) of a head to animate. Generally, a neural network can learn to predict a set of 3D facial landmarks that can be used to drive the animation. In some embodiments, the neural network can learn to detect different speaking styles in the input speech and account for the different speaking styles when predicting the 3D facial landmarks. Generally, template 3D facial landmarks can be identified or extracted from the input image or other representation of the head, and the template 3D facial landmarks can be used with successive windows of audio from the input speech to predict 3D facial landmarks and generate a corresponding animation with plausible 3D effects.

Patent Agency Ranking