Automatic digital parameter adjustment including tone and color correction

    公开(公告)号:US11178368B2

    公开(公告)日:2021-11-16

    申请号:US16696160

    申请日:2019-11-26

    申请人: Adobe Inc.

    IPC分类号: H04N9/31 G06T5/00 H04N9/73

    摘要: Systems and techniques for automatic digital parameter adjustment are described that leverage insights learned from an image set to automatically predict parameter values for an input item of digital visual content. To do so, the automatic digital parameter adjustment techniques described herein captures visual and contextual features of digital visual content to determine balanced visual output in a range of visual scenes and settings. The visual and contextual features of digital visual content are used to train a parameter adjustment model through machine learning techniques that captures feature patterns and interactions. The parameter adjustment model exploits these feature interactions to determine visually pleasing parameter values for an input item of digital visual content. The predicted parameter values are output, allowing further adjustment to the parameter values.

    EDGE-GUIDED RANKING LOSS FOR MONOCULAR DEPTH PREDICTION

    公开(公告)号:US20210256717A1

    公开(公告)日:2021-08-19

    申请号:US16790056

    申请日:2020-02-13

    申请人: Adobe Inc.

    IPC分类号: G06T7/50 G06T7/13 G06N3/08

    摘要: In order to provide monocular depth prediction, a trained neural network may be used. To train the neural network, edge detection on a digital image may be performed to determine at least one edge of the digital image, and then a first point and a second point of the digital image may be sampled, based on the at least one edge. A relative depth between the first point and the second point may be predicted, and the neural network may be trained to perform monocular depth prediction using a loss function that compares the predicted relative depth with a ground truth relative depth between the first point and the second point.

    Transcript-based insertion of secondary video content into primary video content

    公开(公告)号:US11049525B2

    公开(公告)日:2021-06-29

    申请号:US16281903

    申请日:2019-02-21

    申请人: Adobe Inc.

    IPC分类号: G11B27/11

    摘要: Certain embodiments involve transcript-based techniques for facilitating insertion of secondary video content into primary video content. For instance, a video editor presents a video editing interface having a primary video section displaying a primary video, a text-based navigation section having navigable portions of a primary video transcript, and a secondary video menu section displaying candidate secondary videos. In some embodiments, candidate secondary videos are obtained by using target terms detected in the transcript to query a remote data source for the candidate secondary videos. In embodiments involving video insertion, the video editor identifies a portion of the primary video corresponding to a portion of the transcript selected within the text-based navigation section. The video editor inserts a secondary video, which is selected from the candidate secondary videos based on an input received at the secondary video menu section, at the identified portion of the primary video.

    ENHANCED VIDEO SHOT MATCHING USING GENERATIVE ADVERSARIAL NETWORKS

    公开(公告)号:US20210158570A1

    公开(公告)日:2021-05-27

    申请号:US16692503

    申请日:2019-11-22

    申请人: Adobe Inc.

    摘要: This disclosure involves training generative adversarial networks to shot-match two unmatched images in a context-sensitive manner. For example, aspects of the present disclosure include accessing a trained generative adversarial network including a trained generator model and a trained discriminator model. A source image and a reference image may be inputted into the generator model to generate a modified source image. The modified source image and the reference image may be inputted into the discriminator model to determine a likelihood that the modified source image is color-matched with the reference image. The modified source image may be outputted as a shot-match with the reference image in response to determining, using the discriminator model, that the modified source image and the reference image are color-matched.

    Skin Tone Assisted Digital Image Color Matching

    公开(公告)号:US20210142042A1

    公开(公告)日:2021-05-13

    申请号:US17154830

    申请日:2021-01-21

    申请人: Adobe Inc.

    IPC分类号: G06K9/00 G06T7/90

    摘要: In implementations of skin tone assisted digital image color matching, a device implements a color editing system, which includes a facial detection module to detect faces in an input image and in a reference image, and includes a skin tone model to determine a skin tone value reflective of a skin tone of each of the faces. A color matching module can be implemented to group the faces into one or more face groups based on the skin tone value of each of the faces, match a face group pair as an input image face group paired with a reference image face group, and generate a modified image from the input image based on color features of the reference image, the color features including face skin tones of the respective faces in the face group pair as part of the color features applied to modify the input image.

    INCORPORATING BLACK-BOX FUNCTIONS IN NEURAL NETWORKS

    公开(公告)号:US20210012189A1

    公开(公告)日:2021-01-14

    申请号:US16507675

    申请日:2019-07-10

    申请人: Adobe Inc.

    IPC分类号: G06N3/08 G06F17/13 G06N3/10

    摘要: Techniques for incorporating a black-box function into a neural network are described. For example, an image editing function may be the black-box function and may be wrapped into a layer of the neural network. A set of parameters and a source image are provided to the black-box function, and the output image that represents the source image with the set of parameters applied to the source image is output from the black-box function. To address the issue that the black-box function may not be differentiable, a loss optimization may calculate the gradients of the function using, for example, a finite differences calculation, and the gradients are used to train the neural network to ensure the output image is representative of an expected ground truth image.

    3D OBJECT RECONSTRUCTION USING PHOTOMETRIC MESH REPRESENTATION

    公开(公告)号:US20200372710A1

    公开(公告)日:2020-11-26

    申请号:US16985402

    申请日:2020-08-05

    申请人: Adobe, Inc.

    摘要: Techniques are disclosed for 3D object reconstruction using photometric mesh representations. A decoder is pretrained to transform points sampled from 2D patches of representative objects into 3D polygonal meshes. An image frame of the object is fed into an encoder to get an initial latent code vector. For each frame and camera pair from the sequence, a polygonal mesh is rendered at the given viewpoints. The mesh is optimized by creating a virtual viewpoint, rasterized to obtain a depth map. The 3D mesh projections are aligned by projecting the coordinates corresponding to the polygonal face vertices of the rasterized mesh to both selected viewpoints. The photometric error is determined from RGB pixel intensities sampled from both frames. Gradients from the photometric error are backpropagated into the vertices of the assigned polygonal indices by relating the barycentric coordinates of each image to update the latent code vector.

    Deep patch feature prediction for image inpainting

    公开(公告)号:US10740881B2

    公开(公告)日:2020-08-11

    申请号:US15935994

    申请日:2018-03-26

    申请人: Adobe Inc.

    IPC分类号: G06T5/00 G06K9/62

    摘要: Techniques for using deep learning to facilitate patch-based image inpainting are described. In an example, a computer system hosts a neural network trained to generate, from an image, code vectors including features learned by the neural network and descriptive of patches. The image is received and contains a region of interest (e.g., a hole missing content). The computer system inputs it to the network and, in response, receives the code vectors. Each code vector is associated with a pixel in the image. Rather than comparing RGB values between patches, the computer system compares the code vector of a pixel inside the region to code vectors of pixels outside the region to find the best match based on a feature similarity measure (e.g., a cosine similarity). The pixel value of the pixel inside the region is set based on the pixel value of the matched pixel outside this region.

    Dynamically generating and changing view-specific-filter parameters for 360-degree videos

    公开(公告)号:US11539932B2

    公开(公告)日:2022-12-27

    申请号:US17519332

    申请日:2021-11-04

    申请人: Adobe Inc.

    摘要: This disclosure relates to methods, non-transitory computer readable media, and systems that generate and dynamically change filter parameters for a frame of a 360-degree video based on detecting a field of view from a computing device. As a computing device rotates or otherwise changes orientation, for instance, the disclosed systems can detect a field of view and interpolate one or more filter parameters corresponding to nearby spatial keyframes of the 360-degree video to generate view-specific-filter parameters. By generating and storing filter parameters for spatial keyframes corresponding to different times and different view directions, the disclosed systems can dynamically adjust color grading or other visual effects using interpolated, view-specific-filter parameters to render a filtered version of the 360-degree video.