Incorporating black-box functions in neural networks

    公开(公告)号:US11481619B2

    公开(公告)日:2022-10-25

    申请号:US16507675

    申请日:2019-07-10

    Applicant: Adobe Inc.

    Abstract: Techniques for incorporating a black-box function into a neural network are described. For example, an image editing function may be the black-box function and may be wrapped into a layer of the neural network. A set of parameters and a source image are provided to the black-box function, and the output image that represents the source image with the set of parameters applied to the source image is output from the black-box function. To address the issue that the black-box function may not be differentiable, a loss optimization may calculate the gradients of the function using, for example, a finite differences calculation, and the gradients are used to train the neural network to ensure the output image is representative of an expected ground truth image.

    Image composites using a generative neural network

    公开(公告)号:US11328523B2

    公开(公告)日:2022-05-10

    申请号:US16897068

    申请日:2020-06-09

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to an image composite system that employs a generative adversarial network to generate realistic composite images. For example, in one or more embodiments, the image composite system trains a geometric prediction neural network using an adversarial discrimination neural network to learn warp parameters that provide correct geometric alignment of foreground objects with respect to a background image. Once trained, the determined warp parameters provide realistic geometric corrections to foreground objects such that the warped foreground objects appear to blend into background images naturally when composited together.

    LARGE-SCALE OUTDOOR AUGMENTED REALITY SCENES USING CAMERA POSE BASED ON LEARNED DESCRIPTORS

    公开(公告)号:US20220114365A1

    公开(公告)日:2022-04-14

    申请号:US17068429

    申请日:2020-10-12

    Applicant: ADOBE INC.

    Abstract: Methods and systems are provided for facilitating large-scale augmented reality in relation to outdoor scenes using estimated camera pose information. In particular, camera pose information for an image can be estimated by matching the image to a rendered ground-truth terrain model with known camera pose information. To match images with such renders, data driven cross-domain feature embedding can be learned using a neural network. Cross-domain feature descriptors can be used for efficient and accurate feature matching between the image and the terrain model renders. This feature matching allows images to be localized in relation to the terrain model, which has known camera pose information. This known camera pose information can then be used to estimate camera pose information in relation to the image.

    DYNAMICALLY GENERATING AND CHANGING VIEW-SPECIFIC-FILTER PARAMETERS FOR 360-DEGREE VIDEOS

    公开(公告)号:US20220060671A1

    公开(公告)日:2022-02-24

    申请号:US17519332

    申请日:2021-11-04

    Applicant: Adobe Inc.

    Abstract: This disclosure relates to methods, non-transitory computer readable media, and systems that generate and dynamically change filter parameters for a frame of a 360-degree video based on detecting a field of view from a computing device. As a computing device rotates or otherwise changes orientation, for instance, the disclosed systems can detect a field of view and interpolate one or more filter parameters corresponding to nearby spatial keyframes of the 360-degree video to generate view-specific-filter parameters. By generating and storing filter parameters for spatial keyframes corresponding to different times and different view directions, the disclosed systems can dynamically adjust color grading or other visual effects using interpolated, view-specific-filter parameters to render a filtered version of the 360-degree video.

    TEMPORALLY DISTRIBUTED NEURAL NETWORKS FOR VIDEO SEMANTIC SEGMENTATION

    公开(公告)号:US20210319232A1

    公开(公告)日:2021-10-14

    申请号:US16846544

    申请日:2020-04-13

    Applicant: Adobe Inc

    Abstract: A Video Semantic Segmentation System (VSSS) is disclosed that performs accurate and fast semantic segmentation of videos using a set of temporally distributed neural networks. The VSSS receives as input a video signal comprising a contiguous sequence of temporally-related video frames. The VSSS extracts features from the video frames in the contiguous sequence and based upon the extracted features, selects, from a set of labels, a label to be associated with each pixel of each video frame in the video signal. In certain embodiments, a set of multiple neural networks are used to extract the features to be used for video segmentation and the extraction of features is distributed among the multiple neural networks in the set. A strong feature representation representing the entirety of the features is produced for each video frame in the sequence of video frames by aggregating the output features extracted by the multiple neural networks.

    RECONSTRUCTING THREE-DIMENSIONAL SCENES IN A TARGET COORDINATE SYSTEM FROM MULTIPLE VIEWS

    公开(公告)号:US20210295606A1

    公开(公告)日:2021-09-23

    申请号:US16822819

    申请日:2020-03-18

    Applicant: Adobe Inc.

    Abstract: Methods, systems, and non-transitory computer readable storage media are disclosed for reconstructing three-dimensional meshes from two-dimensional images of objects with automatic coordinate system alignment. For example, the disclosed system can generate feature vectors for a plurality of images having different views of an object. The disclosed system can process the feature vectors to generate coordinate-aligned feature vectors aligned with a coordinate system associated with an image. The disclosed system can generate a combined feature vector from the feature vectors aligned to the coordinate system. Additionally, the disclosed system can then generate a three-dimensional mesh representing the object from the combined feature vector.

    Video inpainting via confidence-weighted motion estimation

    公开(公告)号:US11081139B2

    公开(公告)日:2021-08-03

    申请号:US16378906

    申请日:2019-04-09

    Applicant: Adobe Inc.

    Abstract: Certain aspects involve video inpainting via confidence-weighted motion estimation. For instance, a video editor accesses video content having a target region to be modified in one or more video frames. The video editor computes a motion for a boundary of the target region. The video editor interpolates, from the boundary motion, a target motion of a target pixel within the target region. In the interpolation, confidence values assigned to boundary pixels control how the motion of these pixels contributes to the interpolated target motion. A confidence value is computed based on a difference between forward and reverse motion with respect to a particular boundary pixel, a texture in a region that includes the particular boundary pixel, or a combination thereof. The video editor modifies the target region in the video by updating color data of the target pixel to correspond to the target motion interpolated from the boundary motion.

    INTELLIGENT VIDEO REFRAMING
    79.
    发明申请

    公开(公告)号:US20210218929A1

    公开(公告)日:2021-07-15

    申请号:US17217951

    申请日:2021-03-30

    Applicant: Adobe Inc.

    Abstract: Embodiments of the present invention are directed towards reframing videos from one aspect ratio to another aspect ratio while maintaining visibility of regions of interest. A set of regions of interest are determined in frames in a video with a first aspect ratio. The set of regions of interest can be used to estimate an initial camera path. An optimal camera path is determined by leveraging the identified regions of interest using the initial camera path. Sub crops with a second aspect ratio different from the first aspect ratio of the video are identified. The sub crops are placed as designated using the optimal camera path to generate a cropped video with the second aspect ratio.

    Intelligent video reframing
    80.
    发明授权

    公开(公告)号:US10986308B2

    公开(公告)日:2021-04-20

    申请号:US16359876

    申请日:2019-03-20

    Applicant: ADOBE INC.

    Abstract: Embodiments of the present invention are directed towards reframing videos from one aspect ratio to another aspect ratio while maintaining visibility of regions of interest. A set of regions of interest are determined in frames in a video with a first aspect ratio. The set of regions of interest can be used to estimate an initial camera path. An optimal camera path is determined by leveraging the identified regions of interest using the initial camera path. Sub crops with a second aspect ratio different from the first aspect ratio of the video are identified. The sub crops are placed as designated using the optimal camera path to generate a cropped video with the second aspect ratio.

Patent Agency Ranking