Convolutional color correction in digital images

    公开(公告)号:US10237527B2

    公开(公告)日:2019-03-19

    申请号:US16110912

    申请日:2018-08-23

    Applicant: Google LLC

    Abstract: A computing device may obtain an input image. The input image may have a white point represented by chrominance values that define white color in the input image. Possibly based on colors of the input image, the computing device may generate a two-dimensional chrominance histogram of the input image. The computing device may convolve the two-dimensional chrominance histogram with a filter to create a two-dimensional heat map. Entries in the two-dimensional heat map may represent respective estimates of how close respective tints corresponding to the respective entries are to the white point of the input image. The computing device may select an entry in the two-dimensional heat map that represents a particular value that is within a threshold of a maximum value in the heat map, and based on the selected entry, tint the input image to form an output image.

    Hardware-based convolutional color correction in digital images

    公开(公告)号:US10091479B2

    公开(公告)日:2018-10-02

    申请号:US15703571

    申请日:2017-09-13

    Applicant: Google LLC

    Abstract: A computing device may obtain an input image. The input image may have a white point represented by chrominance values that define white color in the input image. Possibly based on colors of the input image, the computing device may generate a two-dimensional chrominance histogram of the input image. The computing device may convolve the two-dimensional chrominance histogram with a filter to create a two-dimensional heat map. Entries in the two-dimensional heat map may represent respective estimates of how close respective tints corresponding to the respective entries are to the white point of the input image. The computing device may select an entry in the two-dimensional heat map that represents a particular value that is within a threshold of a maximum value in the heat map, and based on the selected entry, tint the input image to form an output image.

    Learning-Based Lens Flare Removal
    5.
    发明公开

    公开(公告)号:US20240320808A1

    公开(公告)日:2024-09-26

    申请号:US18734000

    申请日:2024-06-05

    Applicant: Google LLC

    Abstract: A method includes obtaining an input image that contains a particular representation of lens flare, and processing the input image by a machine learning model to generate a de-flared image that includes the input image with at least part of the particular representation of lens flare removed. The machine learning (ML) model may be trained by generating training images that combine respective baseline images with corresponding lens flare images. For each respective training image, a modified image may be determined by processing the respective training image by the ML model, and a loss value may be determined based on a loss function comparing the modified image to a corresponding baseline image used to generate the respective training image. Parameters of the ML model may be adjusted based on the loss value determined for each respective training image and the loss function.

    Learning-Based Lens Flare Removal

    公开(公告)号:US20220375045A1

    公开(公告)日:2022-11-24

    申请号:US17625994

    申请日:2020-11-09

    Applicant: Google LLC

    Abstract: A method includes obtaining an input image that contains a particular representation of lens flare, and processing the input image by a machine learning model to generate a de-flared image that includes the input image with at least part of the particular representation of lens flare removed. The machine learning (ML) model may be trained by generating training images that combine respective baseline images with corresponding lens flare images. For each respective training image, a modified image may be determined by processing the respective training image by the ML model, and a loss value may be determined based on a loss function comparing the modified image to a corresponding baseline image used to generate the respective training image. Parameters of the ML model may be adjusted based on the loss value determined for each respective training image and the loss function.

    Defocus Blur Removal and Depth Estimation Using Dual-Pixel Image Data

    公开(公告)号:US20220375042A1

    公开(公告)日:2022-11-24

    申请号:US17626069

    申请日:2020-11-13

    Applicant: Google LLC

    Abstract: A method includes obtaining dual-pixel image data that includes a first sub-image and a second sub-image, and generating an in-focus image, a first kernel corresponding to the first sub-image, and a second kernel corresponding to the second sub-image. A loss value may be determined using a loss function that determines a difference between (i) a convolution of the first sub-image with the second kernel and (ii) a convolution of the second sub-image with the first kernel, and/or a sum of (i) a difference between the first sub-image and a convolution of the in-focus image with the first kernel and (ii) a difference between the second sub-image and a convolution of the in-focus image with the second kernel. Based on the loss value and the loss function, the in-focus image, the first kernel, and/or the second kernel, may be updated and displayed.

    Learning-based lens flare removal

    公开(公告)号:US12033309B2

    公开(公告)日:2024-07-09

    申请号:US17625994

    申请日:2020-11-09

    Applicant: Google LLC

    Abstract: A method includes obtaining an input image that contains a particular representation of lens flare, and processing the input image by a machine learning model to generate a de-flared image that includes the input image with at least part of the particular representation of lens flare removed. The machine learning (ML) model may be trained by generating training images that combine respective baseline images with corresponding lens flare images. For each respective training image, a modified image may be determined by processing the respective training image by the ML model, and a loss value may be determined based on a loss function comparing the modified image to a corresponding baseline image used to generate the respective training image. Parameters of the ML model may be adjusted based on the loss value determined for each respective training image and the loss function.

    Defocus blur removal and depth estimation using dual-pixel image data

    公开(公告)号:US12008738B2

    公开(公告)日:2024-06-11

    申请号:US17626069

    申请日:2020-11-13

    Applicant: Google LLC

    CPC classification number: G06T5/73 G06T5/50 G06T7/50

    Abstract: A method includes obtaining dual-pixel image data that includes a first sub-image and a second sub-image, and generating an in-focus image, a first kernel corresponding to the first sub-image, and a second kernel corresponding to the second sub-image. A loss value may be determined using a loss function that determines a difference between (i) a convolution of the first sub-image with the second kernel and (ii) a convolution of the second sub-image with the first kernel, and/or a sum of (i) a difference between the first sub-image and a convolution of the in-focus image with the first kernel and (ii) a difference between the second sub-image and a convolution of the in-focus image with the second kernel. Based on the loss value and the loss function, the in-focus image, the first kernel, and/or the second kernel, may be updated and displayed.

    Systems and Methods for Manipulation of Shadows on Portrait Image Frames

    公开(公告)号:US20230351560A1

    公开(公告)日:2023-11-02

    申请号:US17786841

    申请日:2019-12-23

    Applicant: Google LLC

    CPC classification number: G06T5/008 G06T5/50 G06T2207/20081 G06T2207/30201

    Abstract: Systems and methods described herein may relate to potential methods of training a machine learning model to be implemented on a mobile computing device configured to capture, adjust, and/or store image frames. An example method includes supplying a first image frame of a subject in a setting lit within a first lighting environment and supplying a second image frame of the subject lit within a second lighting environment. The method further includes determining a mask. Additionally, the method includes combining the first image frame and the second image frame according to the mask to generate a synthetic image and assigning a score to the synthetic image. The method also includes training a machine learning model based on the assigned score to adjust a captured image based on the synthetic image.

Patent Agency Ranking