-
公开(公告)号:US11589031B2
公开(公告)日:2023-02-21
申请号:US16580802
申请日:2019-09-24
Applicant: Google LLC
Inventor: Sameh Khamis , Yinda Zhang , Christoph Rhemann , Julien Valentin , Adarsh Kowdle , Vladimir Tankovich , Michael Schoenberg , Shahram Izadi , Thomas Funkhouser , Sean Fanello
IPC: H04N13/271 , G06T7/90 , G06T7/521 , H04N5/33 , H04N13/239
Abstract: An electronic device estimates a depth map of an environment based on matching reduced-resolution stereo depth images captured by depth cameras to generate a coarse disparity (depth) map. The electronic device downsamples depth images captured by the depth cameras and matches sections of the reduced-resolution images to each other to generate a coarse depth map. The electronic device upsamples the coarse depth map to a higher resolution and refines the upsampled depth map to generate a high-resolution depth map to support location-based functionality.
-
2.
公开(公告)号:US10997457B2
公开(公告)日:2021-05-04
申请号:US16616235
申请日:2019-10-16
Applicant: Google LLC
Inventor: Christoph Rhemann , Abhimitra Meka , Matthew Whalen , Jessica Lynn Busch , Sofien Bouaziz , Geoffrey Douglas Harvey , Andrea Tagliasacchi , Jonathan Taylor , Paul Debevec , Peter Joseph Denny , Sean Ryan Francesco Fanello , Graham Fyffe , Jason Angelo Dourgarian , Xueming Yu , Adarsh Prakash Murthy Kowdle , Julien Pascal Christophe Valentin , Peter Christopher Lincoln , Rohit Kumar Pandey , Christian Häne , Shahram Izadi
Abstract: Methods, systems, and media for relighting images using predicted deep reflectance fields are provided. In some embodiments, the method comprises: identifying a group of training samples, wherein each training sample includes (i) a group of one-light-at-a-time (OLAT) images that have each been captured when one light of a plurality of lights arranged on a lighting structure has been activated, (ii) a group of spherical color gradient images that have each been captured when the plurality of lights arranged on the lighting structure have been activated to each emit a particular color, and (iii) a lighting direction, wherein each image in the group of OLAT images and each of the spherical color gradient images are an image of a subject, and wherein the lighting direction indicates a relative orientation of a light to the subject; training a convolutional neural network using the group of training samples, wherein training the convolutional neural network comprises: for each training iteration in a series of training iterations and for each training sample in the group of training samples: generating an output predicted image, wherein the output predicted image is a representation of the subject associated with the training sample with lighting from the lighting direction associated with the training sample; identifying a ground-truth OLAT image included in the group of OLAT images for the training sample that corresponds to the lighting direction for the training sample; calculating a loss that indicates a perceptual difference between the output predicted image and the identified ground-truth OLAT image; and updating parameters of the convolutional neural network based on the calculated loss; identifying a test sample that includes a second group of spherical color gradient images and a second lighting direction; and generating a relit image of the subject included in each of the second group of spherical color gradient images with lighting from the second lighting direction using the trained convolutional neural network.
-
公开(公告)号:US10554957B2
公开(公告)日:2020-02-04
申请号:US15996880
申请日:2018-06-04
Applicant: Google LLC
Inventor: Julien Pascal Christophe Valentin , Sean Ryan Fanello , Adarsh Prakash Murthy Kowdle , Christoph Rhemann , Vladimir Tankovich , Philip L. Davidson , Shahram Izadi
IPC: G06K9/64 , H04N13/271 , H04N19/597 , H04N13/128 , G06K9/62 , G06T7/593 , H04N13/00
Abstract: A first and second image of a scene are captured. Each of a plurality of pixels in the first image is associated with a disparity value. An image patch associated with each of the plurality of pixels of the first image and the second image is mapped into a binary vector. Thus, values of pixels in an image are mapped to a binary space using a function that preserves characteristics of values of the pixels. The difference between the binary vector associated with each of the plurality of pixels of the first image and its corresponding binary vector in the second image designated by the disparity value associated with each of the plurality of pixels of the first image is determined. Based on the determined difference between binary vectors, correspondence between the plurality of pixels of the first image and the second image is established.
-
公开(公告)号:US20180352213A1
公开(公告)日:2018-12-06
申请号:US15996880
申请日:2018-06-04
Applicant: Google LLC
Inventor: Julien Pascal Christophe Valentin , Sean Ryan Fanello , Adarsh Prakash Murthy Kowdle , Christoph Rhemann , Vladimir Tankovich , Philip L. Davidson , Shahram Izadi
IPC: H04N13/271 , H04N19/597 , G06K9/62 , H04N13/128
CPC classification number: H04N13/271 , G06K9/6268 , G06T7/593 , G06T2207/20081 , H04N13/128 , H04N19/597 , H04N2013/0081
Abstract: A first and second image of a scene are captured. Each of a plurality of pixels in the first image is associated with a disparity value. An image patch associated with each of the plurality of pixels of the first image and the second image is mapped into a binary vector. Thus, values of pixels in an image are mapped to a binary space using a function that preserves characteristics of values of the pixels. The difference between the binary vector associated with each of the plurality of pixels of the first image and its corresponding binary vector in the second image designated by the disparity value associated with each of the plurality of pixels of the first image is determined. Based on the determined difference between binary vectors, correspondence between the plurality of pixels of the first image and the second image is established.
-
公开(公告)号:US11145075B2
公开(公告)日:2021-10-12
申请号:US16767401
申请日:2019-10-04
Applicant: Google LLC
Inventor: Julien Valentin , Onur G. Guleryuz , Mira Leung , Maksym Dzitsiuk , Jose Pascoal , Mirko Schmidt , Christoph Rhemann , Neal Wadhwa , Eric Turner , Sameh Khamis , Adarsh Prakash Murthy Kowdle , Ambrus Csaszar , João Manuel Castro Afonso , Jonathan T. Barron , Michael Schoenberg , Ivan Dryanovski , Vivek Verma , Vladimir Tankovich , Shahram Izadi , Sean Ryan Francesco Fanello , Konstantine Nicholas John Tsotsos
Abstract: A handheld user device includes a monocular camera to capture a feed of images of a local scene and a processor to select, from the feed, a keyframe and perform, for a first image from the feed, stereo matching using the first image, the keyframe, and a relative pose based on a pose associated with the first image and a pose associated with the keyframe to generate a sparse disparity map representing disparities between the first image and the keyframe. The processor further is to determine a dense depth map from the disparity map using a bilateral solver algorithm, and process a viewfinder image generated from a second image of the feed with occlusion rendering based on the depth map to incorporate one or more virtual objects into the viewfinder image to generate an AR viewfinder image. Further, the processor is to provide the AR viewfinder image for display.
-
6.
公开(公告)号:US20200372284A1
公开(公告)日:2020-11-26
申请号:US16616235
申请日:2019-10-16
Applicant: Google LLC
Inventor: Christoph Rhemann , Abhimitra Meka , Matthew Whalen , Jessica Lynn Busch , Sofien Bouaziz , Geoffrey Douglas Harvey , Andrea Tagliasacchi , Jonathan Taylor , Paul Debevec , Peter Joseph Denny , Sean Ryan Francesco Fanello , Graham Fyffe , Jason Angelo Dourgarian , Xueming Yu , Adarsh Prakash Murthy Kowdle , Julien Pascal Christophe Valentin , Peter Christopher Lincoln , Rohit Kumar Pandey , Christian Häne , Shahram Izadi
Abstract: Methods, systems, and media for relighting images using predicted deep reflectance fields are provided. In some embodiments, the method comprises: identifying a group of training samples, wherein each training sample includes (i) a group of one-light-at-a-time (OLAT) images that have each been captured when one light of a plurality of lights arranged on a lighting structure has been activated, (ii) a group of spherical color gradient images that have each been captured when the plurality of lights arranged on the lighting structure have been activated to each emit a particular color, and (iii) a lighting direction, wherein each image in the group of OLAT images and each of the spherical color gradient images are an image of a subject, and wherein the lighting direction indicates a relative orientation of a light to the subject; training a convolutional neural network using the group of training samples, wherein training the convolutional neural network comprises: for each training iteration in a series of training iterations and for each training sample in the group of training samples: generating an output predicted image, wherein the output predicted image is a representation of the subject associated with the training sample with lighting from the lighting direction associated with the training sample; identifying a ground-truth OLAT image included in the group of OLAT images for the training sample that corresponds to the lighting direction for the training sample; calculating a loss that indicates a perceptual difference between the output predicted image and the identified ground-truth OLAT image; and updating parameters of the convolutional neural network based on the calculated loss; identifying a test sample that includes a second group of spherical color gradient images and a second lighting direction; and generating a relit image of the subject included in each of the second group of spherical color gradient images with lighting from the second lighting direction using the trained convolutional neural network.
-
公开(公告)号:US11037026B2
公开(公告)日:2021-06-15
申请号:US16749626
申请日:2020-01-22
Applicant: Google LLC
Inventor: Sean Ryan Fanello , Julien Pascal Christophe Valentin , Adarsh Prakash Murthy Kowdle , Christoph Rhemann , Vladimir Tankovich , Philip L. Davidson , Shahram Izadi
IPC: G06K9/62
Abstract: Values of pixels in an image are mapped to a binary space using a first function that preserves characteristics of values of the pixels. Labels are iteratively assigned to the pixels in the image in parallel based on a second function. The label assigned to each pixel is determined based on values of a set of nearest-neighbor pixels. The first function is trained to map values of pixels in a set of training images to the binary space and the second function is trained to assign labels to the pixels in the set of training images. Considering only the nearest neighbors in the inference scheme results in a computational complexity that is independent of the size of the solution space and produces sufficient approximations of the true distribution when the solution for each pixel is most likely found in a small subset of the set of potential solutions.
-
公开(公告)号:US10579905B2
公开(公告)日:2020-03-03
申请号:US15925141
申请日:2018-03-19
Applicant: Google LLC
Inventor: Sean Ryan Fanello , Julien Pascal Christophe Valentin , Adarsh Prakash Murthy Kowdle , Christoph Rhemann , Vladimir Tankovich , Philip L. Davidson , Shahram Izadi
IPC: G06K9/62
Abstract: Values of pixels in an image are mapped to a binary space using a first function that preserves characteristics of values of the pixels. Labels are iteratively assigned to the pixels in the image in parallel based on a second function. The label assigned to each pixel is determined based on values of a set of nearest-neighbor pixels. The first function is trained to map values of pixels in a set of training images to the binary space and the second function is trained to assign labels to the pixels in the set of training images. Considering only the nearest neighbors in the inference scheme results in a computational complexity that is independent of the size of the solution space and produces sufficient approximations of the true distribution when the solution for each pixel is most likely found in a small subset of the set of potential solutions.
-
公开(公告)号:US12066282B2
公开(公告)日:2024-08-20
申请号:US17413847
申请日:2020-11-11
Applicant: GOOGLE LLC
Inventor: Sean Ryan Francesco Fanello , Kaiwen Guo , Peter Christopher Lincoln , Philip Lindsley Davidson , Jessica L. Busch , Xueming Yu , Geoffrey Harvey , Sergio Orts Escolano , Rohit Kumar Pandey , Jason Dourgarian , Danhang Tang , Adarsh Prakash Murthy Kowdle , Emily B. Cooper , Mingsong Dou , Graham Fyffe , Christoph Rhemann , Jonathan James Taylor , Shahram Izadi , Paul Ernest Debevec
IPC: G01B11/25 , G01B11/245 , G06T15/50 , G06T17/20
CPC classification number: G01B11/2513 , G01B11/245 , G06T15/506 , G06T17/205
Abstract: A lighting stage includes a plurality of lights that project alternating spherical color gradient illumination patterns onto an object or human performer at a predetermined frequency. The lighting stage also includes a plurality of cameras that capture images of an object or human performer corresponding to the alternating spherical color gradient illumination patterns. The lighting stage also includes a plurality of depth sensors that capture depth maps of the object or human performer at the predetermined frequency. The lighting stage also includes (or is associated with) one or more processors that implement a machine learning algorithm to produce a three-dimensional (3D) model of the object or human performer. The 3D model includes relighting parameters used to relight the 3D model under different lighting conditions.
-
公开(公告)号:US20230360182A1
公开(公告)日:2023-11-09
申请号:US18028930
申请日:2021-05-17
Applicant: Google LLC
Inventor: Sean Ryan Francesco Fanello , Yun-Ta Tsai , Rohit Kumar Pandey , Paul Debevec , Michael Milne , Chloe LeGendre , Jonathan Tilton Barron , Christoph Rhemann , Sofien Bouaziz , Navin Padman Sarma
CPC classification number: G06T5/009 , G06T7/60 , G06T7/70 , G06T15/506 , G06T2200/24 , G06T2207/20081 , G06T2207/20084 , G06T2207/20092 , G06T2207/20208 , G06T2207/30201
Abstract: Apparatus and methods related to applying lighting models to images of objects are provided. An example method includes applying a geometry model to an input image to determine a surface orientation map indicative of a distribution of lighting on an object based on a surface geometry. The method further includes applying an environmental light estimation model to the input image to determine a direction of synthetic lighting to be applied to the input image. The method also includes applying, based on the surface orientation map and the direction of synthetic lighting, a light energy model to determine a quotient image indicative of an amount of light energy to be applied to each pixel of the input image. The method additionally includes enhancing, based on the quotient image, a portion of the input image. One or more neural networks can be trained to perform one or more of the aforementioned aspects.
-
-
-
-
-
-
-
-
-