-
公开(公告)号:US12026833B2
公开(公告)日:2024-07-02
申请号:US17310678
申请日:2020-10-28
Applicant: Google LLC
Inventor: Ricardo Martin Brualla , Moustafa Meshry , Daniel Goldman , Rohit Kumar Pandey , Sofien Bouaziz , Ke Li
CPC classification number: G06T17/20 , G06T7/40 , G06T15/04 , G06T2207/10028 , G06T2207/20081 , G06T2207/30201
Abstract: Systems and methods are described for utilizing an image processing system with at least one processing device to perform operations including receiving a plurality of input images of a user, generating a three-dimensional mesh proxy based on a first set of features extracted from the plurality of input images and a second set of features extracted from the plurality of input images. The method may further include generating a neural texture based on a three-dimensional mesh proxy and the plurality of input images, generating a representation of the user including at least a neural texture, and sampling at least one portion of the neural texture from the three-dimensional mesh proxy. In response to providing the at least one sampled portion to a neural renderer, the method may include receiving, from the neural renderer, a synthesized image of the user that is previously not captured by the image processing system.
-
公开(公告)号:US20240005590A1
公开(公告)日:2024-01-04
申请号:US18251995
申请日:2021-01-14
Applicant: GOOGLE LLC
Inventor: Ricardo Martin Brualla , Keunhong Park , Utkarsh Sinha , Sofien Bouaziz , Daniel Goldman , Jonathan Tilton Barron , Steven Maxwell Seitz
Abstract: Techniques of image synthesis using a neural radiance field (NeRF) includes generating a deformation model of movement experienced by a subject in a non-rigidly deforming scene. For example, when an image synthesis system uses NeRFs, the system takes as input multiple poses of subjects for training data. In contrast to conventional NeRFs, the technical solution first expresses the positions of the subjects from various perspectives in an observation frame. The technical solution then involves deriving a deformation model, i.e., a mapping between the observation frame and a canonical frame in which the subject's movements are taken into account. This mapping is accomplished using latent deformation codes for each pose that are determined using a multilayer perceptron (MLP). A NeRF is then derived from positions and casted ray directions in the canonical frame using another MLP. New poses for the subject may then be derived using the NeRF.
-
公开(公告)号:US20220398705A1
公开(公告)日:2022-12-15
申请号:US17754392
申请日:2021-04-08
Applicant: GOOGLE LLC
Inventor: Ricardo Martin Brualla , Daniel Goldman , Hugues Herve Hoppe , Lynn Tsai , Lars Peter Johannes Hedman
Abstract: Systems and methods are described for receiving a plurality of input images, a plurality of depth images, and a plurality of view parameters for generating a virtual view of a target subject. The systems and methods may generate a plurality of warped images based on the plurality of input images, the plurality of view parameters, and at least one of the plurality of depth images. In response to providing the plurality of depth images, the plurality of view parameters, and the plurality of warped images to a neural network, the systems and methods may receive, from the neural network, blending weights for assigning color to pixels of the virtual view of the target subject and may generate, based on the blending weights and the virtual view, a synthesized image according to the view parameters.
-
公开(公告)号:US20220014723A1
公开(公告)日:2022-01-13
申请号:US17309440
申请日:2019-12-02
Applicant: Google LLC
Inventor: Rohit Pandey , Jonathan Taylor , Ricardo Martin Brualla , Shuoran Yang , Pavlo Pidlypenskyi , Daniel Goldman , Sean Ryan Francesco Fanello
IPC: H04N13/111 , H04N13/161 , H04N13/366 , H04N13/243 , H04N13/332 , G06T7/174 , G06K9/32 , G06K9/00 , G06K9/46
Abstract: Three-dimensional (3D) performance capture and machine learning can be used to re-render high quality novel viewpoints of a captured scene. A textured 3D reconstruction is first rendered to a novel viewpoint. Due to imperfections in geometry and low-resolution texture, the 2D rendered image contains artifacts and is low quality. Accordingly, a deep learning technique is disclosed that takes these images as input and generates more visually enhanced re-rendering. The system is specifically designed for VR and AR headsets, and accounts for consistency between two stereo views.
-
公开(公告)号:US11710287B2
公开(公告)日:2023-07-25
申请号:US17309817
申请日:2020-08-04
Applicant: GOOGLE LLC
Inventor: Ricardo Martin Brualla , Daniel Goldman , Sofien Bouaziz , Rohit Kumar Pandey , Matthew Brown
CPC classification number: G06T19/20 , G06T15/005 , G06T15/04 , G06T15/506 , G06V10/95 , G06T2219/2012 , G06T2219/2021
Abstract: Systems and methods are described for generating a plurality of three-dimensional (3D) proxy geometries of an object, generating, based on the plurality of 3D proxy geometries, a plurality of neural textures of the object, the neural textures defining a plurality of different shapes and appearances representing the object, providing the plurality of neural textures to a neural renderer, receiving, from the neural renderer and based on the plurality of neural textures, a color image and an alpha mask representing an opacity of at least a portion of the object, and generating a composite image based on the pose, the color image, and the alpha mask.
-
公开(公告)号:US11190746B2
公开(公告)日:2021-11-30
申请号:US16710655
申请日:2019-12-11
Applicant: GOOGLE LLC
Inventor: Daniel Goldman , Harris Nover , Supreeth Achar
IPC: H04N13/106 , H04N13/254 , G06T7/593 , G06T7/11 , G06T5/00 , G06T7/174 , H04N13/00
Abstract: According to an aspect, a real-time active stereo system includes a capture system configured to capture stereo image data, where the image data includes a plurality of pairs of a reference image and a secondary image, and each pair of the plurality of pairs relates a different temporal window. The real-time active stereo system includes a depth sensing computing system including at least one processor and a non-transitory computer-readable medium having executable instructions that when executed by the at least one processor are configured to execute a local stereo reconstruction algorithm configured to compute spacetime descriptors from the plurality of pairs of the stereo image data and generate depth maps based on the spacetime descriptors.
-
公开(公告)号:US20190306541A1
公开(公告)日:2019-10-03
申请号:US16443481
申请日:2019-06-17
Applicant: Google LLC
Inventor: Daniel Goldman , Jason Lawrence , Andrew Huibers , Andrew Ian Russell , Steven M. Seitz
IPC: H04N21/222 , H04N7/15 , H04N21/2187 , G02B27/01 , H04N7/14 , H04N21/61 , H04N21/43 , H04N21/4223 , H04N21/242 , G02F1/29 , H04R1/02 , H04R3/00 , H04N13/305 , H04N21/431 , H04N21/239 , H04N21/2365 , H04N21/854 , H04N21/2343
Abstract: An example telepresence terminal includes a lenticular display, an image sensor, an infrared emitter, and an infrared depth sensor. The terminal may determine image data using visible light emitted by the infrared emitter and captured by the image sensor and determine depth data using infrared light captured by the infrared depth sensor. The terminal may also communicate the depth data and the image data to a remote telepresence terminal and receive remote image data and remote depth data. The terminal may also generate a first display image using the lenticular display based on the remote image data that is viewable from a first viewing location and generate a second display image using the lenticular display based on the remote image data and the remote depth data that is viewable from a second viewing location.
-
公开(公告)号:US10327014B2
公开(公告)日:2019-06-18
申请号:US15699651
申请日:2017-09-08
Applicant: Google LLC
Inventor: Daniel Goldman , Jason Lawrence , Andrew Huibers , Andrew Ian Russell , Steven M. Seitz
IPC: G02F1/29 , H04N7/14 , H04N7/15 , H04R1/02 , H04R3/00 , G02B27/01 , H04N21/41 , H04N21/43 , H04N21/61 , H04N13/305 , H04N21/222 , H04N21/239 , H04N21/242 , H04N21/431 , H04N21/854 , H04N21/2187 , H04N21/2343 , H04N21/2365 , H04N21/4223
Abstract: An example telepresence terminal includes a lenticular display, an image sensor, an infrared emitter, and an infrared depth sensor. The terminal may determine image data using visible light emitted by the infrared emitter and captured by the image sensor and determine depth data using infrared light captured by the infrared depth sensor. The terminal may also communicate the depth data and the image data to a remote telepresence terminal and receive remote image data and remote depth data. The terminal may also generate a first display image using the lenticular display based on the remote image data that is viewable from a first viewing location and generate a second display image using the lenticular display based on the remote image data and the remote depth data that is viewable from a second viewing location.
-
公开(公告)号:US20220277485A1
公开(公告)日:2022-09-01
申请号:US17597428
申请日:2021-01-19
Applicant: Google LLC
Inventor: Supreeth Achar , Daniel Goldman
Abstract: A method including capturing a first image of a real-world scene by a first camera sensitive to a first spectrum of light, the first camera having a first light source, capturing a second image of the real-world scene by a second camera sensitive to a second spectrum of light, the second camera having a second light source, identifying at least one feature in the first image, identifying, using a machine learning (ML) model, at least one feature in the second image that matches the at least one feature identified in the first image, mapping pixels in the first image and the second image to rays in a three-dimensional (3D) space based on the matched at least one feature, and calibrating the first camera and the second camera based on the mapping.
-
公开(公告)号:US11288857B2
公开(公告)日:2022-03-29
申请号:US16837612
申请日:2020-04-01
Applicant: Google LLC
Inventor: Moustafa Meshry , Ricardo Martin Brualla , Sameh Khamis , Daniel Goldman , Hugues Hoppe , Noah Snavely , Rohit Pandey
Abstract: According to an aspect, a method for neural rerendering includes obtaining a three-dimensional (3D) model representing a scene of a physical space, where the 3D model is constructed from a collection of input images, rendering an image data buffer from the 3D model according to a viewpoint, where the image data buffer represents a reconstructed image from the 3D model, receiving, by a neural rerendering network, the image data buffer, receiving, by the neural rerendering network, an appearance code representing an appearance condition, and transforming, by the neural rerendering network, the image data buffer into a rerendered image with the viewpoint of the image data buffer and the appearance condition specified by the appearance code.
-
-
-
-
-
-
-
-
-