-
公开(公告)号:US20230419579A1
公开(公告)日:2023-12-28
申请号:US18462310
申请日:2023-09-06
Applicant: Meta Platforms Technologies, LLC
Inventor: Alexander Richard , Michael Zollhoefer , Fernando De la Torre , Yaser Sheikh
CPC classification number: G06T13/205 , G06T13/40 , G06T17/20 , G06T19/006 , G10L21/14 , G10L2021/105
Abstract: A method for training a three-dimensional model face animation model from speech, is provided. The method includes determining a first correlation value for a facial feature based on an audio waveform from a first subject, generating a first mesh for a lower portion of a human face, based on the facial feature and the first correlation value, updating the first correlation value when a difference between the first mesh and a ground truth image of the first subject is greater than a pre-selected threshold, and providing a three-dimensional model of the human face animated by speech to an immersive reality application accessed by a client device based on the difference between the first mesh and the ground truth image of the first subject. A non-transitory, computer-readable medium storing instructions to cause a system to perform the above method, and the system, are also provided.
-
公开(公告)号:US11756250B2
公开(公告)日:2023-09-12
申请号:US17669270
申请日:2022-02-10
Applicant: Meta Platforms Technologies, LLC
Inventor: Alexander Richard , Michael Zollhoefer , Fernando De la Torre , Yaser Sheikh
CPC classification number: G06T13/205 , G06T13/40 , G06T17/20 , G06T19/006 , G10L21/14 , G10L2021/105
Abstract: A method for training a three-dimensional model face animation model from speech, is provided. The method includes determining a first correlation value for a facial feature based on an audio waveform from a first subject, generating a first mesh for a lower portion of a human face, based on the facial feature and the first correlation value, updating the first correlation value when a difference between the first mesh and a ground truth image of the first subject is greater than a pre-selected threshold, and providing a three-dimensional model of the human face animated by speech to an immersive reality application accessed by a client device based on the difference between the first mesh and the ground truth image of the first subject. A non-transitory, computer-readable medium storing instructions to cause a system to perform the above method, and the system, are also provided.
-
公开(公告)号:US20240303951A1
公开(公告)日:2024-09-12
申请号:US18636687
申请日:2024-04-16
Applicant: Meta Platforms Technologies, LLC
Inventor: Stephen Anthony Lombardi , Tomas Simon Kreuz , Jason Saragih , Gabriel Bailowitz Schwartz , Michael Zollhoefer , Yaser Sheikh
CPC classification number: G06T19/20 , G06T15/06 , G06T17/20 , G06T2219/2012
Abstract: A method for training a real-time, modeling for animating an avatar for a subject is provided. The method includes collecting multiple images of a subject. The method also includes selecting a plurality of vertex positions in a guide mesh, indicative of a volumetric primitive enveloping the subject, determining a geometric attribute for the volumetric primitive including a position, a rotation, and a scale factor of the volumetric primitive, determining a payload attribute for each of the volumetric primitive, the payload attribute including a color value and an opacity value for each voxel in a voxel grid defining the volumetric primitive, determining a loss factor for each point in the volumetric primitive based on the geometric attribute, the payload attribute and a ground truth value, and updating a three-dimensional model for the subject. A system and a non-transitory, computer-readable medium storing instructions to perform the above method are also provided.
-
公开(公告)号:US20250131631A1
公开(公告)日:2025-04-24
申请号:US18914741
申请日:2024-10-14
Applicant: Meta Platforms Technologies, LLC
Inventor: Alexander Richard , Michael Zollhoefer , Fernando De la Torre , Yaser Sheikh
Abstract: A method for training a three-dimensional model face animation model from speech, is provided. The method includes determining a first correlation value for a facial feature based on an audio waveform from a first subject, generating a first mesh for a lower portion of a human face, based on the facial feature and the first correlation value, updating the first correlation value when a difference between the first mesh and a ground truth image of the first subject is greater than a pre-selected threshold, and providing a three-dimensional model of the human face animated by speech to an immersive reality application accessed by a client device based on the difference between the first mesh and the ground truth image of the first subject. A non-transitory, computer-readable medium storing instructions to cause a system to perform the above method, and the system, are also provided.
-
公开(公告)号:US12131416B2
公开(公告)日:2024-10-29
申请号:US17556367
申请日:2021-12-20
Applicant: Meta Platforms Technologies, LLC
Inventor: Stephen Anthony Lombardi , Jason Saragih , Tomas Simon Kreuz , Shunsuke Saito , Michael Zollhoefer , Amit Raj , James Henry Hays
CPC classification number: G06T13/40 , G06T7/73 , G06T7/97 , G06T2207/30201
Abstract: A method of forming a pixel-aligned volumetric avatar includes receiving multiple two-dimensional images having at least two or more fields of view of a subject. The method also includes extracting multiple image features from the two-dimensional images using a set of learnable weights, projecting the image features along a direction between a three-dimensional model of the subject and a selected observation point for a viewer, and providing, to the viewer, an image of the three-dimensional model of the subject. A system and a non-transitory, computer readable medium storing instructions to perform the above method, are also provided.
-
公开(公告)号:US12159339B2
公开(公告)日:2024-12-03
申请号:US18462310
申请日:2023-09-06
Applicant: Meta Platforms Technologies, LLC
Inventor: Alexander Richard , Michael Zollhoefer , Fernando De la Torre , Yaser Sheikh
Abstract: A method for training a three-dimensional model face animation model from speech, is provided. The method includes determining a first correlation value for a facial feature based on an audio waveform from a first subject, generating a first mesh for a lower portion of a human face, based on the facial feature and the first correlation value, updating the first correlation value when a difference between the first mesh and a ground truth image of the first subject is greater than a pre-selected threshold, and providing a three-dimensional model of the human face animated by speech to an immersive reality application accessed by a client device based on the difference between the first mesh and the ground truth image of the first subject. A non-transitory, computer-readable medium storing instructions to cause a system to perform the above method, and the system, are also provided.
-
公开(公告)号:US12095975B2
公开(公告)日:2024-09-17
申请号:US17555037
申请日:2021-12-17
Applicant: Meta Platforms Technologies, LLC
Inventor: Nathan Matsuda , Brian Wheelwright , Joel Hegland , Stephen Anthony Lombardi , Jason Saragih , Tomas Simon Kreuz , Shunsuke Saito , Michael Zollhoefer , Amit Raj , James Henry Hays
IPC: H04N13/344 , G02B30/10 , G06T7/80 , G06T15/00 , H04N5/33 , H04N13/117 , H04N13/239 , H04N13/383
CPC classification number: H04N13/344 , G02B30/10 , G06T7/80 , G06T15/00 , H04N5/33 , H04N13/117 , H04N13/239 , H04N13/383 , G06T2207/20081
Abstract: A device for providing a reverse pass-through view of a user of a headset display to an onlooker includes an eyepiece comprising an optical surface configured to provide an image to a user on a first side of the optical surface. The device also includes a first camera configured to collect an image of a portion of a face of the user reflected from the optical surface in a first field of view, a display adjacent to the optical surface and configured to project forward an image of the face of the user, and a screen configured to receive light from the display and provide the image of the face of the user to an onlooker.
-
公开(公告)号:US20230245365A1
公开(公告)日:2023-08-03
申请号:US18074346
申请日:2022-12-02
Applicant: Meta Platforms Technologies, LLC
Inventor: Chen Cao , Stuart Anderson , Tomas Simon Kreuz , Jin Kyu Kim , Gabriel Bailowitz Schwartz , Michael Zollhoefer , Shunsuke Saito , Stephen Anthony Lombardi , Shih-En Wei , Danielle Belko , Shoou-I Yu , Yaser Sheikh , Jason Saragih
IPC: G06T13/40
CPC classification number: G06T13/40
Abstract: A method for generating a subject avatar using a mobile phone scan is provided. The method includes receiving, from a mobile device, multiple images of a first subject, extracting multiple image features from the images of the first subject based on a set of learnable weights, inferring a three-dimensional model of the first subject from the image features and an existing three-dimensional model of a second subject, animating the three-dimensional model of the first subject based on an immersive reality application running on a headset used by a viewer, and providing, to a display on the headset, an image of the three-dimensional model of the first subject. A system and a non-transitory, computer-readable medium storing instructions to perform the above method, are also provided.
-
公开(公告)号:US12243273B2
公开(公告)日:2025-03-04
申请号:US17571285
申请日:2022-01-07
Applicant: META PLATFORMS TECHNOLOGIES, LLC
Inventor: Zhaoyang Lv , Miroslava Slavcheva , Tianye Li , Michael Zollhoefer , Simon Gareth Green , Tanner Schmidt , Michael Goesele , Steven John Lovegrove , Christoph Lassner , Changil Kim
IPC: G06T7/00
Abstract: In one embodiment, a method includes initializing latent codes respectively associated with times associated with frames in a training video of a scene captured by a camera. For each of the frames, a system (1) generates rendered pixel values for a set of pixels in the frame by querying NeRF using the latent code associated with the frame, a camera viewpoint associated with the frame, and ray directions associated with the set of pixels, and (2) updates the latent code associated with the frame and the NeRF based on comparisons between the rendered pixel values and original pixel values for the set of pixels. Once trained, the system renders output frames for an output video of the scene, wherein each output frame is rendered by querying the updated NeRF using one of the updated latent codes corresponding to a desired time associated with the output frame.
-
公开(公告)号:US20240406371A1
公开(公告)日:2024-12-05
申请号:US18800312
申请日:2024-08-12
Applicant: Meta Platforms Technologies, LLC
Inventor: Nathan Matsuda , Brian Wheelwright , Joel Hegland , Stephen Anthony Lombardi , Jason Saragih , Tomas Simon Kreuz , Shunsuke Saito , Michael Zollhoefer , Amit Raj , James Henry Hays
IPC: H04N13/344 , G02B30/10 , G06T7/80 , G06T15/00 , H04N5/33 , H04N13/117 , H04N13/239 , H04N13/383
Abstract: A device for providing a reverse pass-through view of a user of a headset display to an onlooker includes an eyepiece comprising an optical surface configured to provide an image to a user on a first side of the optical surface. The device also includes a first camera configured to collect an image of a portion of a face of the user reflected from the optical surface in a first field of view, a display adjacent to the optical surface and configured to project forward an image of the face of the user, and a screen configured to receive light from the display and provide the image of the face of the user to an onlooker.
-
-
-
-
-
-
-
-
-