Photorealistic talking faces from audio

    公开(公告)号:US12033259B2

    公开(公告)日:2024-07-09

    申请号:US17796399

    申请日:2021-01-29

    Applicant: Google LLC

    CPC classification number: G06T13/205 G06T13/40 G06T17/20

    Abstract: Provided is a framework for generating photorealistic 3D talking faces conditioned only on audio input. In addition, the present disclosure provides associated methods to insert generated faces into existing videos or virtual environments. We decompose faces from video into a normalized space that decouples 3D geometry, head pose, and texture. This allows separating the prediction problem into regressions over the 3D face shape and the corresponding 2D texture atlas. To stabilize temporal dynamics, we propose an auto-regressive approach that conditions the model on its previous visual state. We also capture face illumination in our model using audio-independent 3D texture normalization.

    System and method for grouping related photographs

    公开(公告)号:US10514818B2

    公开(公告)日:2019-12-24

    申请号:US15092102

    申请日:2016-04-06

    Applicant: Google LLC

    Abstract: A computer-implemented method, computer program product, and computing system is provided for interacting with images having similar content. In an embodiment, a method may include identifying a plurality of photographs as including a common characteristic. The method may also include generating a flipbook media item including the plurality of photographs. The method may further include associating one or more interactive control features with the flipbook media item.

    Photorealistic Talking Faces from Audio
    3.
    发明公开

    公开(公告)号:US20230343010A1

    公开(公告)日:2023-10-26

    申请号:US17796399

    申请日:2021-01-29

    Applicant: Google LLC

    CPC classification number: G06T13/205 G06T13/40 G06T17/20

    Abstract: Provided is a framework for generating photorealistic 3D talking faces conditioned only on audio input. In addition, the present disclosure provides associated methods to insert generated faces into existing videos or virtual environments. We decompose faces from video into a normalized space that decouples 3D geometry, head pose, and texture. This allows separating the prediction problem into regressions over the 3D face shape and the corresponding 2D texture atlas. To stabilize temporal dynamics, we propose an auto-regressive approach that conditions the model on its previous visual state. We also capture face illumination in our model using audio-independent 3D texture normalization.

    Classifying facial expressions using eye-tracking cameras

    公开(公告)号:US11042729B2

    公开(公告)日:2021-06-22

    申请号:US15831823

    申请日:2017-12-05

    Applicant: Google LLC

    Abstract: Images of a plurality of users are captured concurrently with the plurality of users evincing a plurality of expressions. The images are captured using one or more eye tracking sensors implemented in one or more head mounted devices (HMDs) worn by the plurality of first users. A machine learnt algorithm is trained to infer labels indicative of expressions of the users in the images. A live image of a user is captured using an eye tracking sensor implemented in an HMD worn by the user. A label of an expression evinced by the user in the live image is inferred using the machine learnt algorithm that has been trained to predict labels indicative of expressions. The images of the users and the live image can be personalized by combining the images with personalization images of the users evincing a subset of the expressions.

    CLASSIFYING FACIAL EXPRESSIONS USING EYE-TRACKING CAMERAS

    公开(公告)号:US20180314881A1

    公开(公告)日:2018-11-01

    申请号:US15831823

    申请日:2017-12-05

    Applicant: Google LLC

    Abstract: Images of a plurality of users are captured concurrently with the plurality of users evincing a plurality of expressions. The images are captured using one or more eye tracking sensors implemented in one or more head mounted devices (HMDs) worn by the plurality of first users. A machine learnt algorithm is trained to infer labels indicative of expressions of the users in the images. A live image of a user is captured using an eye tracking sensor implemented in an HMD worn by the user. A label of an expression evinced by the user in the live image is inferred using the machine learnt algorithm that has been trained to predict labels indicative of expressions. The images of the users and the live image can be personalized by combining the images with personalization images of the users evincing a subset of the expressions.

    Photorealistic Talking Faces from Audio
    6.
    发明公开

    公开(公告)号:US20240320892A1

    公开(公告)日:2024-09-26

    申请号:US18734327

    申请日:2024-06-05

    Applicant: Google LLC

    CPC classification number: G06T13/205 G06T13/40 G06T17/20

    Abstract: Provided is a framework for generating photorealistic 3D talking faces conditioned only on audio input. In addition, the present disclosure provides associated methods to insert generated faces into existing videos or virtual environments. We decompose faces from video into a normalized space that decouples 3D geometry, head pose, and texture. This allows separating the prediction problem into regressions over the 3D face shape and the corresponding 2D texture atlas. To stabilize temporal dynamics, we propose an auto-regressive approach that conditions the model on its previous visual state. We also capture face illumination in our model using audio-independent 3D texture normalization.

Patent Agency Ranking