Enrollment with an automated assistant

    公开(公告)号:US11238142B2

    公开(公告)日:2022-02-01

    申请号:US16403532

    申请日:2019-05-04

    Applicant: Google LLC

    Abstract: Techniques are described herein for dialog-based enrollment of individual users for single- and/or multi-modal recognition by an automated assistant, as well as determining how to respond to a particular user's request based on the particular user being enrolled and/or recognized. Rather than requiring operation of a graphical user interface for individual enrollment, dialog-based enrollment enables users to enroll themselves (or others) by way of a human-to-computer dialog with the automated assistant.

    Invoking automated assistant function(s) based on detected gesture and gaze

    公开(公告)号:US11237630B2

    公开(公告)日:2022-02-01

    申请号:US17110716

    申请日:2020-12-03

    Applicant: Google LLC

    Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.

    SYSTEMS AND METHODS FOR ADJUSTING LIGHT EMITTED FROM A DISPLAY

    公开(公告)号:US20210327393A1

    公开(公告)日:2021-10-21

    申请号:US17282101

    申请日:2018-10-08

    Applicant: Google LLC

    Abstract: Systems and methods for adjusting light emitted from a display of a device are provided. The adjusting includes obtaining, from light of an environment detected by at least one sensor, a measured color of light of the environment, and obtaining, from light of the environment detected by at least one sensor, a measured brightness of light of the environment. In response to the obtaining the measured color and the measured brightness of light, a color of light emitted from the display is adjusted from an initial color prior to the adjusting to a target color that matches the measured color. Further, a brightness of light emitted from the display is adjusted from an initial brightness emitted by the display prior to the adjusting to a target brightness that matches the measured brightness of light.

    GENERATING AND/OR ADAPTING AUTOMATED ASSISTANT CONTENT ACCORDING TO A DISTANCE BETWEEN USER(S) AND AN AUTOMATED ASSISTANT INTERFACE

    公开(公告)号:US20210117720A1

    公开(公告)日:2021-04-22

    申请号:US17135272

    申请日:2020-12-28

    Applicant: Google LLC

    Abstract: Methods, apparatus, systems, and computer-readable media are provided for generating and/or adapting automated assistant content according to a distance of a user relative to an automated assistant interface that renders the automated assistant content. For instance, the automated assistant can provide data for a client device to render. The client device can request additional data when the user relocates closer to, or further from, the client device. In some implementations, a request for additional data can identify a distance between the user and the client device. In this way, the additional data can be generated or selected according to the distance in the request. Other implementations can allow an automated assistant to determine an active user from a group of users in an environment, and determine a distance between the active user and the client device in order that any rendered content can be tailored for the active user.

    ADAPTING AUTOMATED ASSISTANT BASED ON DETECTED MOUTH MOVEMENT AND/OR GAZE

    公开(公告)号:US20200342223A1

    公开(公告)日:2020-10-29

    申请号:US16606030

    申请日:2018-05-04

    Applicant: Google LLC

    Abstract: Adapting an automated assistant based on detecting: movement of a mouth of a user; and/or that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. The detecting of the mouth movement and/or the directed gaze can be based on processing of vision data from one or more vision components associated with the assistant device, such as a camera incorporated in the assistant device. The mouth movement that is detected can be movement that is indicative of a user (to whom the mouth belongs) speaking.

    Conversation-aware proactive notifications for a voice interface device

    公开(公告)号:US10679608B2

    公开(公告)日:2020-06-09

    申请号:US15841284

    申请日:2017-12-13

    Applicant: GOOGLE LLC

    Abstract: A method for proactive notifications in a voice interface device includes: receiving a first user voice request for an action with an future performance time; assigning the first user voice request to a voice assistant service for performance; subsequent to the receiving, receiving a second user voice request and in response to the second user voice request initiating a conversation with the user; and during the conversation: receiving a notification from the voice assistant service of performance of the action; triggering a first audible announcement to the user to indicate a transition from the conversation and interrupting the conversation; triggering a second audible announcement to the user to indicate performance of the action; and triggering a third audible announcement to the user to indicate a transition back to the conversation and rejoining the conversation.

    MULTIMODAL EMBEDDINGS
    28.
    发明申请

    公开(公告)号:US20250078484A1

    公开(公告)日:2025-03-06

    申请号:US18242213

    申请日:2023-09-05

    Applicant: GOOGLE LLC

    Abstract: Implementations relate to generating and using multimodal embeddings. In various implementations, first modality data may be obtained and encoded into first modality embedding(s) using a trained first modality encoder that is stored in memory of edge-based client device(s). Second modality data may be obtained and encoded into second modality embedding(s) using a trained second modality encoder that is also stored in the memory of the edge-based client device(s). The first and second modality embeddings may be processed using an edge-based multimodal LLM that is also stored locally in memory of the edge-based client device(s) to generate a multimodal contextual embedding, which may be provided to a remote server that hosts a central LLM, e.g., in conjunction with a natural language input provided by the user. Information generated using the central LLM, responsive to the natural language input, may be received from the remote server.

Patent Agency Ranking