MULTI-MODAL INTERFACE IN A VOICE-ACTIVATED NETWORK

    公开(公告)号:US20200342856A1

    公开(公告)日:2020-10-29

    申请号:US16923416

    申请日:2020-07-08

    Applicant: Google LLC

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

    Multi-modal interface in a voice-activated network

    公开(公告)号:US10733984B2

    公开(公告)日:2020-08-04

    申请号:US15973447

    申请日:2018-05-07

    Applicant: Google LLC

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

    Multi-modal interface in a voice-activated network

    公开(公告)号:US11776536B2

    公开(公告)日:2023-10-03

    申请号:US16923416

    申请日:2020-07-08

    Applicant: Google LLC

    CPC classification number: G10L15/1822 G10L15/1815

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

    MULTI-MODAL INTERFACE IN A VOICE-ACTIVATED NETWORK

    公开(公告)号:US20190341028A1

    公开(公告)日:2019-11-07

    申请号:US15973447

    申请日:2018-05-07

    Applicant: Google LLC

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

    Multi-modal interface in a voice-activated network

    公开(公告)号:US12106750B2

    公开(公告)日:2024-10-01

    申请号:US18360367

    申请日:2023-07-27

    Applicant: Google LLC

    CPC classification number: G10L15/1822 G10L15/1815

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

    MULTI-MODAL INTERFACE IN A VOICE-ACTIVATED NETWORK

    公开(公告)号:US20240062749A1

    公开(公告)日:2024-02-22

    申请号:US18360367

    申请日:2023-07-27

    Applicant: Google LLC

    CPC classification number: G10L15/1822 G10L15/1815

    Abstract: Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.

Patent Agency Ranking