-
公开(公告)号:US20240078374A1
公开(公告)日:2024-03-07
申请号:US17957489
申请日:2022-09-30
申请人: GOOGLE LLC
发明人: Ajay Gokhale , Jiawei Chen , Alvin Abdagic , Adrien Olczak , Alessandro Agostini , Alexander Robertson , Cristian Udrescu , Jackie Xiang , Jennifer Daniel , Keqian Yan , Mehek Sharma , Nicolo D'Ercole , Yang Lu , Dror Ayalon
IPC分类号: G06F40/166 , G06F3/0482 , G06F3/0488 , G06F40/279 , G10L15/197 , G10L15/22 , G10L25/63
CPC分类号: G06F40/166 , G06F3/0482 , G06F3/0488 , G06F40/279 , G10L15/197 , G10L15/22 , G10L25/63 , G10L2015/223
摘要: Implementations described herein relate to causing emoji(s) that are associated with a given emotion class expressed by a spoken utterance to be visually rendered for presentation to a user at a display of a client device of the user. Processor(s) of the client device may receive audio data that captures the spoken utterance, process the audio data to generate textual data that is predicted to correspond to the spoken utterance, and cause a transcription of the textual data to be visually rendered for presentation to the user via the display. Further, the processor(s) may determine, based on processing the textual data, whether the spoken utterance expresses a given emotion class. In response to determining that the spoken utterance expresses the given emotion class, the processor(s) may cause emoji(s) that are stored in association with the given emotion class to be visually rendered for presentation to the user via the display.
-
2.
公开(公告)号:US20240029728A1
公开(公告)日:2024-01-25
申请号:US17902560
申请日:2022-09-02
申请人: GOOGLE LLC
发明人: Nicolo D'Ercole , Shumin Zhai , Swante Scholz , Mehek Sharma , Adrien Olczak , Akshay Kannan , Alvin Abdagic , Julia Proskurnia , Viesturs Zarins
IPC分类号: G10L15/22 , G10L15/08 , G06F16/683
CPC分类号: G10L15/22 , G10L15/08 , G06F16/685
摘要: Implementations described herein generally relate to generating a modification selectable element that may be provided for presentation to a user in a smart dictation session with an automated assistant. The modification selectable element may, when selected, cause a transcription, that includes textual data generated based on processing audio data that captures a spoken utterance and that is automatically arranged, to be modified. The transcription may be automatically arranged to include spacing, punctuation, capitalization, indentations, paragraph breaks, and/or other arrangement operations that are not specified by the user in providing the spoken utterance. Accordingly, a subsequent selection of the modification selectable element may cause these automatic arrangement operation(s), and/or the textual data locationally proximate to these automatic arrangement operation(s), to be modified. Implementations described herein also relate to generating the transcription and/or the modification selectable element on behalf of a third-party software application.
-