EMOTION CLASSIFICATION BASED ON EXPRESSION VARIATIONS ASSOCIATED WITH SAME OR SIMILAR EMOTIONS

    公开(公告)号:US20190122071A1

    公开(公告)日:2019-04-25

    申请号:US15791821

    申请日:2017-10-24

    Abstract: Techniques are described that facilitate automatically distinguishing between different expressions of a same or similar emotion. In one embodiment, a computer-implemented is provided that comprises partitioning, by a device operatively coupled to a processor, a data set comprising facial expression data into different clusters of the facial expression data based on one or more distinguishing features respectively associated with the different clusters, wherein the facial expression data reflects facial expressions respectively expressed by people. The computer-implemented method can further comprise performing, by the device, a multi-task learning process to determine a final number of the different clusters for the data set using a multi-task learning process that is dependent on an output of an emotion classification model that classifies emotion types respectively associated with the facial expressions.

    SPECIALIST KEYWORDS RECOMMENDATIONS IN SEMANTIC SPACE

    公开(公告)号:US20180137137A1

    公开(公告)日:2018-05-17

    申请号:US15352842

    申请日:2016-11-16

    Abstract: Techniques are provided for generating recommended query terms that are specialized to a topic of desired information based on a query associated with a user. In one example, a computer-implemented method comprising selecting, by a system operatively coupled to a processor, a coarse cluster of corpus terms having a defined relatedness to a query associated with a user from a plurality of coarse clusters of corpus terms; and determining, by the system, a plurality of candidate terms from search results associated with the query. The computer-implemented method can also comprise determining, by the system, at least one recommended query term based on refined clusters of the coarse cluster, the candidate terms, and the query; and displaying, by the system, the at least one recommended query term on a display device associated with the query.

    Correcting text with voice processing
    5.
    发明授权
    Correcting text with voice processing 有权
    使用语音处理校正文本

    公开(公告)号:US09484031B2

    公开(公告)日:2016-11-01

    申请号:US14027664

    申请日:2013-09-16

    Abstract: The present invention relates to voice processing and provides a method and system for correcting a text. The method comprising: determining a target text unit to be corrected in a text; receiving a reference voice segment input by the user for the target text unit; determining a reference text unit whose pronunciation is similar to a word in the target text unit based on the reference voice segment; and correcting the word in the target text unit in the text by the reference text unit. The present invention enables the user to easily correct errors in the text vocally.

    Abstract translation: 本发明涉及语音处理,并且提供了一种用于校正文本的方法和系统。 该方法包括:确定要在文本中校正的目标文本单元; 接收用户为目标文本单元输入的参考语音段; 基于所述参考语音段确定其发音类似于所述目标文本单元中的单词的参考文本单元; 并且通过参考文本单元来校正文本中的目标文本单元中的单词。 本发明使用户能够容易地纠正文本中的错误错误。

    Emotion classification based on expression variations associated with same or similar emotions

    公开(公告)号:US10963756B2

    公开(公告)日:2021-03-30

    申请号:US16587701

    申请日:2019-09-30

    Abstract: Techniques are described that facilitate automatically distinguishing between different expressions of a same or similar emotion. In one embodiment, a computer-implemented is provided that comprises partitioning, by a device operatively coupled to a processor, a data set comprising facial expression data into different clusters of the facial expression data based on one or more distinguishing features respectively associated with the different clusters, wherein the facial expression data reflects facial expressions respectively expressed by people. The computer-implemented method can further comprise performing, by the device, a multi-task learning process to determine a final number of the different clusters for the data set using a multi-task learning process that is dependent on an output of an emotion classification model that classifies emotion types respectively associated with the facial expressions.

    Text-to-articulatory movement
    7.
    发明授权

    公开(公告)号:US10521945B2

    公开(公告)日:2019-12-31

    申请号:US15389502

    申请日:2016-12-23

    Abstract: According to an embodiment of the present invention, a computer-implemented method for modeling text-to-articulatory movement conversion is provided. In the method, text features are generated for a plurality of texts, wherein the text feature comprises a phoneme. Articulatory features are determined for the plurality of texts from a plurality of audio signals and a plurality of visual signals, wherein the audio signals record speeches of the plurality of texts, the visual signals record three-dimensional images of an articulator when speaking the plurality of texts, and the articulatory features indicate articulatory positions of the articulator in the speeches. A text-to-articulatory movement model is established with the text features as inputs and the articulatory features as outputs.

    VISUAL LIVENESS DETECTION
    10.
    发明申请
    VISUAL LIVENESS DETECTION 有权
    视觉生活检测

    公开(公告)号:US20170039440A1

    公开(公告)日:2017-02-09

    申请号:US14821258

    申请日:2015-08-07

    Abstract: In an approach for visual liveness detection, a video-audio signal related to a speaker speaking a text is obtained. The video-audio signal is split into a video signal which records images of the speaker and an audio signal which records a speech spoken by the speaker. Then a first sequence indicating visual mouth openness is obtained from the video signal, and a second sequence indicating acoustic mouth openness is obtained based on the text and the audio signal. Synchrony between the first and second sequences is measured, and the liveness of the speaker is determined based on the synchrony.

    Abstract translation: 在视觉活动检测的方法中,获得与说话者说话的扬声器相关的视频 - 音频信号。 视频 - 音频信号被分割成记录扬声器的图像的视频信号和记录扬声器所说出的语音的音频信号。 然后,从视频信号获得指示视觉开放性的第一序列,并且基于文本和音频信号获得指示声音开口性的第二序列。 测量第一和第二序列之间的同步,并且基于同步来确定说话者的活力。

Patent Agency Ranking