METHOD AND APPARATUS FOR PERFORMING MULTI-LANGUAGE COMMUNICATION

    公开(公告)号:US20200043495A1

    公开(公告)日:2020-02-06

    申请号:US16601787

    申请日:2019-10-15

    Abstract: A method for performing multi-language communication includes receiving an utterance, identifying a language of the received utterance, determining whether the identified language matches a preset reference language, applying, to the received utterance, an interpretation model interpreting the identified language into the reference language when the identified language does not match the reference language, changing, to text, speech data which is outputted in the reference language as a result of applying the interpretation model, generating a response message responding to the text of the speech data, and outputting the response message. Here, the interpretation model may be a deep neural network model generated through machine learning, and the interpretation model may be stored in an edge device or provided through a server in an Internet of things environment through a 5G network.

    EMOTION CLASSIFICATION INFORMATION-BASED TEXT-TO-SPEECH (TTS) METHOD AND APPARATUS

    公开(公告)号:US20210366462A1

    公开(公告)日:2021-11-25

    申请号:US16485421

    申请日:2019-01-11

    Abstract: Disclosed are an emotion classification information-based text-to-speech (TTS) method and device. The emotion classification information-based TTS method according to an embodiment of the present invention may, when emotion classification information is set in a received message, transmit metadata corresponding to the set emotion classification information to a speech synthesis engine and, when no emotion classification information is set in the received message, generate new emotion classification information through semantic analysis and context analysis of sentences in the received message and transmit the metadata to the speech synthesis engine. The speech synthesis engine may perform speech synthesis by carrying emotion classification information based on the transmitted metadata.

    METHOD AND DEVICE FOR SPEECH PROCESSING

    公开(公告)号:US20210082421A1

    公开(公告)日:2021-03-18

    申请号:US16676160

    申请日:2019-11-06

    Abstract: Disclosed are a speech processing method and a speech processing apparatus, characterized in that a speech processing is carried out by executing an artificial intelligence (AI) algorithm and/or a machine learning algorithm, such that the speech processing apparatus, a user terminal, and a server can communicate with each other in a 5G communication environment. The speech processing method according to one exemplary embodiment of the present invention includes converting a response text, which is generated in response to a spoken utterance of a user, to a spoken response utterance, obtaining external situation information while outputting the spoken response utterance, generating a dynamic spoken response utterance by converting the spoken response utterance on the basis of the external situation information, and outputting the dynamic spoken response utterance.

    METHOD FOR SYNTHESIZED SPEECH GENERATION USING EMOTION INFORMATION CORRECTION AND APPARATUS

    公开(公告)号:US20210074261A1

    公开(公告)日:2021-03-11

    申请号:US16928815

    申请日:2020-07-14

    Abstract: A method includes generating first synthesized speech by using text and a first emotion vector configured for the text, extracting a second emotion vector included in the first synthesized speech, determining whether correction of the second emotion information vector is needed by comparing a loss value calculated by using the first emotion information vector and the second emotion information vector with a preconfigured threshold, re-performing speech synthesis by using a third emotion information vector generated by correcting the second emotion information vector, and outputting the generated synthesized speech, thereby configuring emotion information of speech in a more effective manner. A speech synthesis apparatus may be associated with an artificial intelligence module, drone (unmanned aerial vehicle, UAV), robot, augmented reality (AR) devices, virtual reality (VR) devices, devices related to 5G services, and the like.

    APPARATUS AND METHOD FOR INSPECTING SPEECH RECOGNITION

    公开(公告)号:US20200013388A1

    公开(公告)日:2020-01-09

    申请号:US16572955

    申请日:2019-09-17

    Abstract: Disclosed are a speech recognition verification device and a speech recognition verification method, which verify speech recognition results by executing artificial intelligence (AI) algorithms and/or machine learning algorithms in a 5G environment connected for Internet-of-Things. According to an embodiment, the speech recognition verification method includes converting a verification target text item to a verification target spoken utterance by applying a preset utterance condition, analyzing the verification target spoken utterance and outputting a recognition result text item corresponding to an analysis result, and verifying speech recognition performance through comparison between the verification target text item and the recognition result text item. According to the present disclosure, the speech recognition result may be verified objectively by using a spoken utterance generated with random text and various utterance conditions as input of speech recognition.

    SPEECH SYNTHESIS DEVICE AND SPEECH SYNTHESIS METHOD

    公开(公告)号:US20230148275A1

    公开(公告)日:2023-05-11

    申请号:US17959050

    申请日:2022-10-03

    CPC classification number: G10L13/047 G10L25/30

    Abstract: Provided is a speech synthetic device capable of outputting a synthetic voice having various speech styles. The speech synthesis device includes a speaker, and a processor to acquire voice feature information through a text and a user input; generate a synthetic voice, by receiving the text and the voice feature information inputs into a decoder supervised-trained to minimize a difference between feature information of a learning text and characteristic information of a learning voice, and output the generated synthetic voice through the speaker.

Patent Agency Ranking