Abstract:
A recognition method includes extracting target data corresponding to a current window and padding data subsequent to the target data from sequence data; acquiring a state parameter corresponding to a previous window; and calculating a recognition result for the current window based on the state parameter, the extracted target data, and the extracted padding data using a recurrent model.
Abstract:
A speech recognition apparatus includes a probability calculator configured to calculate phoneme probabilities of an audio signal using an acoustic model; a candidate set extractor configured to extract a candidate set from a recognition target list; and a result returner configured to return a recognition result of the audio signal based on the calculated phoneme probabilities and the extracted candidate set.
Abstract:
A terminal and a method for sharing content are provided. A terminal includes an image acquirer configured to acquire face image data from a camera while content is being displayed, a face recognizer configured to recognize a face included in the face image data, a face change detector configured to detect whether the recognized face is different from a face recognized in a previous image data, a contact searcher configured to, in response to the detection that the recognized face is different from the face recognized in the previous image data, search a contact corresponding to the recognized face, and an information transmitter configured to transmit content usage information to the searched contact.
Abstract:
An apparatus and method for executing an application may execute a text string selected by a user or an application associated with a type of text string to input the text string to increase the user's convenience. The apparatus of executing an application includes a text string recognizer to determine a text string, a determiner to determine one or more candidate applications related to the text string, and an input location of the selected text string based on a type of the selected text string and the association model, an application list provider to generate and display a list of the candidate applications, and an application executer to execute a candidate application selected from the list and to input the selected text string into the input location of the candidate application.
Abstract:
Provided is a processor-implemented method of generating a natural language, the method including generating a latent variable from an embedding vector that corresponds to an input utterance, determining attention information related to the input utterance by applying the generated latent variable to a neural network model, and outputting a natural language response that corresponds to the input utterance based on the calculated attention information.
Abstract:
A recognition method performed in a user terminal includes determining a characteristic parameter personalized to a speech of a user based on a reference speech signal input by the user; receiving, as an input, a target speech signal to be recognized from the user; and outputting a recognition result of the target speech signal, wherein the recognition result of the target speech signal is determined based on the characteristic parameter and a model for recognizing the target speech signal.
Abstract:
An automated interpretation method includes: interpreting a source voice signal expressed in a first language by dividing the source voice signal into at least one word as a unit while the source voice signal is being input, and outputting, as an interpretation result in real time, a first target voice signal expressed in a second language by each unit; determining whether to re-output the interpretation result; and in response to a determination of the determining of whether to re-output the interpretation being a determination that the interpretation result is to be re-output, interpreting the source voice signal by a sentence as a unit and outputting, as the interpretation result, a second target voice signal expressed in the second language.
Abstract:
An apparatus for determining a translation word includes a word vector generator configured to generate a word vector corresponding to an input word of a first language with reference to a first word vector space that is related to the first language, a word vector determiner configured to determine a word vector of a second language, wherein the determined word vector of the second language corresponds to the generated word vector, using a matching model, and a translation word selector configured to select a translation word of the second language, wherein the selected translation word corresponds to the input word of the first language, based on the determined word vector of the second language.
Abstract:
A video display method of a user terminal, includes determining whether ambient noise measured when a video is played is in an allowable range, and generating subtitles based on a voice signal included in the video in response to the ambient noise being determined to be out of the allowable range. The method further includes displaying the generated subtitles with the video.
Abstract:
An out-of-service (OOS) sentence generating method includes: training models based on a target utterance template of a target service and a target sentence generated from the target utterance template; generating a similar utterance template that is similar to the target utterance template based on a trained model, among the trained models, and a sentence generated from an utterance template of another service; and generating a similar sentence that is similar to the target sentence based on another trained model, among the trained models, and the similar utterance template.