Abstract:
A speech signal recognition method, apparatus, and system. The speech signal recognition method may include obtaining by or from a terminal an output of a personalization layer, with respect to a speech signal provided by a user of the terminal, having been implemented by input of the speech signal to the personalization layer, the personalization layer being previously trained based on speech features of the user, implementing a global model by providing the obtained output of the personalization layer to the global model, the global model being configured to output a phonemic signal indicating a phoneme included in the speech signal through the global model being previously trained based on speech features common to a plurality of users, and re-training the personalization layer based on the phonemic signal output from the global model, where the personalization layer and the global model collectively represent an acoustic model.
Abstract:
An electronic apparatus is provided. The electronic apparatus includes a storage storing error-related information of an external electronic apparatus, and a processor configured to obtain first error-related information with respect to a target time interval and second error-related information with respect to a standard time interval including the target time interval and time intervals other than the target time interval, from the storage, obtain frequency information for each number of error occurrences with respect to the target time interval based on the first error-related information and frequency information for each number of error occurrences with respect to the standard time interval based on the second error-related information, and compare the frequency information for each number of error occurrences with respect to the target time interval with the frequency information for each number of error occurrences with respect to the standard time interval to identify an error occurrence level with respect to the target time interval.
Abstract:
A processor-implemented machine translation method includes translating a source sentence expressed in a source language into a target language to determine a target sentence, determining, based on a reliability of the target sentence, whether the target sentence is appropriate as a translation result of the source sentence, and re-determining a target sentence corresponding to the source sentence in response to a result of the determining being that the target sentence is inappropriate as the translation result.
Abstract:
A language model training method and an apparatus using the language model training method are disclosed. The language model training method includes assigning a context vector to a target translation vector, obtaining feature vectors based on the target translation vector and the context vector, generating a representative vector representing the target translation vector using an attention mechanism for the feature vectors, and training a language model based on the target translation vector, the context vector, and the representative vector.
Abstract:
A method performed by a speech recognizing apparatus to recognize speech includes: obtaining a distance from the speech recognizing apparatus to a user generating a speech signal; determining a normalization value for the speech signal based on the distance; normalizing a feature vector extracted from the speech signal based on the normalization value; and performing speech recognition based on the normalized feature vector.
Abstract:
A translation method and apparatus may respectively perform or include: using one or more processors, plural different translation processes, in parallel, for a source sentence in a first language, including encoding, to generate respective feature vectors, the source sentence in each of two or more translation processes of the plural translation processes or the source sentence and a variation of the source sentence in respective translation processes of the plural translation processes, and decoding each of the respective feature vectors to generate respective plural candidate sentences in a second language; and selecting a final sentence in the second language from the respective plural candidate sentences in the second language.
Abstract:
Provided is an automated interpretation method, apparatus, and system. The automated interpretation method includes encoding a voice signal in a first language to generate a first feature vector, decoding the first feature vector to generate a first language sentence in the first language, encoding the first language sentence to generate a second feature vector with respect to a second language, decoding the second feature vector to generate a second language sentence in the second language, controlling a generating of a candidate sentence list based on any one or any combination of the first feature vector, the first language sentence, the second feature vector, and the second language sentence, and selecting, from the candidate sentence list, a final second language sentence as a translation of the voice signal.
Abstract:
A method and apparatus for training a language model, include generating a first training feature vector sequence and a second training feature vector sequence from training data. The method is configured to perform forward estimation of a neural network based on the first training feature vector sequence, and perform backward estimation of the neural network based on the second training feature vector sequence. The method is further configured to train a language model based on a result of the forward estimation and a result of the backward estimation.
Abstract:
A real-time processor-implemented translation method and apparatus is provided. The real-time translation method includes receiving a content, determining a delay time for real-time translation based on a silence interval of the received content and an utterance interval of the received content, generating a translation result by translating a language used in the received content, and synthesizing the translation result and the received content.
Abstract:
A real-time processor-implemented translation method and apparatus is provided. The real-time translation method includes receiving a content, determining a delay time for real-time translation based on a silence interval of the received content and an utterance interval of the received content, generating a translation result by translating a language used in the received content, and synthesizing the translation result and the received content.