Abstract:
Features are disclosed for active learning to identify the words which are likely to improve the guessing and automatic speech recognition (ASR) after manual annotation. When a speech recognition system needs pronunciations for words, a lexicon is typically used. For unknown words, pronunciation-guessing (G2P) may be included to provide pronunciations in an unattended (e.g., automatic) fashion. However, having manually (e.g., by a human) annotated pronunciations provides better ASR than having automatic pronunciations that may, in some instances, be wrong. The included active learning features help to direct these limited annotation resources.
Abstract:
Features are disclosed for generating acoustic models from an existing corpus of data. Methods for generating the acoustic models can include receiving at least one characteristic of a desired acoustic model, selecting training utterances corresponding to the characteristic from a corpus comprising audio data and corresponding transcription data, and generating an acoustic model based on the selected training utterances.
Abstract:
A speech recognition system utilizing automatic speech recognition techniques such as end-pointing techniques in conjunction with beamforming and/or signal processing to isolate speech from one or more speaking users from multiple received audio signals and to detect the beginning and/or end of the speech based at least in part on the isolation. Audio capture devices such as microphones may be arranged in a beamforming array to receive the multiple audio signals. Multiple audio sources including speech may be identified in different beams and processed.
Abstract:
In a voice controlled system, multiple applications are configured to respond to various commands. The voice controlled system includes client devices and servers. The correct application to receive a natural language command is identified based on how well the command matches functions of the application. A target application to receive the command may additionally be selected based on which application is most likely to receive a command. Likelihood of an application receiving a command may be determined by considering context. The command may be a voice input to a client device that is analyzed by speech recognition technology to determine word strings representing possible commands. Thus, the selection of a target application to receive the command may be based on word strings from the natural language input, a closeness of fit between the command and an application, and/or the likelihood an application is the target for the next incoming command.
Abstract:
A speech recognition system utilizing automatic speech recognition techniques such as end-pointing techniques in conjunction with beamforming and/or signal processing to isolate speech from one or more speaking users from multiple received audio signals and to detect the beginning and/or end of the speech based at least in part on the isolation. Audio capture devices such as microphones may be arranged in a beamforming array to receive the multiple audio signals. Multiple audio sources including speech may be identified in different beams and processed.
Abstract:
Power consumption for a computing device may be managed by one or more keywords. For example, if an audio input obtained by the computing device includes a keyword, a network interface module and/or an application processing module of the computing device may be activated. The audio input may then be transmitted via the network interface module to a remote computing device, such as a speech recognition server. Alternately, the computing device may be provided with a speech recognition engine configured to process the audio input for on-device speech recognition.
Abstract:
Approaches are described for detecting when an electronic device (such as a mobile phone) has been stolen or is otherwise being used by someone other than an authorized user of the device. At least one sensor of the device can obtain data during a current use of the device, and the device can determine from the data a set of available features. The features can be compared to a corresponding model associated with an owner (or other authorized user) of the device to generate a confidence value indicative of whether the current user operating the device is likely the owner of the device. The confidence value can be compared to at least one confidence threshold, for example, and based on the comparison, the current user can be provided access to at least a portion of functionality of the device and/or a security action can be performed when the confidence value does not at least meet at least one confidence threshold.
Abstract:
A system and method for using speech recognition, natural language understanding, image processing, and facial recognition to automatically analyze the audio and video data of video content and generate enhanced data relating to the video content and characterize the aspects or events of the video content. The results of the analysis and characterization of the aspects of the video content may be used to annotate and enhance the video content to enhance a user's viewing experience by allowing the user to interact with the video content and presenting the user with information related to the video content.
Abstract:
Features are disclosed for generating and using personalized named entity recognition models. A personalized model can be trained for a particular user, and then interpolated with a general model for use in named entity recognition. In some embodiments, a model may be trained for a group of users, where the users share some similarity relevant to language processing. In some embodiments, various base models may be trained so as to provide better accuracy for certain types of language input than a general model. Users may be associated with any number of base models, and the associated based models may then be interpolated for use in named entity recognition on input from the corresponding user.
Abstract:
A method and apparatus for identifying a document in a set of stored documents based on a pattern of characteristics in the document is presented. A digital image including at least a portion of the a document is acquired. A pattern of characteristics is then identified in the digital image. The pattern is matched to the set of stored documents to identify the document in the digital image from the set of stored documents.