Abstract:
A speech-enabled dialog system responds to a plurality of wake-up phrases. Based on which wake-up phrase is detected, the system's configuration is modified accordingly. Various configurable aspects of the system include selection and morphing of a text-to-speech voice; configuration of acoustic model, language model, vocabulary, and grammar; configuration of a graphic animation; configuration of virtual assistant personality parameters; invocation of a particular user profile; invocation of an authentication function; and configuration of an open sound. Configuration depends on a target market segment. Configuration also depends on the state of the dialog system, such as whether a previous utterance was an information query.
Abstract:
A system and method is presented for performing dual mode speech recognition, employing a local recognition module on a mobile device and a remote recognition engine on a server device. The system accepts a spoken query from a user, and both the local recognition module and the remote recognition engine perform speech recognition operations on the query, returning a transcription and confidence score, subject to a latency cutoff time. If both sources successfully transcribe the query, then the system accepts the result having the higher confidence score. If only one source succeeds, then that result is accepted. In either case, if the remote recognition engine does succeed in transcribing the query, then a client vocabulary is updated if the remote system result includes information not present in the client vocabulary.
Abstract:
The present invention extends to methods, systems, and computer program products for a natural language module store. In general, the invention can be used to manage natural language modules offered through a natural language module store. Natural language module (NLM) developers can post NLMs at a NLM store to make the NLMs available for use by others. Developers can select NLMs for inclusion in natural language interpreters (NLIs) containing (and possibly integrating the functionality of) one or more NLMs. Prior to selecting a NLM, a developer can search or browse NLMs to identify an appropriate NLM. Optionally, a developer can test a NLM in the NLM store prior to inclusion in an NLI. For example, multiple NLMs purporting to provide the same specified natural language functionality can be tested relative to one another prior to selection of one of the NLMs for inclusion in an NLI.
Abstract:
A method for matching a query against a broadcast stream includes receiving one or more broadcast streams, from which it generates and stores an audio fingerprint of a selected portion of each received broadcast stream. A query is received from which the method generates an audio fingerprint. From that point, the method continues by identifying audio content from the query, using the query audio fingerprint and a database of indexed audio content. The method concludes by identifying the source of the query using the query audio fingerprint and the stored audio fingerprints. Embodiments of the method further include predictively caching audio fingerprint sequences and corresponding audio item identifiers from a server after storing audio fingerprints extracted from the broadcast stream; and using the predictively cached audio fingerprint sequences to identify an audio item within the audio signal based on at least some additional audio fingerprints of the audio signal.
Abstract:
A system and method is presented for performing dual mode speech recognition, employing a local recognition module on a mobile device and a remote recognition engine on a server device. The system accepts a spoken query from a user, and both the local recognition module and the remote recognition engine perform speech recognition operations on the query, returning a transcription and confidence score, subject to a latency cutoff time. If both sources successfully transcribe the query, then the system accepts the result having the higher confidence score. If only one source succeeds, then that result is accepted. In either case, if the remote recognition engine does succeed in transcribing the query, then a client vocabulary is updated if the remote system result includes information not present in the client vocabulary.
Abstract:
A method for processing a voice message in a computerized system. The method receives and records a speech utterance including a message portion and a communication portion. The method proceeds to parse the input to identify and separate the message portion and the communication portion. It then identifies communication parameters, including one or more destination mailboxes, from the communication portion, and it transmits the message portion to the destination mailbox as a voice message.
Abstract:
A system and method is presented for performing dual mode speech recognition, employing a local recognition module on a mobile device and a remote recognition engine on a server device. The system accepts a spoken query from a user, and both the local recognition module and the remote recognition engine perform speech recognition operations on the query, returning a transcription and confidence score, subject to a latency cutoff time. If both sources successfully transcribe the query, then the system accepts the result having the higher confidence score. If only one source succeeds, then that result is accepted. In either case, if the remote recognition engine does succeed in transcribing the query, then a client vocabulary is updated if the remote system result includes information not present in the client vocabulary.
Abstract:
A method of a local recognition system controlling a host device to perform one or more operations is provided. The method includes receiving, by the local recognition system, a query, performing speech recognition on the received query by implementing, by the local recognition system, a local language context comprising a set of words comprising descriptions in terms of components smaller than the words, and performing speech recognition, using the local language context, to create a transcribed query. Further, the method includes controlling the host device in dependence upon the speech recognition performed on the transcribed query.
Abstract:
Systems and methods for searching databases by sound data input are provided herein. A service provider may have a need to make their database(s) searchable through search technology. However, the service provider may not have the resources to implement such search technology. The search technology may allow for search queries using sound data input. The technology described herein provides a solution addressing the service provider’s need, by giving a search technology that furnishes search results in a fast, accurate manner. In further embodiments, systems and methods to monetize those search results are also described herein.
Abstract:
A method and system for acoustic model conditioning on non-phoneme information features for optimized automatic speech recognition is provided. The method includes using an encoder model to encode sound embedding from a known key phrase of speech and conditioning an acoustic model with the sound embedding to optimize its performance in inferring the probabilities of phonemes in the speech. The sound embedding can comprise non-phoneme information related to the key phrase and the following utterance. Further, the encoder model and the acoustic model can be neural networks that are jointly trained with audio data.