Abstract:
Various systems and methods for implementing skin texture-based authentication are described herein. A system comprises a capture module to obtain at a wearable device worn by a user, an input representation of the user's skin; an analysis module to identify a set of features in the input representation; and an authentication module to authenticate the user based on the set of features.
Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.
Abstract:
Systems, apparatus and methods may provide for audio processing of received user audio input from a microphone that may optionally be a tissue conducting microphone. Audio processing may be further conducted on received ambient audio from one or more additional microphones. A translator may translate the ambient audio into content to be output to a user. In an embodiment, ambient audio is translated into visual content to be displayed on a virtual reality device.
Abstract:
Technologies for automatic speech recognition using articulatory parameters are disclosed. An automatic speech recognition device may capture speech data from a speaker and also capture an image of the speaker. The automatic speech recognition device may determine one or more articulatory parameters based on the image, such as such as a jaw angle, a lip protrusion, or a lip height, and compare those parameters with articulatory parameters of training users. After selecting training users with similar articulatory parameters as the training speaker, the automatic speech recognition device may select training data associated with the selected training speakers, including parameters to use for an automatic speech recognition algorithm. By using the parameters already optimized for training users with similar articulatory parameters as the speaker, the automatic speech recognition device may quickly adapt an automatic speech recognition algorithm to the speaker.
Abstract:
A method for establishing an articulatory speech synthesis model of a person's voice includes acquiring image data representing a visage of a person, in which the visage includes facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice; selecting a predefined articulatory speech synthesis model from among stores of predefined models, the selection based at least in part on one or both of the facial characteristics or the exteriorly visible articulatory speech synthesis model parameters; and associating at least a portion of the selected predefined articulatory speech synthesis model with the articulatory speech synthesis model of the person's voice.
Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.
Abstract:
Technologies for automatic speech recognition using articulatory parameters are disclosed. An automatic speech recognition device may capture speech data from a speaker and also capture an image of the speaker. The automatic speech recognition device may determine one or more articulatory parameters based on the image, such as such as a jaw angle, a lip protrusion, or a lip height, and compare those parameters with articulatory parameters of training users. After selecting training users with similar articulatory parameters as the training speaker, the automatic speech recognition device may select training data associated with the selected training speakers, including parameters to use for an automatic speech recognition algorithm. By using the parameters already optimized for training users with similar articulatory parameters as the speaker, the automatic speech recognition device may quickly adapt an automatic speech recognition algorithm to the speaker.
Abstract:
A method, performed by a user equipment device, for text-to-speech conversion entails sending to an articulatory model server exterior facial structural information of a person, receiving from the articulatory model server at least a portion of a predefined articulatory model that corresponds to the exterior facial structural information, the predefined articulatory model representing a voice of a modeled person who is different from the person, and generating, based at least partly on the predefined articulatory model, speech from text stored in a memory of the user equipment device. Furthermore, a method of configuring text-to-speech conversion for a user equipment device entails determining at least a portion of an articulatory model that corresponds to exterior facial structural information based on a comparison of the exterior facial structural information to exterior facial structural information stored in a database of articulatory models.
Abstract:
Systems, apparatus and methods may provide for audio processing of received user audio input from a microphone that may optionally be a tissue conducting microphone. Audio processing may be further conducted on received ambient audio from one or more additional microphones. A translator may translate the ambient audio into content to be output to a user. In an embodiment, ambient audio is translated into visual content to be displayed on a virtual reality device.
Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.