Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.
Abstract:
Technologies for automatic speech recognition using articulatory parameters are disclosed. An automatic speech recognition device may capture speech data from a speaker and also capture an image of the speaker. The automatic speech recognition device may determine one or more articulatory parameters based on the image, such as such as a jaw angle, a lip protrusion, or a lip height, and compare those parameters with articulatory parameters of training users. After selecting training users with similar articulatory parameters as the training speaker, the automatic speech recognition device may select training data associated with the selected training speakers, including parameters to use for an automatic speech recognition algorithm. By using the parameters already optimized for training users with similar articulatory parameters as the speaker, the automatic speech recognition device may quickly adapt an automatic speech recognition algorithm to the speaker.
Abstract:
Methods and apparatus to modify a binary file for scalable loading on distributed computing systems are disclosed. An example method to modify a binary file for scalable loading on a distributed computing system includes identifying, by executing an instruction with a processor, in the binary file a file name for a shared file, determining, by executing an instruction with the processor, whether the shared file can be loaded based on a path name specified in the binary file, and, responsive to a determination that the shared file can be loaded, generating, by executing an instruction with the processor, an identifier that can be used to access a local file scalably copied from a global file system, and modifying, by executing an instruction with the processor, the binary file by replacing, in the binary file, the file name with the generated identifier.
Abstract:
Methods and apparatus to modify a binary file for scalable loading on distributed computing systems are disclosed. An example method to modify a binary file for scalable loading on a distributed computing system includes identifying, by executing an instruction with a processor, in the binary file a file name for a shared file, determining, by executing an instruction with the processor, whether the shared file can be loaded based on a path name specified in the binary file, and, responsive to a determination that the shared file can be loaded, generating, by executing an instruction with the processor, an identifier that can be used to access a local file scalably copied from a global file system, and modifying, by executing an instruction with the processor, the binary file by replacing, in the binary file, the file name with the generated identifier.
Abstract:
A method for establishing an articulatory speech synthesis model of a person's voice includes acquiring image data representing a visage of a person, in which the visage includes facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice; selecting a predefined articulatory speech synthesis model from among stores of predefined models, the selection based at least in part on one or both of the facial characteristics or the exteriorly visible articulatory speech synthesis model parameters; and associating at least a portion of the selected predefined articulatory speech synthesis model with the articulatory speech synthesis model of the person's voice.
Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.
Abstract:
Technologies for automatic speech recognition using articulatory parameters are disclosed. An automatic speech recognition device may capture speech data from a speaker and also capture an image of the speaker. The automatic speech recognition device may determine one or more articulatory parameters based on the image, such as such as a jaw angle, a lip protrusion, or a lip height, and compare those parameters with articulatory parameters of training users. After selecting training users with similar articulatory parameters as the training speaker, the automatic speech recognition device may select training data associated with the selected training speakers, including parameters to use for an automatic speech recognition algorithm. By using the parameters already optimized for training users with similar articulatory parameters as the speaker, the automatic speech recognition device may quickly adapt an automatic speech recognition algorithm to the speaker.
Abstract:
A method, performed by a user equipment device, for text-to-speech conversion entails sending to an articulatory model server exterior facial structural information of a person, receiving from the articulatory model server at least a portion of a predefined articulatory model that corresponds to the exterior facial structural information, the predefined articulatory model representing a voice of a modeled person who is different from the person, and generating, based at least partly on the predefined articulatory model, speech from text stored in a memory of the user equipment device. Furthermore, a method of configuring text-to-speech conversion for a user equipment device entails determining at least a portion of an articulatory model that corresponds to exterior facial structural information based on a comparison of the exterior facial structural information to exterior facial structural information stored in a database of articulatory models.
Abstract:
Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.