Abstract:
Embodiments are described for rendering spatial audio content through a system that is configured to reflect audio off of one or more surfaces of a listening environment. The system includes an array of audio drivers distributed around a room, wherein at least one driver of the array of drivers is configured to project sound waves toward one or more surfaces of the listening environment for reflection to a listening area within the listening environment and a renderer configured to receive and process audio streams and one or more metadata sets that are associated with each of the audio streams and that specify a playback location in the listening environment.
Abstract:
Embodiments are described for rendering spatial audio content through a system that is configured to reflect audio off of one or more surfaces of a listening environment. The system includes an array of audio drivers distributed around a room, wherein at least one driver of the array of drivers is configured to project sound waves toward one or more surfaces of the listening environment for reflection to a listening area within the listening environment and a renderer configured to receive and process audio streams and one or more metadata sets that are associated with each of the audio streams and that specify a playback location in the listening environment.
Abstract:
Equalizer controller and controlling method are disclosed. In one embodiment, an equalizer controller includes an audio classifier for identifying the audio type of an audio signal in real time; and an adjusting unit for adjusting an equalizer in a continuous manner based on the confidence value of the audio type as identified.
Abstract:
Embodiments are described for a system of rendering object-based audio content through a system that includes individually addressable drivers, including at least one driver that is configured to project sound waves toward one or more surfaces within a listening environment for reflection to a listening area within the listening environment; a renderer configured to receive and process audio streams and one or more metadata sets associated with each of the audio streams and specifying a playback location of a respective audio stream; and a playback system coupled to the renderer and configured to render the audio streams to a plurality of audio feeds corresponding to the array of audio drivers in accordance with the one or more metadata sets.
Abstract:
Embodiments are described for a system of rendering object-based audio content through a system that includes individually addressable drivers, including at least one driver that is configured to project sound waves toward one or more surfaces within a listening environment for reflection to a listening area within the listening environment; a renderer configured to receive and process audio streams and one or more metadata sets associated with each of the audio streams and specifying a playback location of a respective audio stream; and a playback system coupled to the renderer and configured to render the audio streams to a plurality of audio feeds corresponding to the array of audio drivers in accordance with the one or more metadata sets.
Abstract:
Embodiments are described for a method and system of rendering and playing back spatial audio content using a channel-based format. Spatial audio content that is played back through legacy channel-based equipment is transformed into the appropriate channel-based format resulting in the loss of certain positional information within the audio objects and positional metadata comprising the spatial audio content. To retain this information for use in spatial audio equipment even after the audio content is rendered as channel-based audio, certain metadata generated by the spatial audio processor is incorporated into the channel-based data. The channel-based audio can then be sent to a channel-based audio decoder or a spatial audio decoder. The spatial audio decoder processes the metadata to recover at least some positional information that was lost during the down-mix operation by upmixing the channel-based audio content back to the spatial audio content for optimal playback in a spatial audio environment.
Abstract:
Embodiments are described for a method of simultaneously localizing a set of speakers and microphones, having only the times of arrival between each of the speakers and microphones. An autodiscovery process uses an external input to set: a global translation (3 continuous parameters), a global rotation (3 continuous parameters), and discrete symmetries, i.e., an exchange of any axis pairs and/or reversal of any axis. Different time of arrival acquisition techniques may be used, such as ultrasonic sweeps or generic multitrack audio content. The autodiscovery algorithm is based in minimizing a certain cost function, and the process allows for latencies in the recordings, possibly linked to the latencies in the emission.
Abstract:
Embodiments are described for a system of rendering object-based audio content through a system that includes individually addressable drivers, including at least one driver that is configured to project sound waves toward one or more surfaces within a listening environment for reflection to a listening area within the listening environment; a renderer configured to receive and process audio streams and one or more metadata sets associated with each of the audio streams and specifying a playback location of a respective audio stream; and a playback system coupled to the renderer and configured to render the audio streams to a plurality of audio feeds corresponding to the array of audio drivers in accordance with the one or more metadata sets.
Abstract:
Embodiments are described for rendering spatial audio content through a system that is configured to reflect audio off of one or more surfaces of a listening environment. The system includes an array of audio drivers distributed around a room, wherein at least one driver of the array of drivers is configured to project sound waves toward one or more surfaces of the listening environment for reflection to a listening area within the listening environment and a renderer configured to receive and process audio streams and one or more metadata sets that are associated with each of the audio streams and that specify a playback location in the listening environment.
Abstract:
Embodiments are described for a system of rendering object-based audio content through a system that includes individually addressable drivers, including at least one driver that is configured to project sound waves toward one or more surfaces within a listening environment for reflection to a listening area within the listening environment; a renderer configured to receive and process audio streams and one or more metadata sets associated with each of the audio streams and specifying a playback location of a respective audio stream; and a playback system coupled to the renderer and configured to render the audio streams to a plurality of audio feeds corresponding to the array of audio drivers in accordance with the one or more metadata sets.