Abstract:
Embodiments are described for an adaptive audio system that processes audio data comprising a number of independent monophonic audio streams. One or more of the streams has associated with it metadata that specifies whether the stream is a channel-based or object-based stream. Channel-based streams have rendering information encoded by means of channel name; and the object-based streams have location information encoded through location expressions encoded in the associated metadata. A codec packages the independent audio streams into a single serial bitstream that contains all of the audio data. This configuration allows for the sound to be rendered according to an allocentric frame of reference, in which the rendering location of a sound is based on the characteristics of the playback environment (e.g., room size, shape, etc.) to correspond to the mixer's intent. The object position metadata contains the appropriate allocentric frame of reference information required to play the sound correctly using the available speaker positions in a room that is set up to play the adaptive audio content.
Abstract:
Audio signals (201) are received. The audio signals include left and right surround channels (206). The audio signals are played back using far-field loudspeakers (101-108, 401-406) distributed around a space (111, 409) having a plurality of listener positions (112, 410). The left and right surround channels are played back by a pair of far-field loudspeakers (103, 106, 403, 405) arranged at opposite sides of the space having the plurality of listener positions. An audio component (208) coinciding with or approximating audio content common to the left and right surround channels is obtained. The audio component is played back using at least a pair of near-field transducers (109, 110, 407, 408) arranged at one of the listener positions. Associated systems (100, 400), methods (800) and computer program products are provided. Systems (300), methods (900) and computer program products providing a bitstream (303) comprising the audio signals and the audio component are also provided, as well as a computer-readable medium with data (700) representing such audio content.
Abstract:
Example embodiments disclosed herein relate to audio object clustering. A method for metadata-preserved audio object clustering is disclosed. The method comprises classifying a plurality of audio objects into a number of categories based on information to be preserved in metadata associated with the plurality of audio objects. The method further comprises assigning a predetermined number of clusters to the categories and allocating an audio object in each of the categories to at least one of the clusters according to the assigning. Corresponding system and computer program product are also disclosed.
Abstract:
In some embodiments, methods for generating an object based audio program including screen-related metadata indicative of at least one warping degree parameter for at least one audio object, or generating a speaker channel-based program including by warping audio content of an object based audio program to a degree determined at least in part by at least one warping degree parameter, or methods for decoding or rendering any such audio program. Other aspects are systems configured to perform such audio signal generation, decoding, or rendering, and audio processing units (e.g., decoders or encoders) including a buffer memory which stores at least one segment of any such audio program.
Abstract:
Embodiments are described for a method of rendering audio for playback through headphones comprising receiving digital audio content, receiving binaural rendering metadata generated by an authoring tool processing the received digital audio content, receiving playback metadata generated by a playback device, and combining the binaural rendering metadata and playback metadata to optimize playback of the digital audio content through the headphones.
Abstract:
Improved tools for authoring and rendering audio reproduction data are provided. Some such authoring tools allow audio reproduction data to be generalized for a wide variety of reproduction environments. Audio reproduction data may be authored by creating metadata for audio objects. The metadata may be created with reference to speaker zones. During the rendering process, the audio reproduction data may be reproduced according to the reproduction speaker layout of a particular reproduction environment.