Abstract:
Methods and apparatuses are disclosed that can combine audio content from two encoded input signals into a new encoded output signal without requiring a decode or re-encode of audio content in either encoded input signal. Encoded data representing audio content and spatial location of audio objects in two different input encoded signals are combined to generate an encoded output signal that has encoded data representing audio objects from both of the input encoded signals.
Abstract:
Methods for generating an object based audio program which is renderable in a personalizable manner, e.g., to provide an immersive, perception of audio content of the program. Other embodiments include steps of delivering (e.g., broadcasting), decoding, and/or rendering such a program. Rendering of audio objects indicated by the program may provide an immersive experience. The audio content of the program may be indicative of multiple object channels (e.g., object channels indicative of user-selectable and user-configurable objects, and typically also a default set of objects which will be rendered in the absence of a selection by a user) and a bed of speaker channels. Another aspect is an audio processing unit (e.g., encoder or decoder) configured to perform, or which includes a buffer memory which stores at least one frame (or other segment) of an object based audio program (or bitstream thereof) generated in accordance with, any embodiment of the method.
Abstract:
Embodiments are disclosed for channel-based audio (CBA) (e.g., 22.2-ch audio) to object-based audio (OBA) conversion. The conversion includes converting CBA metadata to object audio metadata (OAMD) and reordering the CBA channels based on channel shuffle information derived in accordance with channel ordering constraints of the OAMD. The OBA with reordered channels is rendered in a playback device using the OAMD or in a source device, such as a set-top box or audio/video recorder. In an embodiment, the CBA metadata includes signaling that indicates a specific OAMD representation to be used in the conversion of the metadata. In an embodiment, pre-computed OAMD is transmitted in a native audio bitstream (e.g., AAC) for transmission (e.g., over HDMI) or for rendering in a source device. In an embodiment, pre-computed OAMD is transmitted in a transport layer bitstream (e.g., ISO BMFF, MPEG4 audio bitstream) to a playback device or source device.
Abstract:
Methods for generating an object based audio program which is renderable in a personalizable manner, e.g., to provide an immersive, perception of audio content of the program. Other embodiments include steps of delivering (e.g., broadcasting), decoding, and/or rendering such a program. Rendering of audio objects indicated by the program may provide an immersive experience. The audio content of the program may be indicative of multiple object channels (e.g., object channels indicative of user-selectable and user-configurable objects, and typically also a default set of objects which will be rendered in the absence of a selection by a user) and a bed of speaker channels. Another aspect is an audio processing unit (e.g., encoder or decoder) configured to perform, or which includes a buffer memory which stores at least one frame (or other segment) of an object based audio program (or bitstream thereof) generated in accordance with, any embodiment of the method.
Abstract:
Embodiments are directed to a method of rendering adaptive audio by receiving input audio comprising channel-based audio, audio objects, and dynamic objects, wherein the dynamic objects are classified as sets of low-priority dynamic objects and high-priority dynamic objects, rendering the channel-based audio, the audio objects, and the low-priority dynamic objects in a first rendering processor of an audio processing system, and rendering the high-priority dynamic objects in a second rendering processor of the audio processing system. The rendered audio is then subject to virtualization and post-processing steps for playback through soundbars and other similar limited height capable speakers.
Abstract:
The positions of a plurality of speakers at a media consumption site are determined. Audio information in an object-based format is received. Gain adjustment value for a sound content portion in the object-based format may be determined based on the position of the sound content portion and the positions of the plurality of speakers. Audio information in a ring-based channel format is received. Gain adjustment value for each ring-based channel in a set of ring-based channels may be determined based on the ring to which the ring-based channel belongs and the positions of the speakers at a media consumption site.
Abstract:
Methods for generating an object based audio program which is renderable in a personalizable manner, e.g., to provide an immersive, perception of audio content of the program. Other embodiments include steps of delivering (e.g., broadcasting), decoding, and/or rendering such a program. Rendering of audio objects indicated by the program may provide an immersive experience. The audio content of the program may be indicative of multiple object channels (e.g., object channels indicative of user-selectable and user-configurable objects, and typically also a default set of objects which will be rendered in the absence of a selection by a user) and a bed of speaker channels. Another aspect is an audio processing unit (e.g., encoder or decoder) configured to perform, or which includes a buffer memory which stores at least one frame (or other segment) of an object based audio program (or bitstream thereof) generated in accordance with, any embodiment of the method.
Abstract:
Methods for generating an object based audio program which is renderable in a personalizable manner, e.g., to provide an immersive, perception of audio content of the program. Other embodiments include steps of delivering (e.g., broadcasting), decoding, and/or rendering such a program. Rendering of audio objects indicated by the program may provide an immersive experience. The audio content of the program may be indicative of multiple object channels (e.g., object channels indicative of user-selectable and user-configurable objects, and typically also a default set of objects which will be rendered in the absence of a selection by a user) and a bed of speaker channels. Another aspect is an audio processing unit (e.g., encoder or decoder) configured to perform, or which includes a buffer memory which stores at least one frame (or other segment) of an object based audio program (or bitstream thereof) generated in accordance with, any embodiment of the method.
Abstract:
Embodiments are directed to a method of representing spatial rendering metadata for processing in an object-based audio system that allows for lossless interpolation and/or re-sampling of the metadata. The method comprises time stamping the metadata to create metadata instances, and encoding an interpolation duration to with each metadata instance that specifies the time to reach a desired rendering state for the respective metadata instance. The re-sampling of metadata is useful for re-clocking metadata to an audio coder and for the editing audio content.
Abstract:
Methods for generating or decoding an encoded audio bitstream including audio data and supplementary data (e.g., metadata and/or unrelated audio data), where at least some of the supplementary data is included as LSBs of audio segments, and/or at least some of the supplementary data is included in guard bands. Typical embodiments provide a scalable and video synchronous format compatible with real-time and file-based infrastructure components that support the SMPTE 337 format for carrying data in AES3 serial bitstreams, and/or provide a framework for extending distribution codecs to scale beyond an 8-channel limit to support multiples of 8 channels synchronously across multiple AES3 interfaces. Another aspect is an audio processing unit configured to perform any embodiment of the method or including a buffer memory storing at least one segment of an audio bitstream generated in accordance with any embodiment of the method.