Abstract:
A method for creating a head-related impulse response (HRIR) for use in rendering audio for playback through headphones comprises receiving location parameters for a sound including azimuth, elevation, and range relative to a head of a listener, applying a spherical head model to the azimuth, elevation, and range input parameters to generate binaural HRIR values, computing a pinna model using the azimuth and elevation parameters to apply to the binaural HRIR values to pinna modeled HRIR values, computing a torso model using the azimuth and elevation parameters to apply to the pinna modeled HRIR values to generate pinna and torso modeled HRIR values, and computing a near-field model using the azimuth and range parameters to apply to the pinna and torso modeled HRIR values to generate pinna, torso and near-field modeled HRIR values.
Abstract:
Audio characteristics of audio data corresponding to a plurality of audio channels may be determined. The audio characteristics may include spatial parameter data. Decorrelation filtering processes for the audio data may be based, at least in part, on the audio characteristics. The decorrelation filtering processes may cause a specific inter-decorrelation signal coherence (“IDC”) between channel-specific decorrelation signals for at least one pair of channels. The channel-specific decorrelation signals may be received and/or determined. Inter-channel coherence (“ICC”) between a plurality of audio channel pairs may be controlled. Controlling ICC may involve at receiving an ICC value and/or determining an ICC value based, at least partially, on the spatial parameter data. A set of IDC values may be based, at least partially, on the set of ICC values. A set of channel-specific decorrelation signals, corresponding with the set of IDC values, may be synthesized by performing operations on the filtered audio data.
Abstract:
Audio processing methods may involve receiving audio data corresponding to a plurality of audio channels. The audio data may include a frequency domain representation corresponding to filterbank coefficients of an audio encoding or processing system. A decorrelation process may be performed with the same filterbank coefficients used by the audio encoding or processing system. The decorrelation process may be performed without converting coefficients of the frequency domain representation to another frequency domain or time domain representation. The decorrelation process may involve selective or signal-adaptive decorrelation of specific channels and/or specific frequency bands. The decorrelation process may involve applying a decorrelation filter to a portion of the received audio data to produce filtered audio data. The decorrelation process may involve using a non-hierarchal mixer to combine a direct portion of the received audio data with the filtered audio data according to spatial parameters.
Abstract:
Received audio data may include a first set of frequency coefficients and a second set of frequency coefficients. Spatial parameters for at least part of the second set of frequency coefficients may be estimated, based at least in part on the first set of frequency coefficients. The estimated spatial parameters may be applied to the second set of frequency coefficients to generate a modified second set of frequency coefficients. The first set of frequency coefficients may correspond to a first frequency range (for example, an individual channel frequency range) and the second set of frequency coefficients may correspond to a second frequency range (for example, a coupled channel frequency range). Combined frequency coefficients of a composite coupling channel may be based on frequency coefficients of two or more channels. Cross-correlation coefficients, between frequency coefficients of a first channel and the combined frequency coefficients, may be computed.
Abstract:
Decorrelation filter parameters for audio data may be based, at least in part, on audio characteristics such as tonality information and/or transient information. Determining the audio characteristics may involve receiving explicit audio characteristics with the audio data and/or determining audio characteristics based on one or more attributes of the audio data. The decorrelation filter parameters may include dithering parameters and/or randomly selected pole locations for at least one pole of an all-pass filter. The dithering parameters and/or pole locations may involve a maximum stride value for pole movement. In some examples, the maximum stride value may be substantially zero for highly tonal signals of the audio data. The dithering parameters and/or pole locations may be bounded by constraint areas within which pole movements are constrained. The constraint areas may or may not be fixed. In some implementations, different channels of the audio data may share the same constraint areas.