Abstract:
Example embodiments disclosed herein relate to user experience oriented audio signal processing. There is provided a method for user experience oriented audio signal processing. The method includes obtaining a first audio signal from an audio sensor of an electronic device; computing, based on the first audio signal, a compensation factor for an acoustic path from the electronic device to a listener and applying the compensation factor to a second audio signal outputted from the electronic device. Corresponding system and computer program products are disclosed.
Abstract:
Embodiments of client device and method for audio or video conferencing are described. An embodiment includes an offset detecting unit, a configuring unit, an estimator and an output unit. The offset detecting unit detects an offset of speech input to the client device. The configuring unit determines a voice latency from the client device to every far end. The estimator estimates a time when a user at the far end perceives the offset based on the voice latency. The output unit outputs a perceivable signal indicating that a user at the far end perceives the offset based on the time estimated for the far end. The perceivable signal is helpful to avoid collision between parties.
Abstract:
An audio processing method and apparatus are described. In one embodiment, at least one first sub-band of a first audio signal is suppressed to obtain a reduced first audio signal with reserved sub-bands; suppressing at least one second sub-band of the at least one second audio signal to obtain at least one reduced second audio signal with reserved sub-bands; and mixing the reduced first audio signal and at least one reduced second audio signal. Alternatively, a first spatial auditory property is assigned to a first audio signal so that the first audio signal may be perceived as originating from a first position. Alternatively, rhythmic similarity between at least two audio signals is detected, and time scaling is applied to an audio signal in response to relatively high rhythmic similarity between the audio signal and the other audio signal(s); and then at least two audio signals are mixed.
Abstract:
An audio processing method and an audio processing apparatus are described. A mono-channel audio signal is transformed into a plurality of first subband signals. Proportions of a desired component and a noise component are estimated in each of the subband signals. Second subband signals corresponding respectively to a plurality of channels are generated from each of the first subband signals. Each of the second subband signals comprises a first component and a second component obtained by assigning a spatial hearing property and a perceptual hearing property different from the spatial hearing property to the desired component and the noise component in the corresponding first subband signal respectively, based on a multi-dimensional auditory presentation method. The second subband signals are transformed into signals for rendering with the multi-dimensional auditory presentation method. By assigning different hearing properties to desired sound and noise, the intelligibility of the audio signal can be improved.
Abstract:
An audio signal with a temporal sequence of blocks or frames is received or accessed. Features are determined as characterizing aggregately the sequential audio blocks/frames that have been processed recently, relative to current time. The feature determination exceeds a specificity criterion and is delayed, relative to the recently processed audio blocks/frames. Voice activity indication is detected in the audio signal. VAD is based on a decision that exceeds a preset sensitivity threshold and is computed over a brief time period, relative to blocks/frames duration, and relates to current block/frame features. The VAD and the recent feature determination are combined with state related information, which is based on a history of previous feature determinations that are compiled from multiple features, determined over a time prior to the recent feature determination time period. Decisions to commence or terminate the audio signal, or related gains, are outputted based on the combination.
Abstract:
Some implementations involve analyzing audio packets received during a time interval that corresponds with a conversation analysis segment to determine network jitter dynamics data and conversational interactivity data. The network jitter dynamics data may provide an indication of jitter in a network that relays the audio data packets. The conversational interactivity data may provide an indication of interactivity between participants of a conversation represented by the audio data. A jitter buffer size may be controlled according to the network jitter dynamics data and the conversational interactivity data. The time interval may include a plurality of talkspurts.
Abstract:
In an audio processing system (300), a filtering section (350, 400): receives subband signals (410, 420, 430) corresponding to audio content of a reference signal (301) in respective frequency subbands; receives subband signals (411, 421, 431) corresponding to audio content of a response signal (304) in the respective subbands; and forms filtered inband references (412, 422, 432) by applying respective filters (413, 423, 433) to the subband signals of the reference signal. For a frequency subband: filtered crossband references (424, 425) are formed by multiplying, by scalar factors (426, 427), filtered inband references of other subbands; a composite filtered reference (428) is formed by summing the filtered inband reference of the subband (422) and the filtered crossband references; a residual signal (429) is computed as a difference between the composite filtered reference and the subband signal of the response signal corresponding to the subband; and the scalar factors and the filter applied to the subband signal of the reference signal corresponding to the subband are adjusted based on the residual signal.
Abstract:
Various disclosed implementations involve processing and/or playback of a recording of a conference involving a plurality of conference participants. Some implementations disclosed herein involve receiving audio data corresponding to a recording of at least one conference involving a plurality of conference participants. In some examples, only a portion of the received audio data will be selected as playback audio data. The selection process may involve a topic selection process, a talkspurt filtering process and/or an acoustic feature selection process. Some examples involve receiving an indication of a target playback time duration. Selecting the portion of audio data may involve making a time duration of the playback audio data within a threshold time difference of the target playback time duration.
Abstract:
A service request for communication services for communication clients is received. In response, a communication service network is set up to support the communication services. Routing metadata is generated for each of the communication clients. The routing metadata is to be used by each of the communication clients for sharing service quality information with a respective peer communication client over a light-weight peer-to-peer (P2P) network. The routing metadata is downloaded to each of the communication clients. A communication client may exchange service signaling packets or service data packets over the communication service network. When the communication client determines that there is a problematic region in a bitstream received from the communication server, the communication client can request a peer communication client for a service quality information portion related to the problematic region.
Abstract:
Example embodiments disclosed herein relate to separated audio analysis and processing. A system for processing an audio signal is disclosed. The system includes an audio analysis module configured to analyze an input audio signal to determine a processing parameter for the input audio signal, the input audio signal being represented in time domain. The system also includes an audio processing module configured to process the input audio signal in parallel with the audio analysis module. The audio processing module includes a time domain filter configured to filter the input audio signal to obtain an output audio signal in the time domain, and a filter controller configured to control a filter coefficient of the time domain filter based on the processing parameter determined by the audio analysis module. Corresponding method and computer program product of processing an audio signal are also disclosed.