Abstract:
A device includes one or more processors configured to perform signal processing including a linear transformation and a non-linear transformation of an input signal to generate a reference target signal. The reference target signal has a linear component associated with the linear transformation and a non-linear component associated with the non-linear transformation. The one or more processors are also configured to perform linear filtering of the input signal by controlling adaptation of the linear filtering to generate an output signal that substantially matches the linear component of the reference target signal.
Abstract:
Methods, systems, computer-readable media, and apparatuses for gesture control are presented. One example includes indicating, based on information from a first audio input signal, a presence of an object in proximity to a microphone, and increasing a volume level in response to the indicating.
Abstract:
A device includes a memory configured to store category labels associated with categories of a natural language processing library. A processor is configured to analyze input audio data to generate a text string and to perform natural language processing on at least the text string to generate an output text string including an action associated with a first device, a speaker, a location, or a combination thereof. The processor is configured to compare the input audio data to audio data of the categories to determine whether the input audio data matches any of the categories and, in response to determining that the input audio data does not match any of the categories: create a new category label, associate the new category label with at least a portion of the output text string, update the categories with the new category label, and generate a notification indicating the new category label.
Abstract:
A method of operation of a device includes receiving an input signal at the device. The input signal is generated using at least one microphone. The input signal includes a first signal component having a first amount of wind turbulence noise and a second signal component having a second amount of wind turbulence noise that is greater than the first amount of wind turbulence noise. The method further includes generating, based on the input signal, an output signal at the device. The output signal includes the first signal component and a third signal component that replaces the second signal component. A first frequency response of the input signal corresponds to a second frequency response of the output signal.
Abstract:
A multichannel acoustic system (MAS) comprises an arrangement of microphones and loudspeakers and a multichannel acoustic processor (MAP) to together enhance conversational speech between two or more persons in a shared acoustic space such as an automobile. The enhancements are achieved by receiving sound signals substantially originating from relatively near sound sources; filtering the sound signals to cancel at least one echo signal detected for at least one microphone from among the plurality of microphones; filtering the sound signals received by the plurality of microphones to cancel at least one feedback signal detected for at least one microphone from among the plurality of microphones; and reproducing the filtered sound signals for each microphone from among the plurality of microphones on a subset of loudspeakers corresponding that are relatively far from the source microphone.
Abstract:
A method of selectively authorizing access includes obtaining, at an authentication device, first information corresponding to first synthetic biometric data. The method also includes obtaining, at the authentication device, first common synthetic data and second biometric data. The method further includes generating, at the authentication device, second common synthetic data based on the first information and the second biometric data. The method also includes selectively authorizing, by the authentication device, access based on a comparison of the first common synthetic data and the second common synthetic data.
Abstract:
Systems, devices, and methods are described for recognizing and focusing on at least one source of an audio communication as part of a communication including a video image and an audio communication derived from two or more microphones when a relative position between the microphones is known. In certain embodiments, linked audio and video focus areas providing location information for one or more sound sources may each be associated with different user inputs, and an input to adjust a focus in either the audio or video domain may automatically adjust the focus in the another domain.
Abstract:
A method for noise characteristic dependent speech enhancement by an electronic device is described. The method includes determining a noise characteristic of input audio. Determining a noise characteristic of input audio includes determining whether noise is stationary noise and determining whether the noise is music noise. The method also includes determining a noise reference based on the noise characteristic. Determining the noise reference includes excluding a spatial noise reference from the noise reference when the noise is stationary noise and including the spatial noise reference in the noise reference when the noise is not music noise and is not stationary noise. The method further includes performing noise suppression based on the noise characteristic.
Abstract:
A personalized (i.e., speaker-derivable) bandwidth extension is provided in which the model used for bandwidth extension is personalized (e.g., tailored) to each specific user. A training phase is performed to generate a bandwidth extension model that is personalized to a user. The model may be subsequently used in a bandwidth extension phase during a phone call involving the user. The bandwidth extension phase, using the personalized bandwidth extension model, will be activated when a higher band (e.g., wideband) is not available and the call is taking place on a lower band (e.g., narrowband).
Abstract:
A method for encoding multiple directional audio signals using an integrated codec by a wireless communication device is disclosed. The wireless communication device records a plurality of directional audio signals. The wireless communication device also generates a plurality of audio signal packets based on the plurality of directional audio signals. At least one of the audio signal packets includes an averaged signal. The wireless communication device further transmits the plurality of audio signal packets.