Abstract:
A portable eye tracker device is disclosed which includes a frame, at least one optics holding member, a movement sensor, and a control unit. The frame may be a frame adapted for wearing by a user. The at least one optics holding member may include at least one illuminator configured to selectively illuminate at least a portion of at least one eye of the user, and at least one image sensor configured to capture image data representing images of at least a portion of at least one eye of the user. The movement sensor may be configured to detect movement of the frame. The control unit may be configured to control the at least one illuminator for the selective illumination of at least a portion of at least one eye of the user, receive the image data from the image sensors, and receive information from the movement sensor.
Abstract:
A first apparatus performs the following: determining, using at least two microphone signals corresponding to left and right microphone signals and using at least one further microphone signal, directional information of the left and right microphone signals; outputting a first signal corresponding to the left microphone signal; outputting a second signal corresponding to the right microphone signal; and outputting a third signal corresponding to the determined directional information. Another apparatus performs the following: performing at least one of the following: outputting first and second signals as stereo output signals; or converting the first and second signals to mid and side signals, and converting, using directional information for the first and second signals, the mid and side signals to at least one of binaural signals or multi-channel signals, and outputting the corresponding binaural signals or multi-channel signals. Additional apparatus, program products, and methods are disclosed.
Abstract:
In an oscillator (100), a plurality of piezoelectric vibrators (111 to 113) supported by vibrator support mechanisms (120) individually output highly directional sound waves. The plurality of piezoelectric vibrators (111 to 113) is formed by dividing a laminated body of an elastic member and a piezoelectric substance by the vibrator support mechanisms (120). Since it is not necessary to arrange the plurality of piezoelectric vibrators in a matrix, the entirety of a device can be small-sized. A sound deflection unit that deflects a sound wave which is output by at least one of the piezoelectric vibrators may be further included.
Abstract:
A mobile platform includes a microphone array and is capable of implementing beamforming to amplify or suppress audio information from a sound source. The sound source is indicated through a user input, such as pointing the mobile platform in the direction of the sound source or through a touch screen display interface. The mobile platform further includes orientation sensors capable of detecting movement of the mobile platform. When the mobile platform moves with respect to the sound source, the beamforming is adjusted based on the data from the orientation sensors so that beamforming is continuously implemented in the direction of the sound source. The audio information from the sound source may be included or suppressed from a telephone or video-telephony conversation. Images or video from a camera may be likewise controlled based on the data from the orientation sensors.
Abstract:
Spatial audio signals can include audio objects that can be respectively encoded and rendered at each of multiple different depths. In an example, a method for encoding a spatial audio signal can include receiving audio scene information from an audio capture source in an environment, and receiving a depth characteristic of a first object in the environment. The depth characteristic can be determined using information from a depth sensor. A correlation can be identified between at least a portion of the audio scene information and the first object. The spatial audio signal can be encoded using the portion of the audio scene and the depth characteristic of the first object.
Abstract:
An audio signal playing method and apparatus, and an electronic device are provided. The method comprises: separating, from a first audio signal, a recorded audio signal corresponding to each of at least one sound source; on the basis of the first audio signal, determining a real-time orientation of each of the at least one sound source relative to the head of a user; for each sound source, according to the real-time orientation of the sound source and the recorded audio signal corresponding to the sound source, generating a target direct audio signal corresponding to the sound source, and generating a target reverberated audio signal corresponding to the sound source; and playing a second audio signal that is generated by means of fusing the target direct audio signal and the target reverberated audio signal corresponding to each sound source.
Abstract:
A system comprises a microphone array and a processor. The microphone array detects first sound signal and the second sound signal. Each of the first and second sound signals has a particular frequency band. Each of the first and second sound signals is originated from a particular sound source. The processor receives the first and second sound signals. The processor amplifies the first sound signals, each with a different amplification order. The processor disregards the second sound signal, where the second sound signal includes interference noise signals. The processor determines that the first sound signals indicate that a vehicle is within a threshold distance from an autonomous vehicle and traveling in a direction toward the autonomous vehicle. In response, the processor instructs the autonomous vehicle to perform a minimal risk condition operation. The minimal risk condition operation includes pulling over or stopping the autonomous vehicle.
Abstract:
Spatial audio signals can include audio objects that can be respectively encoded and rendered at each of multiple different depths. In an example, a method for encoding a spatial audio signal can include receiving audio scene information from an audio capture source in an environment, and receiving a depth characteristic of a first object in the environment. The depth characteristic can be determined using information from a depth sensor. A correlation can be identified between at least a portion of the audio scene information and the first object. The spatial audio signal can be encoded using the portion of the audio scene and the depth characteristic of the first object.
Abstract:
A method for capturing a sound source includes: capturing a space where a microphone array is located to generate an image by a camera, wherein the microphone array is configured to receive a sound generated by the sound source and generate a sound source coordinate of the sound source relative to the microphone array; searching for a sub-image belonging to the microphone array within the images by a computing device connected to the camera; calculating a microphone coordinate of the microphone array relative to the camera by the computing device according to the sub-image; calculating a required control parameter by the computing device at least according to the sound source coordinate and the microphone coordinate; adjusting a capturing direction by the camera to capture the sound source at least according to the required control parameter.
Abstract:
Microphone arrays comprise several microphone capsules, the outputs of which being electronically combined for directional recording of sound. The directional and frequency properties of the microphone array depend on the number and positions of the microphone array. In order to obtain the smallest possible microphone array with only few microphone capsules, which, however, has an essentially uniform directional and frequency dependence over a speech frequency range, is scalable and robust against small incorrect positioning of the capsules, fifteen or twenty-one microphone capsules (K15,11-K15,35, K21,11-K21,37) are arranged on a carrier such that they lie on three similar branches, each with the same number of microphone capsules, which are rotated against each other by 120°. Each of the microphone capsules lies on a corner of a triangle of a grid in a flat isometric coordinate system with three axes rotated by 120° against each other and forming the grid of equilateral triangles.