Abstract:
Disclosed herein are systems and methods for presenting audio content in mixed reality environments. A method may include receiving a first input from an application program; in response to receiving the first input, receiving, via a first service, an encoded audio stream; generating, via the first service, a decoded audio stream based on the encoded audio stream; receiving, via a second service, the decoded audio stream; receiving a second input from one or more sensors of a wearable head device; receiving, via the second service, a third input from the application program, wherein the third input corresponds to a position of one or more virtual speakers; generating, via the second service, a spatialized audio stream based on the decoded audio stream, the second input, and the third input; presenting, via one or more speakers of the wearable head device, the spatialized audio stream.
Abstract:
A method for improving speech recognition by a speech recognition system includes obtaining a voice sample from a speaker; storing the voice sample of the speaker as a voice model in a voice model database; identifying an area from which sound matching the voice model for the speaker is coming; providing one or more audio signals corresponding to sound received from the identified area to the speech recognition system for processing.
Abstract:
A method, an apparatus, and a computer program product for initiating at least one process in a vehicle are provided. The apparatus determines an approximate position of the vehicle. In addition, the apparatus determines a shortest time period based on the determined approximate position in which one or more registered drivers of the vehicle is in proximity to the vehicle. Furthermore, the apparatus determines whether to initiate the at least one process within the vehicle based on the determined time period.
Abstract:
Various arrangements for selecting a function to be executed on a mobile device are presented. A processor may receive proximity data that indicates the mobile device is proximate to a user. The processor may receive acceleration data that indicates a direction of acceleration. In response to receiving the proximity data indicating that the mobile device is proximate to the user, the processor may identify a position of the mobile device in relation to the user using the acceleration data. In response to identifying the position of the mobile device in relation to the user using the acceleration data, the processor may select a function to be executed.
Abstract:
Apparatuses, methods, systems and computer-readable media for using proximity inputs on or near a touch screen lens to select objects within a field of view are presented. In some embodiments, a viewing apparatus (e.g. head mounted display, augmented reality goggles) may include at least one lens, wherein the lens can sense touches or near-touches and output data indicative of a location of the proximity input by the user. A processor may receive the data and may select an object within the field of view of the user corresponding to the data, wherein the object and the location of the proximity input on or near the lens by the user are on a common line of sight. In some embodiments, the viewing apparatus may include at least one camera that is configured to record at least one image representative of the user's field of view.
Abstract:
FIG. 1 is a front view of a display panel with a graphical user interface showing a first embodiment of our design; FIG. 2 is a front view of a display panel with a graphical user interface showing a second embodiment thereof; and FIG. 3 is a front view of a display panel with a graphical user interface showing a third embodiment thereof. The broken lines, including all text, illustrate portions of the graphical user interface or mixed reality environment and form no part of the design.
Abstract:
Apparatuses, methods, systems and computer-readable media for using proximity inputs on or near a touch screen lens to select objects within a field of view are presented. In some embodiments, a viewing apparatus (e.g. head mounted display, augmented reality goggles) may include at least one lens, wherein the lens can sense touches or near-touches and output data indicative of a location of the proximity input by the user. A processor may receive the data and may select an object within the field of view of the user corresponding to the data, wherein the object and the location of the proximity input on or near the lens by the user are on a common line of sight. In some embodiments, the viewing apparatus may include at least one camera that is configured to record at least one image representative of the user's field of view.
Abstract:
A method for improving speech recognition by a speech recognition system includes obtaining a voice sample from a speaker; storing the voice sample of the speaker as a voice model in a voice model database; identifying an area from which sound matching the voice model for the speaker is coming; providing one or more audio signals corresponding to sound received from the identified area to the speech recognition system for processing.
Abstract:
Various arrangements for selecting a function to be executed on a mobile device are presented. A processor may receive proximity data that indicates the mobile device is proximate to a user. The processor may receive acceleration data that indicates a direction of acceleration. In response to receiving the proximity data indicating that the mobile device is proximate to the user, the processor may identify a position of the mobile device in relation to the user using the acceleration data. In response to identifying the position of the mobile device in relation to the user using the acceleration data, the processor may select a function to be executed.
Abstract:
A method of enabling speech commands in an application includes identifying, by a computer processor, a user interaction element within a resource of the application; extracting, by the computer processor, text associated with the identified user interaction element; generating, by the computer processor, a voice command corresponding to the extracted text; and adding the generated voice command to a grammar associated with the application.