Abstract:
This disclosure describes an apparatus, such as a wireless communication device, that applies a direct evaluation technique to render triangles for the 3D graphical environment. The apparatus includes a rendering engine that defines a rectangular area of pixels, referred to as a bounding box, that bounds the area to be rendered. The rendering engine evaluates coordinates associated with the pixels of the rectangular area to selectively render those pixels that fall within the triangular area. The direct evaluation triangle rendering algorithm may require fewer complex operations than the more computationally intensive interpolation process employed by other systems. As a result, the apparatus may present a 3D graphical environment while preserving as much as possible the available power.
Abstract:
This disclosure describes an apparatus, such as a wireless communication device, that applies a direct evaluation technique to render triangles for the 3D graphical environment. The apparatus includes a rendering engine that defines a rectangular area of pixels, referred to as a bounding box, that bounds the area to be rendered. The rendering engine evaluates coordinates associated with the pixels of the rectangular area to selectively render those pixels that fall within the triangular area. The direct evaluation triangle rendering algorithm may require fewer complex operations than the more computationally intensive interpolation process employed by other systems. As a result, the apparatus may present a 3D graphical environment while preserving as much as possible the available power.
Abstract:
A speech processing system modifies various aspects of input speech according to a user-selected one of various preprogrammed voice fonts. Initially, the speech converter receives a formants signal representing an input speech signal and a pitch signal representing the input signal's fundamental frequency. One or both of the following may also be received: a voicing signal comprising an indication of whether the input speech signal is voiced, unvoiced, or mixed, and/or a gain signal representing the input speech signal's energy. The speech converter also receives user selection of one of multiple preprogrammed voice fonts, each specifying a manner of modifying one or more of the received signals (i.e., formants, voicing, pitch, gain). The speech converter modifies at least one of the formants, voicing, pitch, and/or gain signals as specified by the selected voice font.
Abstract:
A voice recognition rejection scheme for capturing an utterance includes the steps accepting the utterance, applying an N-best algorithm to the utterance, or rejecting the utterance. The utterance is accepted if a first predefined relationship exists between one or more closest comparison results for the utterance with respect to a stored word and one or more differences between the one or more closest comparison results and one or more other comparison results between the utterance and one or more other stored words. An N-best algorithm is applied to the utterance if a second predefined relationship exists between the one or more closest comparison results and the one or more differences between the one or more closest comparison results and the one or more other comparison results. The utterance is rejected if a third predefined relationship exists between the one or more closest comparison results and the one or more differences between the one or more closest comparison results and the one or more other comparison results. One of the one or more other comparison results may advantageously be a next-closest comparison result for the utterance and another store word. The first, second, and third predefined relationships may advantageously be linear relationships.
Abstract:
A system and method for collision resolution in a communication system having multiple transmission stations communicating on a shared slotted communication channel. The system and method includes a procedure for organizing the stations in a structure wherein each station has a fixed address corresponding to a static location in the structure. The procedure further indicates to the multiple transmission stations a slot type for each communication slot, the slot type selected from a group including contention slot type and reservation slot type, and accepts transmissions of contention slot data packets in contention slots on the communication channel from the stations, the contention slot data packets each optionally containing a reservation quantity for reservation slot data packets to be sent by an originating station upon successful transmission of a contention slot data packet originating at the originating station, the reservation slot data packets transmitted in reservation slots dedicated to the originating station. The procedure detects collisions in the transmissions of packets originating from the transmitting station with packets transmitted from other transmitting stations; and upon detecting a collision in a contention slots, recursively entitles a subset of the locations in the structure to transmit in the subsequent contention slot regardless of whether the station transmitted during the contention slot in which the collision occurred, wherein membership in the subset indicates that a station corresponding to a location in the subset is entitled. Further systems and methods are disclosed.
Abstract:
In general, this disclosure describes techniques for providing a gesture-based user interface. For example, according to some aspects of the disclosure, a user interface generally includes a camera and a computing device that identifies and tracks the motion of one or more fingertips of a user. In some examples, the user interface is configured to identify predefined gestures (e.g., patterns of motion) associated with certain motions of the user's fingertips. In another example, the user interface is configured to identify hand postures (e.g., patterns of showing up of fingertips). Accordingly, the user can interact with the computing device by performing the gestures.
Abstract:
This disclosure describes a multi-stage tessellation technique for tessellating a curve during graphics rendering. In particular, a first tessellation stage tessellates the curve into a first set of line segments that each represents a portion of the curve. A second tessellation stage further tessellates the portion of the curve represented by each of the line segments of the first set into additional line segments that more finely represent the shape of the curve. In this manner, each portion of the curve that was represented by only one line segment after the first tessellation stage is represented by more than one line segment after the second tessellation stage. In some instances, more than two tessellation stages may be performed to tessellate the curve.
Abstract:
A method and system for speech recognition combines different types of engines in order to recognize user-defined digits and control words, predefined digits and control words, and nametags. Speaker-independent engines are combined with speaker-dependent engines. A Hidden Markov Model (HMM) engine is combined with Dynamic Time Warping (DTW) engines.
Abstract:
A method and apparatus for constructing voice templates for a speaker-independent voice recognition system includes segmenting a training utterance to generate time-clustered segments, each segment being represented by a mean. The means for all utterances of a given word are quantized to generate template vectors. Each template vector is compared with testing utterances to generate a comparison result. The comparison is typically a dynamic time warping computation. The training utterances are matched with the template vectors if the comparison result exceeds at least one predefined threshold value, to generate an optimal path result, and the training utterances are partitioned in accordance with the optimal path result. The partitioning is typically a K-means segmentation computation. The partitioned utterances may then be re-quantized and re-compared with the testing utterances until the at least one predefined threshold value is not exceeded.
Abstract:
A method and apparatus for real time speech recognition with and without speaker dependency which includes the following steps. Converting the speech signals into a series of primitive sound spectrum parameter frames; detecting the beginning and ending of speech according to the primitive sound spectrum parameter frame, to determine the sound spectrum parameter frame series; performing non-linear time domain normalization on the sound spectrum parameter frame series using sound stimuli, to obtain speech characteristic parameter frame series with predefined lengths on the time domain; performing amplitude quantization normalization on the speech characteristic parameter frames; comparing the speech characteristic parameter frame series with the reference samples, to determine the reference sample which most closely matches the speech characteristic parameter frame series; and determining the recognition result according to the most closely matched reference sample.