Abstract:
An example system for game status detection and trajectory fusion is described herein. The system includes a tracker to obtain a ball position and a player position based on images from a plurality of cameras and a fusion controller to combine multiple trajectories that are detected via the ball position to obtain a fused trajectory. The system also includes a finite state machine configured to model a game pattern, wherein a game status is determined via the ball position, the player position and the fused trajectory as input to the finite state machine.
Abstract:
A stacked transformer or inductor apparatus including a first layer with a first layer wire element extending around a center axis and a second layer with a second layer wire element. The second layer element includes side by side first and second wire components in parallel spaced relation extending around the center axis and the first wire component is connected to the first layer wire element to form a primary turn winding. A third layer includes a third layer wire element extending around the center axis and connected to the second wire component of the second layer wire element to form a secondary turn winding partially overlapping with the primary turn winding.
Abstract:
Systems and techniques for sensor-derived swing hit and direction detection are described herein. A set of sensor values may be compressed into a first lower dimension (2105). Features may be extracted from the compressed set of sensor values (2110). The features may be clustered into a set of clusters (2115). A swing action may be detected based on a distance between members of the set of clusters (2120).
Abstract:
In response to movement of an underlying structure, motion of complex objects connected to that structure may be simulated relatively quickly and without requiring extensive processing capabilities. A skeleton extraction method is used to simplify the complex object. Tracking is used to track the motion of the underlying structure, such as the user's head in a case where motion of hair is being simulated. Thus, the simulated motion is driven in response to the extent and direction of head or facial movement. A mass-spring model may be used to accelerate the simulation in some embodiments.
Abstract:
A mechanism is described to facilitate gesture matching according to one embodiment. A method of embodiments, as described herein, includes selecting a gesture from a database during an authentication phase, translating the selected gesture into an animated avatar, displaying the avatar, prompting a user to perform the selected gesture, capturing a real-time image of the user and comparing the gesture performed by the user in the captured image to the selected gesture to determine whether there is a match.
Abstract:
Examples of systems and methods for non-facial animation in facial performance driven avatar system are generally described herein. A method for facial gesture driven body animation may include capturing a series of images of a face, and computing facial motion data for each of the images in the series of images. The method may include identifying an avatar body animation based on the facial motion data, and animating a body of an avatar using the avatar body animation.
Abstract:
Examples of systems and methods for augmented facial animation are generally described herein. A method for mapping facial expressions to an alternative avatar expression may include capturing a series of images of a face, and detecting a sequence of facial expressions of the face from the series of images. The method may include determining an alternative avatar expression mapped to the sequence of facial expressions, and animating an avatar using the alternative avatar expression.
Abstract:
Various embodiments are generally directed to techniques for employing a hybrid of sequential and parallel processing to perform random sample and consensus (RANSAC). A device to perform RANSAC includes a derivation component to derive a first set of proposed models in parallel from a first set of minimal sample sets of a data set;and a comparison component to recalculate a required quantity of proposed models to derive an accurate model if a proposed model of the first set of proposed models better fits the data set than any proposed model derived prior to derivation of the first set of proposed models,and to determine whether to derive a second set of proposed models following derivation of the first set of proposed models based on a comparison of the required quantity to a quantity of previously derived proposed models that includes the first set. Other embodiments are described and claimed.
Abstract:
Apparatuses, methods and storage medium associated with processing an image are disclosed herein. In embodiments, a method for processing one or more images may include generating a plurality of pairs of keypoint features for a pair of images. Each pair of keypoint features may include a keypoint feature from each image. Further, for each pair of keypoint features, corresponding adjoin features may be generated. Additionally, for each pair of keypoint features, whether the adjoin features are similar may be determined. Whether the pair of images have at least one similar object may also be determined, based at least in part on a result of the determination of similarity between the corresponding adjoin features. Other embodiments may be described and claimed.
Abstract:
A device, method and system of video and audio sharing among communication devices, may comprise a communication device for generating and sending a packet containing information related to the video and audio, and another communication device for receiving the packet and rendering the information related to the audio and video. In some embodiments, the communication device may comprise: an audio encoding module to encode a piece of audio into an audio bit stream; an avatar data extraction module to extract avatar data from a piece of video and generate an avatar data bit stream; and a synchronization module to generate synchronization information for synchronizing the audio bit stream with the avatar parameter stream. In some embodiments, the another communication device may comprise: an audio decoding module to decode an audio bit stream into decoded audio data; an Avatar animation module to animate an Avatar model based on an Avatar data bit stream to generate an animated Avatar model; and a synchronizing and rendering module to synchronize and render the decoded audio data and the animated Avatar model by utilizing the synchronization information.