Abstract:
Techniques related to game focus estimation in team sports for multi-camera immersive video are discussed. Such techniques include selecting regions of a scene comprising a sporting event, generating a node graph and sets of features for the selected regions, and determining a game focus region of the selected regions by applying a graph node classification model based on the node graph and sets of features.
Abstract:
A system (600) includes multiple cameras (104) disposed about an area (102), a processor (606), and a memory (608) communicatively coupled to the processor. The memory stores instructions that cause the processor to receive a set of video data (602) associated with a first change in acceleration and a second image frame (628) associated with a second change in acceleration. In an embodiment, the operations include generating a set of virtual camera actions (630) based on the first image frame and the second image frame.
Abstract:
Apparatuses, methods and storage medium associated with generating and animating avatars are disclosed. The apparatus may comprise an avatar generator to receive an image having a face of a user; analyze the image to identify various facial and related components of the user; access an avatar database to identify corresponding artistic renditions for the various facial and related components stored in the database; and combine the corresponding artistic renditions for the various facial and related components to form an avatar, without user intervention. The apparatus may further comprise an avatar animation engine to animate the avatar in accordance with a plurality of animation messages having facial expression or head pose parameters that describe facial expressions or head poses of a user determined from an image of the user.
Abstract:
Apparatuses, methods and storage medium associated with animating and rendering an avatar are disclosed herein. In embodiments, an apparatus may include a facial expression and speech tracker to respectively receive a plurality of image frames and audio of a user, and analyze the image frames and the audio to determine and track facial expressions and speech of the user. The tracker may further select a plurality of blend shapes, including assignment of weights of the blend shapes, for animating the avatar, based on tracked facial expressions or speech of the user. The tracker may select the plurality of blend shapes, including assignment of weights of the blend shapes, based on the tracked speech of the user, when visual conditions for tracking facial expressions of the user are determined to be below a quality threshold. Other embodiments may be disclosed and/or claimed.
Abstract:
A video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar, initiating communication, capturing an image, detecting a face in the image, extracting features from the face, converting the facial features to avatar parameters, and transmitting at least one of the avatar selection or avatar parameters. The system is configured to allow a user to select a new avatar during active communication with a remote user.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar; initiating communication; detecting a user input; identifying the user input; identifying an animation command based on the user input; generating avatar parameters; and transmitting at least one of the animation command and the avatar parameters.
Abstract:
A method for player and ball tracking post processing is described herein. The method includes optically tracking a trajectory of a player and ball and determining at least one affinity player for the currently tracked player. Additionally, the method includes completing the optical trajectory for the currently tracked player based on an estimated trajectory from the affinity player and stabilizing ball tracking during game break. The method further includes calculating virtual camera movement.
Abstract:
Methods, systems and apparatuses may provide for technology that identifies a player captured in a multi-camera video feed of a game that involves the identified player and estimates a first field of view from the perspective of the identified player for a selected from of the multi-camera video feed. Additionally, the technology automatically generates, based on the first field of view, a camera path for a replay of the selected frame from the perspective of the identified player. In one example, the technology also determines a trajectory of a projectile captured in the multi-camera video feed, estimates, based on the trajectory, a second field of view from the perspective of the projectile, and automatically generates, based on the second field of view, a replay of one or more selected frames of the multi-camera video feed from the perspective of the projectile.
Abstract:
A multi-camera architecture for detecting and tracking a ball in real-time. The multi-camera architecture includes network interface circuitry to receive a plurality of real-time videos taken from a plurality of high-resolution cameras. Each of the high-resolution cameras simultaneously captures a sports event, wherein each of the plurality of high-resolution cameras includes a viewpoint that covers an entire playing field where the sports event is played. The multi-camera architecture further includes one or more processors coupled to the network interface circuity and one or more memory devices coupled to the one or more processors. The one or more memory devices includes instructions to determine the location of the ball for each frame of the plurality of real-time videos, which when executed by the one or more processors, cause the multi-camera architecture to simultaneously perform one of a detection scheme or a tracking scheme on a frame from each of the plurality of real-time videos to detect the ball used in the sports event and perform a multi-camera build to determine a location of the ball in 3D for the frame from each of the plurality of real-time videos using one of detection or tracking results for each of the cameras.
Abstract:
A mechanism is described for facilitating real-time multi-view detection of objects in multi-camera environments, according to one embodiment. A method of embodiments, as described herein, includes mapping first lines associated with objects to a ground plane; and forming clusters of second lines corresponding to the first lines such that an intersection point in a cluster represents a position of an object on the ground plane.