Abstract:
Techniques related to game focus estimation in team sports for multi-camera immersive video are discussed. Such techniques include selecting regions of a scene comprising a sporting event, generating a node graph and sets of features for the selected regions, and determining a game focus region of the selected regions by applying a graph node classification model based on the node graph and sets of features.
Abstract:
A system (600) includes multiple cameras (104) disposed about an area (102), a processor (606), and a memory (608) communicatively coupled to the processor. The memory stores instructions that cause the processor to receive a set of video data (602) associated with a first change in acceleration and a second image frame (628) associated with a second change in acceleration. In an embodiment, the operations include generating a set of virtual camera actions (630) based on the first image frame and the second image frame.
Abstract:
Methods and apparatus to generate photo-realistic three-dimensional models of a photographed environment are disclosed. An apparatus includes an object position calculator to determine a three-dimensional (3D) position of an object detected within a first image of an environment and within a second image of the environment. The apparatus further includes a 3D model generator to generate a 3D model of the environment based on the first image and the second image. The apparatus also includes a model integrity analyzer to detect a difference between the 3D position of the object and the 3D model. The 3D model generator automatically modifies the 3D model based on the difference in response to the difference satisfying a confidence threshold.
Abstract:
Apparatuses, methods and storage medium associated with generating and animating avatars are disclosed. The apparatus may comprise an avatar generator to receive an image having a face of a user; analyze the image to identify various facial and related components of the user; access an avatar database to identify corresponding artistic renditions for the various facial and related components stored in the database; and combine the corresponding artistic renditions for the various facial and related components to form an avatar, without user intervention. The apparatus may further comprise an avatar animation engine to animate the avatar in accordance with a plurality of animation messages having facial expression or head pose parameters that describe facial expressions or head poses of a user determined from an image of the user.
Abstract:
Apparatuses, methods and storage medium associated with animating and rendering an avatar are disclosed herein. In embodiments, an apparatus may include a facial expression and speech tracker to respectively receive a plurality of image frames and audio of a user, and analyze the image frames and the audio to determine and track facial expressions and speech of the user. The tracker may further select a plurality of blend shapes, including assignment of weights of the blend shapes, for animating the avatar, based on tracked facial expressions or speech of the user. The tracker may select the plurality of blend shapes, including assignment of weights of the blend shapes, based on the tracked speech of the user, when visual conditions for tracking facial expressions of the user are determined to be below a quality threshold. Other embodiments may be disclosed and/or claimed.
Abstract:
Apparatus, systems, media and/or methods may involve animating avatars. User facial motion data may be extracted that corresponds to one or more user facial gestures observed by an image capture device when a user emulates a source object. An avatar animation may be provided based on the user facial motion data. Also, script data may be provided to the user and/or the user facial motion data may be extracted when the user utilizes the script data. Moreover, audio may be captured and/or converted to a predetermined tone. Source facial motion data may be extracted and/or an avatar animation may be provided based on the source facial motion data. A degree of match may be determined between the user facial motion data of a plurality of users and the source facial motion data. The user may select an avatar as a user avatar and/or a source object avatar.
Abstract:
A video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar, initiating communication, capturing an image, detecting a face in the image, extracting features from the face, converting the facial features to avatar parameters, and transmitting at least one of the avatar selection or avatar parameters. The system is configured to allow a user to select a new avatar during active communication with a remote user.
Abstract:
Various embodiments are generally directed to cooperation among networked devices to obtain and use a multiple-frame screenshot. In one embodiment, an apparatus comprises a processor circuit executing instructions that cause the processor circuit to receive a signal conveying a video stream from a source device; visually present video frames of the video stream on a display associated with the apparatus; maintain a rolling buffer comprising a plurality of video frames; recurringly update the plurality of video frames to represent a subset of video frames of the video stream most recently presented on the display; receive a signal indicative of a capture command; and preserve the subset of video frames as a multiple-frame screenshot in response to the capture command.
Abstract:
This invention provides a computer-implemented method for a 3D graphical user interface. The computer-implemented method comprises: receiving visual data of a user, wherein the visual data includes 3D visual data; determining a 3D distance from a 3D display to the user based at least in part on the received 3D visual data; and adjusting a 3D projection distance from the 3D display to the user based at least in part on the determined 3D distance to the user.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar; initiating communication; detecting a user input; identifying the user input; identifying an animation command based on the user input; generating avatar parameters; and transmitting at least one of the animation command and the avatar parameters.