摘要:
A multimedia application system uses metadata for sensory devices. The system includes: a sensory-device engine for generating a sensory device command (SDC) for controlling the sensory devices based on sensory effect information (SEI) generated to represent sensory effects by using the sensory devices depending on video contents, user preference information (UPI) of the sensory devices and device capability information (DCI) indicative of reproducing capability of the sensory devices; and a sensory-device controller for controlling sensory devices to perform sensory effect reproduction in response to the generated SDC.
摘要:
A head motion tracking method for three-dimensional facial model animation, the head motion tracking method includes acquiring initial facial motion to be fit to an image of a three-dimensional model from an image inputted by a video camera; creating a silhouette of the three-dimensional model and projecting the silhouette; matching the silhouette created from the three-dimensional model with a silhouette acquired by a statistical feature point tracking scheme; and obtaining a motion parameter for the image of the three-dimensional model through motion correction using a texture to perform three-dimensional head motion tracking. In accordance with the present invention, natural three-dimensional facial model animation based on a real image acquired with a video camera can be performed automatically, thereby reducing time and cost.
摘要:
Provided is a method for automatic rigging and shape surface transfer of a 3-D standard mesh model based on muscles and Nurbs by using parametric control. In some embodiments, the method includes: extracting a joint-skeleton structure from scan data of dynamic objects having a skeleton structure, or input 3-D mesh models, generating a muscle and an appearance surface based on anatomy information, and generating a standard mesh model that approximates the appearance of the dynamic objects by using the joint-skeleton structure and the generated muscle and appearance surface; and performing the shape surface transfer of the standard mesh model by comparing multi-view image information of the dynamic objects with projection image information of the standard mesh model.
摘要:
An apparatus for 3D mesh compression based on quantization, includes a data analyzing unit (510) for decomposing data of an input 3D mesh model into vertices information (511) property information (512) representing property of the 3D mesh model, and connectivity information (515) between vertices constituting the 3D mesh model: and a mesh model quantizing unit (520) for producing quantized vertices and property information of the 3D mesh model by using the vertices, property and connectivity information (511, 512, 513). Further, the apparatus for 3D mesh compression based on quantization includes a decision bit encoding unit (535) for calculating a decision bit by using the quantized connectivity information and then encoding the quantized vertex information, property information and connectivity information (511, 512, 513) by using the decision bit.
摘要:
A method for creating a 3D face model by using multi-view image information, includes: creating a mesh structure for expressing an appearance of a 3D face model by using a first multi-view image obtained by capturing an expressionless face of a performer; and locating joints of a hierarchical structure in the mesh structure, by using a second multi-view image obtained by capturing an expression performance of the performer. Further, the method includes creating a 3D face model that is animated to enable reproduction of a natural expression of the performer, by setting dependency between the joints and the mesh structure.
摘要:
The present invention provides a complex comprising an aggregate of primary particles of an electrode-active transition metal compound and a fibrous carbon material, wherein said fibrous carbon material is present more densely in the surface region of the aggregate than in the inside of the aggregate.
摘要:
A method for producing a 3D facial animation using a single facial video stream, includes producing a representative joint focused on a major expression producing element in a 3D standard facial model, producing a statistical feature-point model of various facial expressions of different people in the 3D standard facial model, moving each feature-point of the statistical feature-point model by tracking a change in facial expressions of the video stream, calculating a transformation coefficient of the representative joint corresponding to a variation of the feature-point of the 3D standard facial model, and producing a 3D facial animation by applying the calculated transformation coefficient to transform the representative joint.