Abstract:
The present invention provides a local multi-resolution 3-D face-inherent model generation apparatus, including one or more 3-D facial model generation cameras for photographing a face of an object at various angles in order to obtain one or more 3-D face models, a 3-D face-inherent model generation unit for generating a 3-D face-inherent model by composing the one or more 3-D face models, a local photographing camera for photographing a local part of the face of the object, a control unit for controlling the position of the local photographing camera on the 3-D face-inherent model, and a local multi-resolution 3-D face-inherent model generation unit for generating a local multi-resolution face-inherent model by composing an image captured by the local photographing camera and the 3-D face-inherent model, a local multi-resolution 3-D face-inherent model generation using the local multi-resolution 3-D face-inherent model generation apparatus, and a skin management system.
Abstract:
Disclosed herein are an apparatus and method for creating animation by capturing the motions of a non-rigid object. The apparatus includes a geometry mesh reconstruction unit, a motion capture unit, and a content creation unit. The geometry mesh reconstruction unit receives moving images captured by a plurality of cameras, and generates a reconstruction mesh set for each frame. The motion capture unit generates mesh graph sets for the reconstruction mesh set and generates motion data, including motion information, using the mesh graph sets. The content creation unit creates three-dimensional (3D) content for a non-rigid object by generating a final transformation mesh set, having a topology similar to that of the reconstruction mesh set, using the motion data.
Abstract:
A face avatar generating apparatus includes: a face feature information extraction unit for receiving a face photo and extracting face feature information from the face photo and a two-dimensional (2D) avatar generation unit for selecting at least one region from the face photo based on the face feature information, and exaggerating or beautifying the selected region to create a 2D avatar image. The apparatus further includes a 3D avatar generation unit for modifying a standard 3D face model through a comparison with the standard 3D model based on the face feature information and pre-stored standard information to create a 3D avatar image.
Abstract:
A multimedia application system uses metadata for sensory devices. The system includes: a sensory-device engine for generating a sensory device command (SDC) for controlling the sensory devices based on sensory effect information (SEI) generated to represent sensory effects by using the sensory devices depending on video contents, user preference information (UPI) of the sensory devices and device capability information (DCI) indicative of reproducing capability of the sensory devices; and a sensory-device controller for controlling sensory devices to perform sensory effect reproduction in response to the generated SDC.
Abstract:
A method for providing knowledge includes creating a three-dimensional virtual space, generating temporal and spatial information including knowledge and space information concerning the three-dimensional virtual space at a specific time point, and associating the temporal and spatial information with the three-dimensional virtual space to be stored in a memory. The creating the three-dimensional virtual space may be created based on a map or a two-dimensional image.
Abstract:
There is provided a method of representing and animating a 2D (Two-Dimensional) character in a 3D (Three-Dimensional) space for a character animation. The method includes performing a pre-processing operation in which data of a character that is required to represent and animate the 2D character like a 3D character is prepared and stored and producing the character animation using the stored data.
Abstract:
A system for managing face data includes a global face capturing unit configured to capture a global face image; and a global face data generation unit configured to obtain shape information and texture information of global face data, and generate the global face data. Further, the system includes a local face capturing unit configured to capture a plurality of local face images; and a global face posture extraction unit configured to estimate a position and a direction of the face of a captured user. Furthermore, the system includes a local capturing device posture extraction unit configured to extract posture information of the local face capturing unit; and a local face data generation unit configured to generate texture information and shape information, and generate local face data.
Abstract:
A method of deforming a shape of a human body model includes the steps of reorganizing human body model data into a joint-skeleton structure-based Non-Uniform Rational B-spline (NURBS) surface model, generating statistical deformation information about control parameters of the NURBS surface model based on parameters of joints and key section curves for specific motions, and deforming the shape of the human body model based on the NURBS surface model and the statistical deformation information. The human body model data includes three-dimensional (3D) human body scan data and a 3D polygon mesh.
Abstract:
Disclosed herein is a motion capture apparatus and method. The motion capture apparatus includes a three-dimensional (3D) appearance reconstruction unit for reconstructing a 3D appearance of an entity using a multi-view image obtained by capturing the entity at a T+1-th frame. A 3D skeleton generation unit generates a 3D skeleton of the T+1-th frame using the 3D appearance and a model of the entity having a motion of a T-th frame. A motion reconstruction unit reconstructs a motion of the T+1-th frame of the model by changing the motion of the model based on the 3D skeleton of the T+1-th frame.
Abstract:
The present invention relates to a method of creating an animatable digital clone includes receiving input multi-view images of an actor captured by at least two cameras and reconstructing a three-dimensional appearance therefrom, accepting shape information selectively based on a probability of photo-consistency in the input multi-view images obtained from the reconstruction and transferring a mesh topology of a reference human body model onto a shape of the actor obtained from the reconstruction. The method further includes generating an initial human body model of the actor via transfer of the mesh topology utilizing sectional shape information of the actor's joints, and generating a genuine human body model of the actor from learning genuine behavioral characteristics of the actor by applying the initial human body model to multi-view posture learning images where performance of a predefined motion by the actor is recorded.