Abstract:
Provided are a face de-identification method and system and a graphical user interface (GUI) provision method for face de-identification employing facial image generation. According to the face de-identification method and system and the GUI provision method, a facial area including eyes, a nose, and a mouth in a face of a person detected in an input image is replaced with a de-identified facial area generated through deep learning to maintain the face in a natural shape while protecting the person's portrait right. Accordingly, qualitative degradation of content is prevented, and viewers' concentration on the image is increased.
Abstract:
A method for receiving a mono sound source audio signal including phase information as an input, and separating into a plurality of signals may comprise performing initial convolution and down-sampling on the inputted mono sound source audio signal; generating an encoded signal by encoding the inputted signal using at least one first dense block and at least one down-transition layer; generating a decoded signal by decoding the encoded signal using at least one second dense block and at least one up-transition layer; and performing final convolution and resize on the decoded signal.
Abstract:
An apparatus for generating text from an image may comprise: a memory configured to store at least one instruction; and a processor configured to execute the at least one instruction, wherein the processor is further configured to generate encoding information for an image based on the image and extract text information related to content of the image based on a degree of association with the encoding information.
Abstract:
Disclosed are a learning data generation method and apparatus needed to learn animation characters on the basis of deep learning. The learning data generation method needed to learn animation characters on the basis of deep learning may include collecting various images from an external source using wired/wireless communication, acquiring character images from the collected images using a character detection module, clustering the acquired character images, selecting learning data from among the clustered images, and inputting the selected learning data to an artificial neural network for character recognition.
Abstract:
An apparatus and method for providing projection mapping-based augmented reality (AR). According to an exemplary embodiment, the apparatus includes an input to acquire real space information and user information; and a processor to recognize a real environment by using the acquired real space information and the acquired user information, map the recognized real environment to a virtual environment, generate augmented content that changes corresponding to a change in space or a user's movement, and project and visualize the generated augmented content through a projector.
Abstract:
The present invention relates to an apparatus and method for identifying music in a content, The present invention includes extracting and storing a fingerprint of an original audio in an audio fingerprint DB; extracting a first fingerprint of a first audio in the content; and searching for a fingerprint corresponding to the fingerprint of the first audio in the audio fingerprint DB, wherein the first audio is audio data in a music section detected from the content.
Abstract:
The present invention relates to a method and an apparatus for matching a virtual object in a virtual environment, the method including generating a point cloud of a non-rigid object, matching a low resolution virtual model to the point cloud, and implementing a high resolution model using the matched model.
Abstract:
Provided are an apparatus and method for detecting a plurality of arms and hands by using a three-dimensional (3D) image. The apparatus includes an image input unit configured to acquire a 3D image of an object, an arm detecting unit configured to detect one or more component-unit candidate regions of the object in the 3D image, and detect one or more arm regions by using arm detection feature information, extracted from each of the candidate regions, and a pattern recognition algorithm, and a hand detecting unit configured to calculate a position of a hand and a position of a wrist in each of the arm regions detected by the arm detecting unit, and detect a hand region by using the position of the hand and the position of the wrist.