Abstract:
Disclosed herein are an apparatus and method for classifying clothing attributes based on deep learning. The apparatus includes memory for storing at least one program and a processor for executing the program, wherein the program includes a first classification unit for outputting a first classification result for one or more attributes of clothing worn by a person included in an input image, a mask generation unit for outputting a mask tensor in which multiple mask layers respectively corresponding to principal part regions obtained by segmenting a body of the person included in the input image are stacked, a second classification unit for outputting a second classification result for the one or more attributes of the clothing by applying the mask tensor, and a final classification unit for determining and outputting a final classification result for the input image based on the first classification result and the second classification result.
Abstract:
Disclosed herein are an apparatus and method for generating robot interaction behavior. The method for generating robot interaction behavior includes generating co-speech gesture of a robot corresponding to utterance input of a user, generating a nonverbal behavior of the robot, that is a sequence of next joint positions of the robot, which are estimated from joint positions of the user and current joint positions of the robot based on a pre-trained neural network model for robot pose estimation, and generating a final behavior using at least one of the co-speech gesture and the nonverbal behavior.
Abstract:
Disclosed herein are an apparatus and method for determining the speech and motion properties of an interactive robot. The method for determining the speech and motion properties of an interactive robot includes receiving interlocutor conversation information including at least one of voice information and image information about an interlocutor that interacts with an interactive robot, extracting at least one of a verbal property and a nonverbal property of the interlocutor by analyzing the interlocutor conversation information, determining at least one of a speech property and a motion property of the interactive robot based on at least one of the verbal property, the nonverbal property, and context information inferred from a conversation between the interactive robot and the interlocutor, and controlling the operation of the interactive robot based on at least one of the determined speech property and motion property of the interactive robot.
Abstract:
A virtual content provision apparatus and method for augmenting usability of a real object. The virtual content provision apparatus includes a real object information acquisition unit for acquiring real object information corresponding to a real object through an input module, a virtual content search unit for searching for any one piece of virtual content based on the real object information, a virtual interface projection unit for projecting a virtual interface corresponding to the virtual content onto the real object through an output module, a user input detection unit for detecting user input on the virtual interface based on the input module, and a virtual content provision unit for, when the user input is detected, extracting virtual information related to the user input based on the virtual content, and projecting and providing the virtual information to correspond to the real object information through the output module.
Abstract:
Disclosed are a bin-picking system and a bin-picking method. A bin-picking system includes a transformable bin-picking box; a supporting unit configured to support the bottom part of the bin-picking box and be movable upward and downward; and a control unit configured to change an alignment of at least one bin-picking candidate object by transforming the bin-picking box by controlling a movement of the supporting unit if no bin-picking target object is detected from the at least one bin-picking candidate object existing inside bin-picking box and placed on the supporting unit to facilitate detection of a bin-picking target object and bin-picking.
Abstract:
Disclosed herein are a facial recognition method and apparatus based on masking. The facial recognition method based on masking includes performing detection and normalization of a face area including five landmarks in an input image, separating a face portion other than a portion covered by an obstacle in the detected face area, generating a mask corresponding to the separated face portion when there is a portion covered by an obstacle in the separated face portion, applying the mask to an image in a face image registration database and extracting a feature of a masking region, and determining whether it is an identical person based on the extracted feature value.
Abstract:
Disclosed herein are a human behavior recognition apparatus and method. The human behavior recognition apparatus includes a multimodal sensor unit for generating at least one of image information, sound information, location information, and Internet-of-Things (IoT) information of a person using a multimodal sensor, a contextual information extraction unit for extracting contextual information for recognizing actions of the person from the at least one piece of generated information, a human behavior recognition unit for generating behavior recognition information by recognizing the actions of the person using the contextual information and recognizing a final action of the person using the behavior recognition information and behavior intention information, and a behavior intention inference unit for generating the behavior intention information based on context of action occurrence related to each of the actions of the person included in the behavior recognition information.
Abstract:
Disclosed herein is a method for integrated anomaly detection. The method includes detecting a thing object and a human object in input video using a first neural network, and tracking the human object, and detecting an anomalous situation based on an object detection result and a human object tracking result.
Abstract:
Disclosed herein are an apparatus and method for determining a modality of interaction between a user and a robot. The apparatus includes memory in which at least one program is recorded and a processor for executing the program. The program may perform recognizing a user state and an environment state by sensing circumstances around a robot, determining an interaction capability state associated with interaction with a user based on the recognized user state and environment state, and determining the interaction behavior of the robot for the interaction with the user based on the user state, the environment state, and the interaction capability state.
Abstract:
Disclosed herein are an apparatus and method for registering face poses for face recognition. The apparatus includes a face detection unit for detecting the face of a user from an image including the face of the user; a pose recognition unit for recognizing the face pose of the face of the user based on the degree of rotation of the face of the user; a registration interface unit for providing an interface for showing information about whether the face poses of the face of the user are registered; and a face registration unit for registering the face pose of the face of the user when the face pose of the face of the user is recognized as an unregistered face pose based on the interface.