Abstract:
Techniques are provided for estimation of human orientation and facial pose, in images that include depth information. A methodology embodying the techniques includes detecting a human in an image generated by a depth camera and estimating an orientation category associated with the detected human. The estimation is based on application of a random forest classifier, with leaf node template matching, to the image. The orientation category defines a range of angular offsets relative to an angle corresponding to the human facing the depth camera. The method also includes performing a three dimensional (3D) facial pose estimation of the detected human, based on detected facial landmarks, in response to a determination that the estimated orientation category includes the angle corresponding to the human facing the depth camera.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar, initiating communication, capturing an image, detecting a face in the image, extracting features from the face, converting the facial features to avatar parameters, and transmitting at least one of the avatar selection or avatar parameters.
Abstract:
The present disclosure is directed to face detection window refinement using depth. Existing face detection systems may perform face detection by analyzing portions of visual data such as an image, video, etc. identified by sub-windows. These sub-windows are now determined only based on pixels, and thus may number in the millions. Consistent with the present disclosure, at least depth data may be utilized to refine the size and appropriateness of sub-windows that identify portions of the visual data to analyze during face detection, which may substantially reduce the number of sub-windows to be analyzed, the total data processing burden, etc. For example, at least one device may comprise user interface circuitry including capture circuitry to capture both visual data and depth data. Face detection circuitry in the at least one device may refine face detection by determining criteria for configuring the sub-windows that will be used in face detection.
Abstract:
A system and method for selecting an advertisement to present to a consumer includes detecting facial regions in the image, identifying one or more consumer characteristics (mood, gender, age, etc.) of said consumer in the image, identifying one or more advertisements to present to the consumer based on a comparison of the consumer characteristics with an advertisement database including a plurality of advertisement profiles, and presenting a selected one of the identified advertisement to the consumer on a media device.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar; initiating communication; detecting a user input; identifying the user input; identifying an animation command based on the user input; generating avatar parameters; and transmitting at least one of the animation command and the avatar parameters.
Abstract:
A mechanism is described for facilitating dynamic user-based customization of advertisement content at computing devices according to one embodiment. A method of embodiments, as described herein, includes receiving an advertiser content to be published on an avatar list, where the advertiser content is associated with an advertising entity, and verifying the advertiser content for publication, where verifying further includes assigning a ranking to the advertiser content. The ranking represents a position on the avatar list. The method may further include transmitting a publication notification identifying the ranking assigned to the advertiser content, and facilitating an auction for bidding to allow the advertising entity to obtain a higher ranking for the advertiser content than the assigned ranking, if the assigned ranking is rejected by the advertising entity.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar; initiating communication; detecting a user input; identifying the user input; identifying an animation command based on the user input; generating avatar parameters; and transmitting at least one of the animation command and the avatar parameters.
Abstract:
Techniques are provided for detection of humans in images that include depth information. A methodology embodying the techniques includes segmenting an image into multiple windows and estimating the distance to a subject in each window based on depth pixel values in that window, and filtering to reject windows with sizes that are outside of a desired window size range. The desired window size range is based on the estimated subject distance and the focal length of the depth camera that produced the image. The method further includes generating classifier features for each remaining windows (post-filtering) for use by a cascade classifier. The cascade classifier creates candidate windows for further consideration based on a preliminary detection of a human in any of the remaining windows. The method further includes merging neighboring candidate windows and executing a linear classifier on the merged candidate windows to verify the detection of a human.
Abstract:
A video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar, initiating communication, capturing an image, detecting a face in the image, determining facial characteristics from the face, including eye movement and eyelid movement of a user indicative of direction of user gaze and blinking, respectively, converting the facial features to avatar parameters, and transmitting at least one of the avatar selection or avatar parameters.
Abstract:
Generally this disclosure describes a video communication system that replaces actual live images of the participating users with animated avatars. A method may include selecting an avatar; initiating communication; detecting a user input; identifying the user input; identifying an animation command based on the user input; generating avatar parameters; and transmitting at least one of the animation command and the avatar parameters.