Abstract:
A system, article, and method of curved object recognition using image matching for image processing, comprising: using paired 2D-3D point(s) to form a perspective projection function to determine a geometric correspondence between target object and reference object(s) and that converts the 2D points into 3D points at the target object.
Abstract:
Attributes of large scale computer vision systems may be made available to users of more limited processer-based systems by dynamically and adaptively updating recognizers into a smaller scale device from a connected larger scale device, based on the user's situational context and behavior. A recognizer is a hardware, software or firmware module specialized to use computer vision to recognize a defined class of imaged objects.
Abstract:
Computer-readable storage media, computing devices and methods are discussed herein. In embodiments, a computing device may be configured to perform facial recognition based on gradient based feature extractions of images of faces. In embodiments, the computing device may be configured to determine directional matching patterns of the images from the gradient based feature extraction and may utilize these directional matching patterns in performing a facial recognition analysis of the images of faces. Other embodiments may be described and/or claimed.
Abstract:
Avatar animation systems disclosed herein provide high quality, real-time avatar animation that is based on the varying countenance of a human face. The real-time provision of high quality avatar animation is enabled, at least in part, by a multi-frame regressor that is configured to map information descriptive of facial expressions depicted in two or more images to information descriptive of a single avatar blend shape. The two or more images may be temporally sequential images. This multi-frame regressor implements a machine learning component that generates the high quality avatar animation from information descriptive of a subject's face and/or information descriptive of avatar animation frames previously generated by the multi-frame regressor. The machine learning component may be trained using a set of training images that depict human facial expressions and avatar animation authored by professional animators to reflect facial expressions depicted in the set of training images.
Abstract:
Techniques are provided for motion estimation using hybrid video imaging based on frame-based capture and event-based capture. A methodology implementing the techniques according to an embodiment includes receiving a sequence of pixel events, generated asynchronously by an event-based video camera, and receiving a sequence of image frames generated by a frame-based video camera at a frame sampling period. The method also includes integrating a subset of the sequence of pixel events, occurring within the frame sampling period between two of the image frames, to generate a pixel motion vector representing motion of the pixel between the two image frames. The method further includes mapping the pixel motion vector to a tile of one of the image frames to generate an estimated motion vector associated with that tile.
Abstract:
Apparatuses, methods and storage medium associated with face detection are disclosed. In embodiments, an apparatus may comprise one or more processors, cache memory, a lookup table of feature patterns, and a face detector. The lookup table is to be loaded into the cache memory, during operation, for use to detect whether a face is present in an image. The feature patterns within the lookup table are organized within the lookup table in accordance with relative connectivity of the feature patterns. The face detector to detect whether a face is present in an image, may use the lookup table of feature patterns loaded into the cache memory during detection for the face. Other embodiments may be described and/or claimed.
Abstract:
Apparatuses, methods and storage medium associated with 3D face model reconstruction are disclosed herein. In embodiments, an apparatus may include a facial landmark detector, a model fitter and a model tracker. The facial landmark detector may be configured to detect a plurality of landmarks of a face and their locations within each of a plurality of image frames. The model fitter may be configured to generate a 3D model of the face from a 3D model of a neutral face, in view of detected landmarks of the face and their locations within a first one of the plurality of image frames. The model tracker may be configured to maintain the 3D model to track the face in subsequent image frames, successively updating the 3D model in view of detected landmarks of the face and their locations within each of successive ones of the plurality of image frames. In embodiments, the facial landmark detector may include a face detector, an initial facial landmark detector, and one or more facial landmark detection linear regressors. Other embodiments may be described and/or claimed.
Abstract:
Apparatuses, methods and storage medium associated with animating and rendering an avatar are disclosed herein. In embodiments, an apparatus may include a facial mesh tracker to receive a plurality of image frames, detect facial action movements of a face and head pose gestures of a head within the plurality of image frames, and output a plurality of facial motion parameters and head pose parameters that depict facial action movements and head pose gestures detected, all in real time, for animation and rendering of an avatar. The facial action movements and head pose gestures may be detected through inter-frame differences for a mouth and an eye, or the head, based on pixel sampling of the image frames. The facial action movements may include opening or closing of a mouth, and blinking of an eye. The head pose gestures may include head rotation such as pitch, yaw, roll, and head movement along horizontal and vertical direction, and the head comes closer or goes farther from the camera. Other embodiments may be described and/or claimed.
Abstract:
Methods and apparatus to convert images for computer-vision systems are disclosed. An example computer-vision system includes an image converter (112) to convert a near infrared light first image (202) to form a visible light image (206), and to update a coefficient of the image converter (112) based on a difference (214), an object recognizer (102) to recognize an object (208) in the first visible light image (206), and an object recognition analyzer (210) to determine the difference (214) between the object (208) recognized in the first visible light image (206) and an object (212) associated with the near infrared light image (202).
Abstract:
System, apparatus, method, and computer readable media for on-the-fly captured image data object tracking. An image or video stream is processed to detect and track an object in concurrence with generation of the stream by a camera module. In one exemplary embodiment, HD image frames are processed at a rate of 30 fps, or more, to track one or more target object. In embodiments, object detection is validated prior to employing detected object descriptor(s) as learning data to generate or update an object model. A device platform including a camera module and comporting with the exemplary architecture may provide 3A functions based on objects robustly tracked in accordance with embodiments.