Abstract:
Techniques for facial expression capture for character animation are described. In one or more implementations, facial key points are identified in a series of images. Each image, in the series of images, is normalized from the identified facial key points. Facial features are determined from each of the normalized images. Then a facial expression is classified, based on the determined facial features, for each of the normalized images. In additional implementations, a series of images are captured that include performances of one or more facial expressions. The facial expressions in each image of the series of images are classified by a facial expression classifier. Then the facial expression classifications are used by a character animator system to produce a series of animated images of an animated character that include animated facial expressions that are associated with the facial expression classification of the corresponding image in the series of images.
Abstract:
Different candidate windows in an image are identified, such as by sliding a rectangular or other geometric shape of different sizes over an image to identify portions of the image (groups of pixels in the image). The candidate windows are analyzed by a set of convolutional neural networks, which are cascaded so that the input of one convolutional neural network layer is based on the input of another convolutional neural network layer. Each convolutional neural network layer drops or rejects one or more candidate windows that the convolutional neural network layer determines does not include an object (e.g., a face). The candidate windows that are identified as including an object (e.g., a face) are analyzed by another one of the convolutional neural network layers. The candidate windows identified by the last of the convolutional neural network layers are the indications of the objects (e.g., faces) in the image.
Abstract:
Feature interpolation techniques are described. In a training stage, features are extracted from a collection of training images and quantized into visual words. Spatial configurations of the visual words in the training images are determined and stored in a spatial configuration database. In an object detection stage, a portion of features of an image are extracted from the image and quantized into visual words. Then, a remaining portion of the features of the image are interpolated using the visual words and the spatial configurations of visual words stored in the spatial configuration database.
Abstract:
Each of multiple images is analyzed to determine how the colors of the pixels of the image are distributed throughout the color space of the image. Different covariance based characteristics of the image are determined that identify a direction, as well as magnitude in each direction, of the distribution of colors of the image pixels. These different covariance based characteristics that are determined for an image can be saved as associated with the image, allowing the characteristics to be accessed and used as a basis for searching the images to identify particular types of images. These different covariance based characteristics can also be used to order the images identified by a search.
Abstract:
Content update and suggestion techniques are described. In one or more implementations, techniques are implemented to generate suggestions that are usable to guide creative professionals in updating content such as images, video, sound, multimedia, and so forth. A variety of techniques are usable to generate suggestions for the content professionals. In one example, suggestions are based on shared characteristics of images licensed by users of a content sharing service, e.g., licensed by the users. In another example, suggestions are based on metadata of the images licensed by the users, the metadata describing characteristics of how the images are created. These suggestions are then used to guide transformation of a user's image such that the image exhibits these characteristics and thus has an increased likelihood of being desired for licensing by customers of the service.
Abstract:
Content update and suggestion techniques are described. In one or more implementations, techniques are implemented to generate suggestions that are usable to guide creative professionals in updating content such as images, video, sound, multimedia, and so forth. A variety of techniques are usable to generate suggestions for the content professionals. In one example, suggestions are based on shared characteristics of images licensed by users of a content sharing service, e.g., licensed by the users. In another example, suggestions are based on metadata of the images licensed by the users, the metadata describing characteristics of how the images are created. These suggestions are then used to guide transformation of a user's image such that the image exhibits these characteristics and thus has an increased likelihood of being desired for licensing by customers of the service.
Abstract:
Techniques for facial expression capture for character animation are described. In one or more implementations, facial key points are identified in a series of images. Each image, in the series of images, is normalized from the identified facial key points. Facial features are determined from each of the normalized images. Then a facial expression is classified, based on the determined facial features, for each of the normalized images. In additional implementations, a series of images are captured that include performances of one or more facial expressions. The facial expressions in each image of the series of images are classified by a facial expression classifier. Then the facial expression classifications are used by a character animator system to produce a series of animated images of an animated character that include animated facial expressions that are associated with the facial expression classification of the corresponding image in the series of images.
Abstract:
Accelerating object detection techniques are described. In one or more implementations, adaptive sampling techniques are used to extract features from an image. Coarse features are extracted from the image and used to generate an object probability map. Then, dense features are extracted from high-probability object regions of the image identified in the object probability map to enable detection of an object in the image. In one or more implementations, cascade object detection techniques are used to detect an object in an image. In a first stage, exemplars in a first subset of exemplars are applied to features extracted from the multiple regions of the image to detect object candidate regions. Then, in one or more validation stages, the object candidate regions are validated by applying exemplars from the first subset of exemplars and one or more additional subsets of exemplars.
Abstract:
Different candidate windows in an image are identified, such as by sliding a rectangular or other geometric shape of different sizes over an image to identify portions of the image (groups of pixels in the image). The candidate windows are analyzed by a set of convolutional neural networks, which are cascaded so that the input of one convolutional neural network layer is based on the input of another convolutional neural network layer. Each convolutional neural network layer drops or rejects one or more candidate windows that the convolutional neural network layer determines does not include an object (e.g., a face). The candidate windows that are identified as including an object (e.g., a face) are analyzed by another one of the convolutional neural network layers. The candidate windows identified by the last of the convolutional neural network layers are the indications of the objects (e.g., faces) in the image.
Abstract:
Accelerating object detection techniques are described. In one or more implementations, adaptive sampling techniques are used to extract features from an image. Coarse features are extracted from the image and used to generate an object probability map. Then, dense features are extracted from high-probability object regions of the image identified in the object probability map to enable detection of an object in the image. In one or more implementations, cascade object detection techniques are used to detect an object in an image. In a first stage, exemplars in a first subset of exemplars are applied to features extracted from the multiple regions of the image to detect object candidate regions. Then, in one or more validation stages, the object candidate regions are validated by applying exemplars from the first subset of exemplars and one or more additional subsets of exemplars.