Abstract:
This document describes techniques and apparatuses for area-dependent image enhancement. These techniques are capable of enabling selection, through a touch-enabled mobile-device display, of an area of a photographic image through movement of a spatially-variable implement, such as brush icon moved over the image. Selected areas can be enhanced differently than other areas, such as to apply sharpening to the selected area and blurring to a non-selected area.
Abstract:
Neural network patch aggregation and statistical techniques are described. In one or more implementations, patches are generated from an image, e.g., randomly, and used to train a neural network. An aggregation of outputs of patches processed by the neural network may be used to label an image using an image descriptor, such as to label aesthetics of the image, classify the image, and so on. In another example, the patches may be used by the neural network to calculate statistics describing the patches, such as to describe statistics such as minimum, maximum, median, and average of activations of image characteristics of the individual patches. These statistics may also be used to support a variety of functionality, such as to label the image as described above.
Abstract:
Image zooming is described. In one or more implementations, zoomed croppings of an image are scored. The scores calculated for the zoomed croppings are indicative of a zoomed cropping's inclusion of content that is captured in the image. For example, the scores are indicative of a degree to which a zoomed cropping includes salient content of the image, a degree to which the salient content included in the zoomed cropping is centered in the image, and a degree to which the zoomed cropping preserves specified regions-to-keep and excludes specified regions-to-remove. Based on the scores, at least one zoomed cropping may be chosen to effectuate a zooming of the image. Accordingly, the image may be zoomed according to the zoomed cropping such that an amount the image is zoomed corresponds to a scale of the zoomed cropping.
Abstract:
In techniques for image foreground detection, a foreground detection module is implemented to generate varying levels of saliency thresholds from a saliency map of an image that includes foreground regions. The saliency thresholds can be generated based on an adaptive thresholding technique applied to the saliency map of the image and/or based on multi-level segmentation of the saliency map. The foreground detection module applies one or more constraints that distinguish the foreground regions in the image, and detects the foreground regions of the image based on the saliency thresholds and the constraints. Additionally, different ones of the constraints can be applied to detect different ones of the foreground regions, as well as to detect multi-level foreground regions based on the saliency thresholds and the constraints.
Abstract:
Techniques for detecting and recognizing text may be provided. For example, an image may be analyzed to detect and recognize text therein. The analysis may involve detecting text components in the image. For example, multiple color spaces and multiple-stage filtering may be applied to detect the text components. Further, the analysis may involve extracting text lines based on the text components. For example, global information about the text components can be analyzed to generate best-fitting text lines. The analysis may also involve pruning and splitting the text lines to generate bounding boxes around groups of text components. Text recognition may be applied to the bounding boxes to recognize text therein.
Abstract:
In techniques for adaptive denoising with internal and external patches, example image patches taken from example images are grouped into partitions of similar patches, and a partition center patch is determined for each of the partitions. An image denoising technique is applied to image patches of a noisy image to generate modified image patches, and a closest partition center patch to each of the modified image patches is determined. The image patches of the noisy image are then classified as either a common patch or a complex patch of the noisy image, where an image patch is classified based on a distance between the corresponding modified image patch and the closest partition center patch. A denoising operator can be applied to an image patch based on the classification, such as applying respective denoising operators to denoise the image patches that are classified as the common patches of the noisy image.
Abstract:
Systems and methods are discussed to localize facial landmarks using a test facial image and a set of training images. The landmarks can be localized on a test facial image using training facial images. A plurality of candidate landmark locations on the test facial image can be determined. A subset of the training facial images with facial features similar to the facial features in the test facial image can be identified. A plurality of shape constraints can be determined for each test facial image in the subset of test facial images. These shape constraints graphically relate to one landmark location from a linear combination of the other landmark locations in the test facial image. Shape constraints can be determined for every landmark within each test facial image. A candidate landmark can be chosen from the plurality of candidate landmarks using the shape constraints.
Abstract:
Systems and methods are provided for providing learned, piece-wise patch regression for image enhancement. In one embodiment, an image manipulation application generates training patch pairs that include training input patches and training output patches. Each training patch pair includes a respective training input patch from a training input image and a respective training output patch from a training output image. The training input image and the training output image include at least some of the same image content. The image manipulation application determines patch-pair functions from at least some of the training patch pairs. Each patch-pair function corresponds to a modification to a respective training input patch to generate a respective training output patch. The image manipulation application receives an input image generates an output image from the input image by applying at least some of the patch-pair functions based on at least some input patches of the input image.
Abstract:
One exemplary embodiment involves receiving a test image generating, by a plurality of maps for the test image based on a plurality of object images. Each of the object images comprises an object of a same object type, e.g., each comprising a different face. Each of the plurality of maps is generated to provide information about the similarity of at least a portion of a respective object image to each of a plurality of portions of the test image. The exemplary embodiment further comprises detecting a test image object within the test image based at least in part on the plurality of maps.
Abstract:
Image classification techniques using images with separate grayscale and color channels are described. In one or more implementations, an image classification network includes grayscale filters and color filters which are separate from the grayscale filters. The grayscale filters are configured to extract grayscale features from a grayscale channel of an image, and the color filters are configured to extract color features from a color channel of the image. The extracted grayscale features and color features are used to identify an object in the image, and the image is classified based on the identified object.