Abstract:
Techniques for enhancing an image using pixel-specific processing are disclosed. An image can be enhanced by updating certain pixels through patch aggregation. Neighboring pixels of a selected pixel are identified. Respective patch values for patches containing the selected pixel are determined. Patch values provide update information for updating the respective pixels in the patch. Relevant patch values for the selected pixel are identified by identifying associated patches of the pixel. Information from the relevant patch values of the selected pixel may be obtained. Using this information, pixel-specific processing may be performed to determine an updated pixel value for the selected pixel or for neighboring pixels of the selected pixel. Pixel-specific processes may be executed for each of the selected or neighboring pixels. These pixel-specific processes can be executed in parallel. Therefore, through the execution of pixel-specific processes, which may be performed concurrently, an enhanced image may be determined.
Abstract:
Systems and methods are disclosed herein for using one or more computing devices to automatically segment an object in an image by referencing a dataset of already-segmented images. The technique generally involves identifying a patch of an already-segmented image in the dataset based on the patch of the already-segmented image being similar to an area of the image including a patch of the image. The technique further involves identifying a mask of the patch of the already-segmented image, the mask representing a segmentation in the already-segmented image. The technique also involves segmenting the object in the image based on at least a portion of the mask of the patch of the already-segmented image.
Abstract:
Systems and methods are disclosed herein for using one or more computing devices to automatically segment an object in an image by referencing a dataset of already-segmented images. The technique generally involves identifying a patch of an already-segmented image in the dataset based on the patch of the already-segmented image being similar to an area of the image including a patch of the image. The technique further involves identifying a mask of the patch of the already-segmented image, the mask representing a segmentation in the already-segmented image. The technique also involves segmenting the object in the image based on at least a portion of the mask of the patch of the already-segmented image.
Abstract:
This document describes techniques and apparatuses for area-dependent image enhancement. These techniques are capable of enabling selection, through a touch-enabled mobile-device display, of an area of a photographic image through movement of a spatially-variable implement, such as brush icon moved over the image. Selected areas can be enhanced differently than other areas, such as to apply sharpening to the selected area and blurring to a non-selected area.
Abstract:
Systems and methods are provided for content-based selection of style examples used in image stylization operations. For example, training images can be used to identify example stylized images that will generate high-quality stylized images when stylizing input images having certain types of semantic content. In one example, a processing device determines which example stylized images are more suitable for use with certain types of semantic content represented by training images. In response to receiving or otherwise accessing an input image, the processing device analyzes the semantic content of the input image, matches the input image to at least one training image with similar semantic content, and selects at least one example stylized image that has been previously matched to one or more training images having that type of semantic content. The processing device modifies color or contrast information for the input image using the selected example stylized image.
Abstract:
Joint depth estimation and semantic labeling techniques usable for processing of a single image are described. In one or more implementations, global semantic and depth layouts are estimated of a scene of the image through machine learning by the one or more computing devices. Local semantic and depth layouts are also estimated for respective ones of a plurality of segments of the scene of the image through machine learning by the one or more computing devices. The estimated global semantic and depth layouts are merged with the local semantic and depth layouts by the one or more computing devices to semantically label and assign a depth value to individual pixels in the image.
Abstract:
Saliency map computation is described. In one or more implementations, a base saliency map is generated for an image of a scene. The base saliency map may be generated from intermediate saliency maps computed for boundary regions of the image. Each of the intermediate saliency maps may represent visual saliency of portions of the scene that are captured in the corresponding boundary region. The boundary regions may include, for instance, a top boundary region, a bottom boundary region, a left boundary region, and a right boundary region. Further, the intermediate saliency maps may be combined in such a way that an effect of a foreground object on the saliency map is suppressed. The foreground objects for which the effect is suppressed are those that occupy a majority of one of the boundary regions.
Abstract:
Techniques for enhancing an image using pixel-specific processing. An image can be enhanced by updating selected pixels through patch aggregation. Respective patch values for patches of any size of the image are determined. Patch values provide update information for updating the respective pixels in the patch. Relevant patch values for the selected pixel are identified by identifying associated patches of the pixel. Information from the relevant patch values of the selected pixel may be obtained by averaging the relevant patch values or determining the maximum or minimum patch value. Using this information, pixel-specific processing may be performed to determine an updated pixel value for the selected pixel. Pixel-specific processes may be executed for each of the selected pixels. These pixel-specific processes can be executed in parallel. Therefore, through the execution of pixel-specific processes, which may be performed concurrently, an enhanced image may be determined.
Abstract:
A convolutional neural network is trained to analyze input data in various different manners. The convolutional neural network includes multiple layers, one of which is a convolution layer that performs a convolution, for each of one or more filters in the convolution layer, of the filter over the input data. The convolution includes generation of an inner product based on the filter and the input data. Both the filter of the convolution layer and the input data are binarized, allowing the inner product to be computed using particular operations that are typically faster than multiplication of floating point values. The possible results for the convolution layer can optionally be pre-computed and stored in a look-up table. Thus, during operation of the convolutional neural network, rather than performing the convolution on the input data, the pre-computed result can be obtained from the look-up table
Abstract:
In techniques for video denoising using optical flow, image frames of video content include noise that corrupts the video content. A reference frame is selected, and matching patches to an image patch in the reference frame are determined from within the reference frame. A noise estimate is computed for previous and subsequent image frames relative to the reference frame. The noise estimate for an image frame is computed based on optical flow, and is usable to determine a contribution of similar motion patches to denoise the image patch in the reference frame. The similar motion patches from the previous and subsequent image frames that correspond to the image patch in the reference frame are determined based on the optical flow computations. The image patch is denoised based on an average of the matching patches from reference frame and the similar motion patches determined from the previous and subsequent image frames.