Abstract:
Techniques and apparatus for automatic upright adjustment of digital images. An automatic upright adjustment technique is described that may provide an automated approach for straightening up slanted features in an input image to improve its perceptual quality. This correction may be referred to as upright adjustment. A set of criteria based on human perception may be used in the upright adjustment. A reprojection technique that implements an optimization framework is described that yields an optimal homography for adjustment based on the criteria and adjusts the image according to new camera parameters generated by the optimization. An optimization-based camera calibration technique is described that simultaneously estimates vanishing lines and points as well as camera parameters for an image; the calibration technique may, for example, be used to generate estimates of camera parameters and vanishing points and lines that are input to the reprojection technique.
Abstract:
A computer-implemented method and apparatus are described for deblurring an image. The method may include accessing the image that has at least one blurred region and, automatically, without user input, determining a first value for a first size for a blur kernel for the at least one blurred region. Thereafter, automatically, without user input, a second value for a second size for the blur kernel is determined for the at least one blurred region. A suggested size for the blur kernel is then determined based on the first value and the second value.
Abstract:
Techniques for detecting and recognizing text may be provided. For example, an image may be analyzed to detect and recognize text therein. The analysis may involve detecting text components in the image. For example, multiple color spaces and multiple-stage filtering may be applied to detect the text components. Further, the analysis may involve extracting text lines based on the text components. For example, global information about the text components can be analyzed to generate best-fitting text lines. The analysis may also involve pruning and splitting the text lines to generate bounding boxes around groups of text components. Text recognition may be applied to the bounding boxes to recognize text therein.
Abstract:
Systems and methods are provided for providing improved de-noising image content by using directional noise filters to accurately estimate a blur kernel from a noisy blurry image. In one embodiment, an image manipulation application applies multiple directional noise filters to an input image to generate multiple filtered images. Each of the directional noise filters has a different orientation with respect to the input image. The image manipulation application determines multiple two-dimensional blur kernels from the respective filtered images. The image manipulation application generates a two- two-dimensional blur kernel for the input image from the two-dimensional blur kernels for the filtered images. The image manipulation application generates a de-blurred version of the input image by executing a de-blurring algorithm based on the two-dimensional blur kernel for the input image.
Abstract:
Methods for enhancing images with increased efficiency include using a discriminative index tree to expedite image optimization processes. The discriminative index tree indexes patch-based image priors for modifying an image by using classifiers determined by exploiting a structure of the patch-based image priors. The discriminative index tree quickly and efficiently parses a space of patch-based image patches to determine approximate dominant patch-based image priors for the space of image patches. To further improve the efficiency of the discriminative index tree, one or more embodiments can limit a number of potential patch-based image priors from which a dominant patch-based image prior is selected.
Abstract:
Techniques are disclosed for collaborative and synchronized photography across multiple digital camera devices. A panoramic photograph of a scene can be generated from separate photographs taken by each of the cameras simultaneously. During composition, the viewfinder images from each camera are collected and stitched together on the fly to create a panoramic preview image. The panoramic preview is then displayed on the camera devices as live visual guidance, which each user can use to change the orientation of the camera and thus change the composition of the panoramic photograph. In some cases, the host sends visual instructions to other camera devices to guide users in camera adjustment. When the desired composition is achieved, the host sends a trigger command to all of the cameras to take photographs simultaneously. Each of these separate photographs can then be stitched together to form a panoramic photograph.
Abstract:
A simulated tracking shot is generated from an image sequence in which a foreground feature moves relative to a background during capturing of the image sequence. The background is artificially blurred in the simulated tracking shot in a spatially-invariant manner corresponding to foreground motion relative to the background during a time span of the image sequence. The foreground feature can be substantially unblurred relative to a reference image selected from the image sequence. A system to generate the simulated tracking shot can be configured to derive spatially invariant blur kernels for a background portion by reconstructing or estimating a 3-D space of the captured scene, placing virtual cameras along a foreground trajectory in the 3-D space, and projecting 3-D background points on to the virtual cameras.
Abstract:
Techniques for detecting and recognizing text may be provided. For example, an image may be analyzed to detect and recognize text therein. The analysis may involve detecting text components in the image. For example, multiple color spaces and multiple-stage filtering may be applied to detect the text components. Further, the analysis may involve extracting text lines based on the text components. For example, global information about the text components can be analyzed to generate best-fitting text lines. The analysis may also involve pruning and splitting the text lines to generate bounding boxes around groups of text components. Text recognition may be applied to the bounding boxes to recognize text therein.
Abstract:
Multi-feature image haze removal is described. In one or more implementations, feature maps are extracted from a hazy image of a scene. The feature maps convey information about visual characteristics of the scene captured in the hazy image. Based on the feature maps, portions of light that are not scattered by the atmosphere and are captured to produce the hazy image are computed. Additionally, airlight of the hazy image is ascertained based on at least one of the feature maps. The calculated airlight represents constant light of the scene. Using the computed portions of light and the ascertained airlight, a dehazed image is generated from the hazy image.
Abstract:
Joint video deblurring and stabilization techniques are described. In one or more implementations, a deblurring and stabilization module is configured to jointly deblur and stabilize a video by grouping video frames into spatial-neighboring frame clusters, and building local mesh homographies for video frames in each spatial-neighboring frame cluster.