Abstract:
Wide dynamic range depth imaging in a structured light device is provided that improves depth maps for scenes with a wide range of albedo values under varying light conditions. A structured light pattern, e.g., a time-multiplexed structured light pattern, is projected into a scene at various projection times and a camera captures images of the scene for at least the same exposure times as the projection times. A depth image is computed for each of the projection/exposure times and the resulting depth images are combined to generate a composite depth image.
Abstract:
Disclosed examples include three-dimensional imaging systems and methods to reconstruct a three-dimensional scene from first and second image data sets obtained from a single camera at first and second times, including computing feature point correspondences between the image data sets, computing an essential matrix that characterizes relative positions of the camera at the first and second times, computing pairs of first and second projective transforms that individually correspond to regions of interest that exclude an epipole of the captured scene, as well as computing first and second rectified image data sets in which the feature point correspondences are aligned on a spatial axis by respectively applying the corresponding first and second projective transforms to corresponding portions of the first and second image data sets, and computing disparity values of a stereo disparity map according to the rectified image data sets to construct.
Abstract:
A method, apparatus and a system multi-camera image processing method. The method includes performing geometric alignment to produce a geometric output, performing photometric alignment to produce a photometric output and blending output, using data from the geometric alignment and the photometric alignment for performing synthesis function for at least one of blending and stitching images from the multi-cameras, and displaying an image from the synthesis function.
Abstract:
For reducing a delay from panning a camera system, an estimate is received of a physical movement of the camera system. In response to the estimate, a determination is made of whether the camera system is being panned. In response to determining that the camera system is not being panned, most effects of the physical movement are counteracted in a video sequence from the camera system. In response to determining that the camera system is being panned, most effects of the panning are preserved in the video sequence, while concurrently the video sequence is shifted toward a position that balances flexibility in counteracting effects of a subsequent physical movement of the camera system.
Abstract:
A first depth map is generated in response to a stereoscopic image from a camera. The first depth map includes first pixels having valid depths and second pixels having invalid depths. In response to the first depth map, a second depth map is generated for replacing at least some of the second pixels with respective third pixels having valid depths. For generating the second depth map, a particular one of the third pixels is generated for replacing a particular one of the second pixels. For generating the particular third pixel, respective weight(s) is/are assigned to a selected one or more of the first pixels in response to value similarity and spatial proximity between the selected first pixel(s) and the particular second pixel. The particular third pixel is computed in response to the selected first pixel(s) and the weight(s).
Abstract:
A method, apparatus and a system multi-camera image processing method. The method includes performing geometric alignment to produce a geometric output by estimating fish eye distortion correction parameters, performing initial perspective correction on related frame, running corner detection in the overlapping areas, locating the stronger corner, calculating BRIEF descriptors for features and match feature point from two cameras using BRIEF scores, performing checks and rejecting wrong feature matches, finding perspective matrices to minimize distance between matched features; and creating a geometric lookup table.
Abstract:
A method of local tone mapping of a high dynamic range (HDR) image is provided that includes dividing a luminance image of the HDR image into overlapping blocks and computing a local tone curve for each block, computing a tone mapped value for each pixel of the luminance image as a weighted sum of values computed by applying local tone curves of neighboring blocks to the pixel value, computing a gain for each pixel as a ratio of the tone mapped value to the value of the pixel, and applying the gains to corresponding pixels in the HDR image. A weight for each value is computed based on distance from the pixel to the center point of the block having the local tone curve applied to compute the value and the intensity difference between the value of the pixel and the block mean pixel value.
Abstract:
A method of noise filtering of a digital video sequence is provided that includes computing a motion image for a frame, wherein the motion image includes a motion value for each pixel in the frame, and wherein the motion values are computed as differences between pixel values in a luminance component of the frame and corresponding pixel values in a luminance component of a reference frame, applying a first spatial noise filter to the motion image to obtain a final motion image, computing a blending factor image for the frame, wherein the blending factor image includes a blending factor for each pixel in the frame, and wherein the blending factors are computed based on corresponding motion values in the final motion image, generating a filtered frame, wherein the blending factors are applied to corresponding pixel values in the reference frame and the frame, and outputting the filtered frame.