Abstract:
A game controller includes a plurality of LEDs formed on the rear of a case. The plurality of LEDs are arranged two-dimensionally in its layout area. The game controller has a plurality of PWM control units which are provided inside the case and control the lighting of the plurality of LEDs, respectively. The PWM control units control the lighting of the LEDs based on a control signal from a game apparatus. The game apparatus acquires a captured image of the game controller, and acquires the position of the game controller in the captured image based on the positions of the LEDs in the captured image.
Abstract:
An image acquisition section of an information processor acquires stereo images from an imaging device. An input information acquisition section acquires an instruction input from a user. A depth image acquisition portion of a position information generation section generates a depth image representing a position distribution of subjects existing in the field of view of the imaging device in the depth direction using stereo images. A matching portion adjusts the size of a reference template image in accordance with the position of each of the subjects in the depth direction represented by the depth image first, then performs template matching on the depth image, thus identifying the position of a target having a given shape and size in the three-dimensional space. An output information generation section generates output information by performing necessary processes based on the target position.
Abstract:
A frame sequence of moving picture data is divided into a tile image sequence 250, and the color space of the tile image sequence 250 is converted to generate a YCbCr image sequence 252 (S10). Each frame is reduced to ½ time in the vertical and horizontal directions (S12), and a compression process is carried out to generate compression data 260 of a reference image (S14). The compression data 260 of the reference image is decoded and decompressed similarly as upon image display to restore a YCbCr image as the reference image, and a difference image sequence 262 is generated from the reference image and the original YCbCr image 252 (S16). Then, compression data 266 of a difference image is generated (S18), and compression data 268 obtained by connecting the compression data 260 of the reference image and the compression data 266 of the difference image is generated for every four frames of a tile image (S20).
Abstract:
An information processor includes a detection plane definition portion defining a detection plane in a 3D space of a camera coordinate system of a first camera and calculates vertex coordinates of a detection area by projecting the detection plane onto a plane of a left image shot by the first camera. A feature quantity calculation portion generates feature point image of left and right images. A parallax correction area derivation portion derives an area as a parallax correction area. The parallax correction area is obtained by moving, to a left, an area of a right image identical to the detection area of the left image by as much as the parallax appropriate to a position of the detection plane in a depth direction. A matching portion performs block matching for the feature point images of each area, thus deriving a highly rated feature point. A position information output portion generates information to be used by an output information generation section based on the matching result and output that information.
Abstract:
Frames of a moving image are configured as a hierarchical structure where each frame is represented with a plurality of resolutions. Some layers are set as original image layers, and the other layers are set as difference image layers in hierarchical data representing a frame at each time step. In the case that an area is to be displayed in the resolution of the difference image layer, to respective pixel values of a difference image of the area, respective pixel values of an image of a corresponding area retained by the original image layer of lower resolution, the image enlarged to the resolution of the difference image layer, are added. A layer to be set as a difference image layer is switched to another layer as time passes.
Abstract:
An imaging device includes a first camera and a second camera and shoots the same object under different shooting conditions. A shot-image data acquirer of an image analyzer acquires data of two images simultaneously shot from the imaging device. A correcting section aligns the distributions of the luminance value between the two images by carrying out correction for either one of the two images. A correction table managing section switches and generates a correction table to be used according to the function implemented by the information processing device. A correction table storage stores the correction table showing the correspondence relationship between the luminance values before and after correction. A depth image generator performs stereo matching by using the two images and generates a depth image.
Abstract:
An image storage section 48 stores shot image data with a plurality of resolutions transmitted from an imaging device. Depth images 152 with a plurality of resolutions are generated using stereo images with a plurality of resolution levels from the shot image data (S10). Next, template matching is performed using a reference template image 154 that represents a desired shape and size, thus extracting a candidate area for a target picture having the shape and size for each distance range associated with one of the resolutions (S12). A more detailed analysis is performed on the extracted candidate areas using the shot image stored in the image storage section 48 (S14). In some cases, a further image analysis is performed based on the analysis result using a shot image with a higher resolution level (S16a and S16b).
Abstract:
A viewpoint detection unit detects a user viewing a stereoscopic image, including a parallax image of a subject as viewed from a predetermined position defined as a reference view position, and tracks a viewpoint of the detected user. A motion parallax correction unit determines, if a speed of movement of the viewpoint becomes equal to or higher than a predetermined level, an amount of motion parallax correction for the parallax image, on the basis of an amount of movement of the viewpoint, so as to generate a stereoscopic image corrected for motion parallax, generates, if the speed of movement of the viewpoint subsequently becomes lower than a predetermined level, a stereoscopic image by changing the amount of motion parallax correction in steps until the parallax image return to parallax images as seen from the reference view position.
Abstract:
A tile image sequence 250 obtained by dividing a frame into a predetermined size is further divided into another predetermined size on an image plane to generate a voxel (for example, a voxel 252) (S10). If a redundancy in a space direction or a time direction exists, then data is reduced in the direction (S12), and sequences in the time direction are deployed on a two-dimensional plane (S14). Voxel images are placed on an image plane of a predetermined size to generate one integrated image 258 (S16). In a grouping pattern which exhibits a minimum quantization error, pixels are collectively placed in the region of each voxel image for each group (integrated image 262) (S18). The integrated image 262 after the re-placement is compressed in accordance with a predetermined compression method to generate a compressed image 266 and reference information 264 for determining the position of a needed pixel (S20).
Abstract:
A hard disk drive stores hierarchical image data, a speed map holding, for each tile image, an index of the processing time required to render a tile image having a predetermined image size obtained by partitioning the image, and scenario data which defines viewpoint shifting. In a control unit of an information processing apparatus having a function of displaying an image, an input information acquisition unit acquires information with respect to the user's input operation via an input device. A loading unit loads necessary data for image displaying from the hard disk drive. A shifting condition adjustment unit adjusts the viewpoint shifting speed based upon the speed map. A frame coordinate determination unit sequentially determines frame coordinates of a display area. A decoding unit decodes compressed image data. A display image processing unit renders a display image.