摘要:
The invention relates to a method for generating a motion field between a current frame and a reference frame belonging to a video sequence from an input set of motion fields. An motion field is associated to an ordered pair of frames comprises for a group of pixels belonging to a first frame of the ordered pair of frames, a motion vector computed from a location of the pixel in the first frame to an endpoint in a second frame of the ordered pair of frames. The method comprises a step for determining a plurality of motion paths from a current frame to a reference frame wherein a motion path comprises a sequence of N ordered pairs of frames associated to the input set of motion fields and wherein a first frame of an ordered pair corresponds to a second frame of the previous ordered pair in the sequence; the first image of the first ordered pair is the current frame; the second frame of the last ordered pair is the reference frame; and N is an integer. The method then comprises a step for determining, for the group of pixels belonging to the current frame, a plurality of candidate motion vectors from the current frame to the reference frame wherein a candidate motion vector is the result of a sum of motion vectors; each motion vector belonging to a motion field associated to an ordered pair of frames according to a determined motion path. And the method then comprises a step for selecting, for the group of pixels belonging to the current frame, a candidate motion vector among the plurality of candidate motion vectors.
摘要:
The invention relates to a method of filtering a map of disparity associated with one of the views of a stereoscopic image comprising the steps of identifying the pixels (x n ) with which is associated an erroneous disparity value (d(x n )) by comparison with the disparity values of the neighbouring pixels (x n-1 ,x n+1 ) and of correcting the disparity values of the identified pixels.
摘要:
In video encoding, the video frames are spatio-temporally filtered for reduction of spatial and temporal redundancy before they are entropy encoded. Known filtering schemes consider temporally successive frames and are static. It is probable but not necessary that successive frames are most efficient to encode. Therefore, a plurality or all possible frame order permutations are considered for a group of frames (GOP) and evaluated based on a global criterion, which is the sum of local criterion values computed over disjoint sets of successive frames considered in the new order. The local criterion value is deduced from motion estimation processed on each considered set of frames. The best ordering is chosen as the one that minimizes the global criterion value.
摘要:
An augmented reality (AR) interactive system and method is provided. In one embodiment the systems comprises a head mounted user interface configured to receive user input, a processor configured to manage data based on user input, a camera and a display. The camera and the display are in processing communication with one another and the head mounted user interface via the processor. The processor is configured to determine a user's field of view and a center of the user's field of view based on output of a sensor and rendering images for output to the display. Each of the images include one or more objects and a plurality of signs each corresponding to a selectable object in the user's field of view. The rendering of images include altering a first display attribute of a given sign of the plurality of displayed signs based on determining that the user's field of view is centered on the given sign. The processor is also configured to select the given sign based on determining that the user's field of view remains centered on the given sign for an amount of time exceeding a threshold.
摘要:
A 3D immersive method for a user in a virtual 3D scene includes: for at least one event in said virtual 3D scene, obtaining (21) metadata linking said event and at least one place of said event in said virtual 3D scene; determining (22) a position and an orientation of said user in said virtual 3D scene; executing (24) in said virtual 3D scene, for said user, at least one action regarding a given event and a given place linked by said metadata, in function of said position and orientation of said user with respect to said given place, and of a relationship between said given event and a user profile of said user; and outputting (25) a result of said at least one action to the attention of said user.
摘要:
The present invention relates to a method for estimating cast shadow regions and highlight regions in a scene, said scene being represented by a low dynamic range image, called LDR image, and a high dynamic range, called HDR image, said method comprising the steps of: - computing (S1) a detail map, called LDR detail map, of the LDR image, said LDR detail map associating a detail level to each pixel of the LDR image, - computing (S2) a detail map, called HDR detail map, of the HDR image, said HDR detail map associating a detail level to each pixel of the HDR image, and - detecting (S3) regions of the scene where the detail level in the HDR detail map is greater than the detail level in the LDR detail map, said detected regions corresponding to cast shadow regions or highlights region of the scene.
摘要:
In order to blur a virtual object in a video in real time as the video is acquired by a device capturing a real scene, a salient idea comprises estimating an apparent motion vector between two successive images, being captured at two successive device poses, wherein the apparent motion vector estimation is based on a motion of the device. The successive images are then filtered based on the estimated apparent motion vector.
摘要:
A method and device for rendering a scene comprising one or more real objects. A virtual object is associated with each real object, a virtual object associated with a real object corresponding to a virtual replica of this real object. The virtual replica is used to render a transformation that may be applied to the real object when for example hit by a virtual object, the virtual object then replacing the corresponding real object within the scene. To bring realism to the scene, texture information obtained from image(s) of the real object(s) is used to texture the visible part(s) of the transformed virtual object(s). The texture information is selected in the images by using information on the visibility of the parts of the real object(s) that correspond to the visible parts of the transformed virtual object(s).