Abstract:
A method and apparatus for generating a virtual object are provided, the method includes acquiring a point cloud of an object to generate a virtual object, determining shape attribute information of the object based on an image of the object, changing a position of at least one point in the point cloud based on the shape attribute information, and generating a virtual object for the object based on a changed point cloud.
Abstract:
Provided are an electronic device and a method of controlling an electronic device. The electronic device includes: a memory storing at least one instruction; and at least one processor configured to execute the at least one instruction, wherein one or more of the at least one processor is configured to: acquire a first vector corresponding to each of a plurality of sections of a voice signal by inputting the voice signal to a common encoder based on acquiring the voice signal; acquire a second vector corresponding to each of the plurality of sections and independent on a context of the voice signal by inputting the first vector into a first individual encoder; acquire a phoneme sequence corresponding to the second vector by inputting the second vector into a first decoder; acquire a third vector corresponding to at least two sections among the plurality of sections and dependent on the context of the voice signal by inputting the first vectors into a second individual encoder; acquire a sub-word sequence corresponding to the third vector by inputting the third vector into a second decoder; and acquire text information corresponding to the plurality of sections by correcting the sub-word sequence based on the phoneme sequence, through a text information acquisition module.
Abstract:
An electronic device is provided. The electronic device includes a camera including a plurality of lenses, a display, and a processor, in which the processor is configured to display a plurality of icons corresponding to the plurality of lenses, based on first position information in a first photographing mode, and upon selection of a first icon by a first gesture from the plurality of icons in the first photographing mode, switch to a second photographing mode and display a zoom control region including a plurality of zoom levels having a first zoom level of a first lens corresponding to the first icon as a reference zoom level and the plurality of icons rearranged based on second position information corresponding to the plurality of zoom levels.
Abstract:
A processor-implemented method with three-dimensional (3D) modeling includes: determining a type of movement of an object detected in an image received from a sensor, based on a variability of a position of the object and a variability of a shape of the object; segmenting the object into one or more chunks each corresponding to a unit of movement, based on the determined type of movement; correcting the determined type of movement based on a change in position of one or more feature points of the one or more chunks of the object in the image; determining a state of the movement of the object based on the corrected type of the movement of the object; correcting a pose of the sensor based on a state of the object; and updating a pose for each of the one or more chunks of the object based on the determined state of the object and the corrected pose of the sensor.
Abstract:
A processor-implement method with image processing includes: generating a feature map of a first image and detecting a target region in the first image based on the feature map; correcting the detected target region; and processing an object corresponding to the target region, based on the corrected target region.
Abstract:
A method, apparatus, electronic device, and non-transitory computer-readable storage medium with multi-modal feature fusion are provided. The method includes generating three-dimensional (3D) feature information and two-dimensional (2D) feature information based on a color image and a depth image, generating fused feature information by fusing the 3D feature information and the 2D feature information based on an attention mechanism, and generating predicted image information by performing image processing based on the fused feature information.
Abstract:
A processor-implemented method with object pose estimation includes: determining an image feature corresponding to a point cloud of an input image; determining semantic segmentation information, instance mask information, and keypoint information of an object, based on the image feature; and estimating a pose of the object based on the semantic segmentation information, the instance mask information, and the keypoint information.
Abstract:
A stereo matching method includes extracting feature points of a first image and feature points of a second image, the first image and the second image together constituting a stereo image, determining reference points by matching the feature points of the second image to the feature points of the first image, classifying the reference points, and performing stereo matching on pixels of which disparities are not determined in the first image and the second image based on disparities of the reference points in the pixels determined based on a result of the classifying.
Abstract:
Provided is a stereo matching method and apparatus. A stereo matching method includes acquiring detection information based on an edge detection of a first image, determining an edge line and a feature point based on the detection information, sampling a reference pixel from the edge line and the feature point, determining a polygon with the reference pixel as a vertex, predicting a search range comprising a disparity of a pixel within the polygon based on a disparity of the reference pixel, and performing stereo matching of the pixel based on the search range.
Abstract:
Provided is a method and apparatus for modeling and restoring a target object, the method including generating a key structure of the target object based on deformation information of the target object extracted from a pre-modeling result based on a shape and a material property of the target object, and calculating a virtual material property corresponding to the key structure based on the virtual material property and the key structure of the target object.