Abstract:
An image encoding method using a Binary Partition Tree (BPT) includes performing the BPT on a reference frame, detecting blocks, each having a difference in a pixel value exceeding a threshold value in a current frame, based on a result of the BPT of the reference frame, and performing the BPT of the current frame on the detected blocks. In accordance with the present invention, block partition is not applied to all frames, but a partial partition method based on a difference between the pixel values of a reference frame and a current frame to be encoded is provided. Accordingly, the encoding speed within the P frame or the B frame can be improved. Furthermore, the PSNR of a corresponding frame can be maintained within a specific range of the PSNR of a reference frame, and a compression effect can be improved.
Abstract:
The present disclosure provides a method and a device for training a neural network model for use in analyzing captured images, and an intelligent image capturing apparatus employing the same. The neural network model can be trained by performing the image reconstruction and the image classification using based on image data received from a plurality of image capturing devices installed in the monitoring area, calculating at least one loss function based on data processed by the neural network model or the neural network model training device, and determining parameters minimizing the loss function. In addition, the neural network model can be updated through the re-training taking into account the newly acquired image data. Accordingly, the image analysis neural network model can operate with high precision and accuracy.
Abstract:
A method and an apparatus for authoring a machine learning-based immersive media are provided. The apparatus determines an immersive effect type of an original image of image contents to be converted into an immersive media by using an immersive effect classifier learned using an existing immersive media that the immersive effect is already added to an image, detects an immersive effect section of the original image based on the immersive effect type determination result, and generates metadata of the detected immersive effect section.
Abstract:
A method for generating a super resolution image may comprise up-scaling an input low resolution image; determining a directivity for each patch included in the up-scaled image; selecting an orientation-specified neural network or an orientation-non-specified neural network according to the directivity of the patch; applying the selected neural network to the patch; and obtaining a super resolution image by combining one or more patches output from the orientation-specified neural network and the orientation-non-specified neural network.
Abstract:
A method for removing compressed Poisson noises in an image, based on deep neural networks, may comprise generating a plurality of block-aggregation images by performing block transform on low-frequency components of an input image; obtaining a plurality of restored block-aggregation images by inputting the plurality of block-aggregation images into a first deep neural network; generating a low-band output image from which noises for the low-frequency components are removed by performing inverse block transform on the plurality of restored block-aggregation images; and generating an output image from which compressed Poisson noises are removed by adding the low-band output image to a high-band output image from which noises for high-frequency components of the input image are removed.
Abstract:
The present invention relates to an image encoding method using an adaptive preprocessing scheme, including loading an input image for each frame, determining an encoding type of each of the frames, determining the size of a block to be encoded in each frame according to the determined encoding type, determining blocks that can be replicated from the blocks having the determined size and performing an intra-picture replication preprocessing or inter-picture replication preprocessing procedure on the determined blocks according to the encoding types of the frames, and encoding the frames on which the preprocessing procedure has been performed.
Abstract:
Provided is a sensory effect adaptation method performed by an adaptation engine, the method including identifying first metadata associated with an object in a virtual world and used to describe the object and converting the identified first metadata into second metadata to be applied to a sensory device in a real world, wherein the second metadata is obtained by converting the first metadata based on a scene determined by a gaze of a user in the virtual world.
Abstract:
Disclosed is a sensory information providing apparatus. The sensory information providing apparatus may comprise a learning model database storing a plurality of learning models related to sensory effect information with respect to a plurality of videos; and a video analysis engine generating the plurality of learning models by extracting sensory effect association information by analyzing the plurality of videos and sensory effect meta information of the plurality of videos, and extracting sensory information corresponding to an input video stream by analyzing the input video stream based on the plurality of learning model.
Abstract:
Disclosed are a method and an apparatus for generating metadata of immersive media and disclosed also are an apparatus and a method for transmitting metadata related information. The apparatus includes: at least one of a camera module photographing or capturing the image; a gyro module sensing horizontality; a global positioning sensor (GPS) module calculating a position by receiving a satellite signal; and an audio module recording audio; and a network module receiving sensor effect information from a sensor aggregator through a wireless communication network; and an application generating metadata by performing timer-synchronization of an image photographed based on the camera module, a sensor effect collected by using the gyro module or the GPS module, or audio collected based on the audio module.