Abstract:
Disclosed herein are a method, apparatus, system, and computer-readable recording medium for image compression. An encoding apparatus performs preprocessing of feature map information, frame packing, frame classification, and encoding. A decoding apparatus performs decoding, frame depacking, and postprocessing in order to reconstruct feature map information. By encoding the feature map information, inter-prediction and intra-block prediction for a frame are performed. The encoding apparatus provides the decoding apparatus with a feature map information bitstream for reconstructing the feature map information along with an image information bitstream.
Abstract:
There are provided a method, apparatus, system, and computer-readable recording medium for image compression. An encoding apparatus performs domain transformation and quantization on feature map information and image information. The encoding apparatus rearranges the result of domain transformation and quantization so as to have a form advantageous to the encoding procedure and encodes the result of rearrangement, thereby generating a bitstream. A decoding apparatus receives the bitstream, decodes the received bitstream, and performs inverse transformation, dequantization, and inverse rearrangement using information transmitted through the bitstream. The result of inverse transformation, dequantization, and inverse rearrangement is used for the machine-learning task of a neural network.
Abstract:
There are provided an apparatus, method, system, and recording medium for performing selective encoding/decoding on feature information. An encoding apparatus generates residual feature information. The encoding apparatus transmits the residual feature information to a decoding apparatus through a residual feature map bitstream. The residual feature information is the difference between feature information extracted from an original image and feature information extracted from a reconstructed image. Feature information of the reconstructed image is generated using the reconstructed image. Reconstructed feature information is generated using the feature information of the reconstructed image and reconstructed residual feature information.
Abstract:
A user terminal, hands-free device and method for hands-free automatic interpretation service. The user terminal includes an interpretation environment initialization unit, an interpretation intermediation unit, and an interpretation processing unit. The interpretation environment initialization unit performs pairing with a hands-free device in response to a request from the hands-free device, and initializes an interpretation environment. The interpretation intermediation unit sends interpretation results obtained by interpreting a user's voice information received from the hands-free device to a counterpart terminal, and receives interpretation results obtained by interpreting a counterpart's voice information from the counterpart terminal. The interpretation processing unit synthesizes the interpretation results of the counterpart into a voice form based on the initialized interpretation environment when the interpretation results are received from the counterpart terminal, and sends the synthesized voice information to the hands-free device.
Abstract:
Described herein is a speech recognition device comprising: a communication module receiving speech data corresponding to speech input from a speech recognition terminal and multi-sensor data corresponding to input environment of the speech; a model selection module selecting a language and acoustic model corresponding to the multi-sensor data among a plurality of language and acoustic models classified according to the speech input environment on the basis of previous multi-sensor data; and a speech recognition module controlling the communication module to apply a feature vector extracted from the speech data to the language and acoustic model and transmit speech recognition result for the speech data to the speech recognition terminal.
Abstract:
An encoding apparatus extracts features of an image by applying multiple padding operations and multiple downscaling operations to an image represented by data and transmits feature information indicating the features to a decoding apparatus. The multiple padding operations and the multiple downscaling operations are applied to the image in an order in which one padding operation is applied and thereafter one downscaling operation corresponding to the padding operation is applied. A decoding method receives feature information from an encoding apparatus, and generates a reconstructed image by applying multiple upscaling operations and multiple trimming operations to an image represented by the feature information. The multiple upscaling operations and the multiple trimming operations are applied to the image in an order in which one upscaling operation is applied and thereafter one trimming operation corresponding to the upscaling operation is applied.
Abstract:
Disclosed herein is an encoding method. The encoding method includes extracting a feature map from an input image, determining an encoding feature map based on the extracted feature map, generating a converted feature map by performing conversion on the encoding feature map, and performing encoding on the converted feature map.
Abstract:
Disclosed herein are a method, an apparatus and a storage medium for processing a feature map. An encoding method for a feature map includes configuring a feature frame for feature maps, and generating encoded information by performing encoding on the feature frame. A decoding method for a feature map includes reconstructing a feature frame by performing decoding on encoded information, and reconstructing feature maps using the feature frame. A feature frame is configured using feature maps, and compression using a video compression codec or a deep learning-based image compression method is applied to the feature frame.
Abstract:
Disclosed herein are a method, an apparatus, and a storage medium for image encoding/decoding using filtering. The image encoding method and the image decoding method perform classification on a classification unit of a target image, and perform filtering that uses a filter on the classification unit. A classification index of the classification unit is determined, and the classification index indicates a filter to be used for filtering on the classification unit, among multiple filters. The classification index is determined by various coding parameters, such as an encoding mode of the target block and an encoding mode of a neighboring block adjacent to the target block.
Abstract:
Disclosed herein are a method, an apparatus and a storage medium for image encoding/decoding using a binary mask. An encoding method includes generating a latent vector using an input image, generating a selected latent vector component set using a binary mask, and generating a main bitstream by performing entropy encoding on the selected latent vector component set. A decoding method includes generating a selected latent vector component set including one or more selected latent vector components by performing entropy decoding on a main bitstream and generating the latent vector in which the one or more selected latent vector components are relocated by relocating the selected latent vector component set in the latent vector.