Abstract:
A liveness test method and apparatus is disclosed. A processor implemented liveness test method includes extracting an interest region of an object from a portion of the object in an input image, performing a liveness test on the object using a neural network model-based liveness test model, the liveness test model using image information of the interest region as provided first input image information to the liveness test model and determining liveness based at least on extracted texture information from the information of the interest region by the liveness test model, and indicating a result of the liveness test.
Abstract:
Disclosed is a face verification method and apparatus. The method including analyzing a current frame of a verification image, determining a current frame state score of the verification image indicating whether the current frame is in a state predetermined as being appropriate for verification, determining whether the current frame state score satisfies a predetermined validity condition, and selectively, based on a result of the determining of whether the current frame state score satisfies the predetermined validity condition, extracting a feature from the current frame and performing verification by comparing a determined similarity between the extracted feature and a registered feature to a set verification threshold.
Abstract:
A training method of training an illumination compensation model includes extracting, from a training image, an albedo image of a face area, a surface normal image of the face area, and an illumination feature, the extracting being based on an illumination compensation model; generating an illumination restoration image based on the albedo image, the surface normal image, and the illumination feature; and training the illumination compensation model based on the training image and the illumination restoration image.
Abstract:
Face recognition of a face, to determine whether the face correlates with an enrolled face, may include generating a personalized three-dimensional (3D) face model based on a two-dimensional (2D) input image of the face, acquiring 3D shape information and a normalized 2D input image of the face based on the personalized 3D face model, generating feature information based on the 3D shape information and pixel color values of the normalized 2D input image, and comparing the feature information with feature information associated with the enrolled face. The feature information may include first and second feature information generated based on applying first and second deep neural network models to the pixel color values of the normalized 2D input image and the 3D shape information, respectively. The personalized 3D face model may be generated based on transforming a generic 3D face model based on landmarks detected in the 2D input image.
Abstract:
A method of recognizing a feature of an image may include receiving an input image including an object; extracting first feature information using a first layer of a neural network, the first feature information indicating a first feature corresponding to the input image among a plurality of first features; extracting second feature information using a second layer of the neural network, the second feature information indicating a second feature among a plurality of second features, the indicated second feature corresponding to the first feature information; and recognizing an element corresponding to the object based on the first feature information and the second feature information.
Abstract:
A method of generating a three-dimensional (3D) face model includes extracting feature points of a face from input images comprising a first face image and a second face image; deforming a generic 3D face model to a personalized 3D face model based on the feature points; projecting the personalized 3D face model to each of the first face image and the second face image; and refining the personalized 3D face model based on a difference in texture patterns between the first face image to which the personalized 3D face model is projected and the second face image to which the personalized 3D face model is projected.
Abstract:
At least one example embodiment discloses a method of extracting a feature of an input image. The method includes constructing an example pyramid including at least one hierarchical level based on stored example images, generating a codebook in each of the at least one hierarchical level, calculating a similarity between the codebook and the input image, and extracting a feature of the input image based on the similarity.
Abstract:
A method and apparatus for processing an image is disclosed, wherein the apparatus for processing the image may set blocks in an input image, perform an orthogonal transform on pixel values in the blocks, obtain orthogonal transform coefficients, and generate a resulting image by normalizing the obtained orthogonal transform coefficients.
Abstract:
A processor-implemented method with a neural network includes: generating a first intermediate vector by applying a first activation function to first nodes in a first intermediate layer adjacent to an input layer among intermediate layers of the neural network; transferring the first intermediate vector to second nodes in a second intermediate layer adjacent to an output layer among the intermediate layers; generating a second intermediate vector by applying a second activation function to the second nodes; and applying the second intermediate vector to an output layer of the neural network, wherein the second activation function is determined by a first hyperparameter of which a multiplier of the second activation function is associated with an ascending slope of the second activation function and a second hyperparameter of which the multiplier is associated with a descending slope of the second activation function to fix a peak value of the second activation function.
Abstract:
Disclosed is a convolutional neural network (CNN) processing apparatus and method, the apparatus configured to determine a loading space unit for at least one loading space in an input based on a height or a width for an input feature map of the input and an extent of a dimension of a kernel feature map, load target input elements corresponding to a target loading space, among the at least one loading space, from a memory and store the target input elements in an allocated input buffer having a size corresponding to the loading space unit, and perform a convolution operation between the target input elements stored in the input buffer and at least one kernel element of a kernel.