Abstract:
The model training apparatus trains an image conversion model to generate, from an input image representing a scene in a first environment, an output image representing the scene in a second environment. The model training apparatus inputs a training image to the image conversion model to obtain a first feature map and an output image, input the output image to the image conversion model to obtain a second feature map, computes a patch-wise loss using the features corresponding to a positive example patch and a negative example patch extracted from the training image and a positive example patch extracted from the output image, and trains the image conversion model based on the patch-wise loss, which is extracted intensively from the region representing an object of a specific type.
Abstract:
In the object detection device, the plurality of object detection units output a score indicating a probability that a predetermined object exists, for each partial region set with respect to image data inputted. The weight computation unit computes a weight for each of the plurality of object detection units by using weight computation parameters and based on the image data. The weights are used when the scores outputted by the plurality of object detection units are merged. The weight redistribution unit changes the weight for a predetermined object detection unit, among the weights computed by the weight computation unit, to 0 and output the weights. The merging unit merges the scores outputted by the plurality of object detection units for each of the partial regions, by using the weights computed by the weight computation unit and including the weight changed by the weight redistribution unit. The loss computation unit computes a difference between a ground truth label of the image data and the merged score merged by the merging unit as a loss. Then, the parameter correction unit corrects the weight computation parameters so as to reduce the loss.
Abstract:
Provided is an object detection device for efficiently and simply selecting an image for creating instructor data on the basis of the number of detected objects. The object detection device is provided with: a detection unit for detecting an object from each of a plurality of input images using a dictionary; an acceptance unit for displaying, on a display device, a graph indicating the relationship between the input images and the number of subregions in which the objects are detected, and displaying, on the display device, in order to create instructor data, one input image among the plurality of input images in accordance with a position on the graph accepted by operation of an input device; a generation unit for generating the instructor data from the input image; and a learning unit for learning a dictionary from the instructor data.
Abstract:
To accurately detect theft of a product. An information processing apparatus (2000) acquires a reference image (13) in which a product exhibition location is imaged. In addition, the information processing apparatus (2000) acquires a comparison image (14), in which a product shelf is imaged at a time after the reference image (13) is imaged. Furthermore, the information processing apparatus (2000) computes the quantity of reduction in products inside a surveillance area (15) included in both the reference image (13) and the comparison image (14). Furthermore, the information processing apparatus (2000) outputs the warning in a case where the computed quantity of reduction is equal to or larger than the reference value.
Abstract:
To provide a POS terminal or the like which enables adjustment of white balance of an object to be sufficiently accurate even in the change of external light illuminating the object. The POS terminal includes: an image capture unit which captures an object including a plurality of respectively different reference colors and generates an image; a comparison unit which compares a portion of the image relating to each of the reference colors with a standard image which is a standard of the reference colors; and an adjustment unit which adjusts white balance based on a result of the comparison.
Abstract:
A POS terminal device capable of improving a recognition rate in a process for recognizing a commodity irrespective of the surrounding environment is provided. A POS terminal device (1) includes a brightness measurement unit (2), an irradiation unit (4), an image pickup unit (6), and a recognition process unit (8). The brightness measurement unit (2) measures the brightness of environmental light around the POS terminal device (1). The irradiation unit (4) irradiates a commodity with light, the light being adjusted according to the brightness of the environmental light measured by the brightness measurement unit (2). The image pickup unit (6) shoots the commodity irradiated with the light by the irradiation unit (4) and thereby generates an image thereof. The recognition process unit (8) performs a process for recognizing the commodity based on the image generated by the shooting performed by the image-pickup unit (6).
Abstract:
In an object detection device, a plurality of object detection units output a score indicating probability that a predetermined object exists, for each partial region set to image data inputted. The weight computation unit computes weights for merging the scores outputted by the plurality of object detection units, using weight calculation parameters, based on the image data. The merging unit merges the scores outputted by the plurality of object detection units, for each partial region, with the weights computed by the weight computation unit. The target model object detection unit configured to output a score indicating probability that the predetermined object exists, for each partial region set to the image data. The first loss computation unit computes a first loss indicating a difference of the score of the target model object detection unit from a ground truth label of the image data and the score merged by the merging unit. The first parameter correction unit corrects parameters of the target model object detection unit to reduce the first loss.
Abstract:
To improve pose estimation accuracy, a pose estimation apparatus according to the present invention extracts a person area from an image, and generates person area image information, based on an image of the extracted person area. The pose estimation apparatus according to the present invention further extracts a joint point of a person from the image, and generates joint point information, based on the extracted joint point. Then, the pose estimation apparatus according to the present invention generates feature value information, based on both of the person area image information and the joint point information. Then, the pose estimation apparatus according to the present invention estimates a pose of a person included in the image, based on an estimation model in which the feature value information is an input, and pose estimation result is an output.
Abstract:
Concerning a partial area image that constitutes a wide area image, to control a flying body in accordance with a flight altitude at a past point of time of image capturing, an information processing apparatus includes a wide area image generator that extracts, from a flying body video obtained when a flying body captures a ground area spreading below while moving, a plurality of video frame images and combines the video frame images, thereby generating a captured image in a wide area, an image capturing altitude acquirer that acquires a flight altitude at a point of time of image capturing by the flying body for each of the plurality of video frame images, and an image capturing altitude output unit that outputs a difference of the flight altitude for each video frame image.
Abstract:
Learning means 701 learns a model for identifying an object indicated by data by using training data. First identification means 702 identifies the object indicated by the data by using the model learned by the learning means 701. Second identification means 703 identifies the object indicated by the data as an identification target used by the first identification means 702 by using a model different from the model learned by the learning means 701. The learning means 701 re-learns the model by using the training data including the label for the data determined based on the identification result derived by the second identification means 703 and the data.