Abstract:
At least one storage stores a dictionary of a discriminator acquired by machine learning by use of a plurality of pairs of crowd state image as an image which expresses a crowd state at a predetermined size and includes a person whose reference site is expressed as large as the size of the reference site of a person defined for the predetermined size, and training label for the crowd state image. At least one processor extracts regions from a given image and recognizes states of the crowds shot in the extracted regions based on the dictionary.
Abstract:
A resolution evaluation section calculates a resolution evaluation value for each location within a monitoring target region excluding the presence area of an obstacle. A gazing point angle evaluation section calculates a gazing point angle evaluation value for each location within the monitoring target region excluding the presence area of the obstacle. The suitability calculation section calculates the suitability for each location within the monitoring target region excluding the presence area of the obstacle, on the basis of at least the resolution evaluation values and the gazing point angle evaluation values, the suitability indicating the degree to which the image of the monitoring target object placed at each location is suitable for the image recognition process. A display control section causes a display device to show regions within the monitoring target region which correspond to the suitabilities, in a mode commensurate with these suitabilities.
Abstract:
At least one processor generates a crowd state image as an image in which a person image corresponding to a person state is synthesized with previously-prepared image at a predetermined size. The at least one processor specifies a training label for the crowd state image. The at least one processor outputs a pair of crowd state image and training label.
Abstract:
An image processing device (3000) comprises an input unit (3020) and a presentation unit (3040). The input unit (3020) accepts an input of an operation for movement, on a captured image captured by a camera, of a first image which is superimposed on the captured image on the basis of a predetermined camera parameter indicating the position and attitude of the camera and which indicates a target object having a predetermined shape and a predetermined size set in a real space. The presentation unit (3040) presents the first image indicating the target object in a manner of view corresponding to a position on the captured image after the movement on the basis of the camera parameter.
Abstract:
An information processing apparatus (2000) includes a first analyzing unit (2020), a second analyzing unit (2040), and an estimating unit (2060). The first analyzing unit (2020) calculates a flow of a crowd in a capturing range of a fixed camera (10) using a first surveillance image (12). The second analyzing unit (2040) calculates a distribution of an attribute of objects in a capturing range of a moving camera (20) using a second surveillance image (22). The estimating unit (2060) estimates an attribute distribution for a range that is not included in the capturing range of the moving camera (20).
Abstract:
An image processing system, an image processing method, and a program capable of implementing an association of a person appearing in a video image through a simple operation are provided. The image processing system includes an input device which accepts input of video images captured by a plurality of video cameras, a display screen generating unit which causes a display device to display at least one video image among the video images inputted from the input device, and a tracked person registering unit which is capable of registering one or more persons appearing in the video image displayed by the display device. When a person appears in the video image displayed by the display device, the display screen generating unit selectably displays person images of one or more persons, which are associable with the person appearing in the video image and which are registered by the tracked person registering unit, in a vicinity of the video image.
Abstract:
An information processing apparatus (2000) includes a recognizer (2020). An image (10) is input to the recognizer (2020). The recognizer (2020) outputs, for a crowd included in the input image (10), a label (30) describing a type of the crowd and structure information (40) describing a structure of the crowd. The structure information (40) indicates a location and a direction of an object included in the crowd. The information processing apparatus (2000) acquires training data (50) which includes a training image (52), a training label (54), and training structure information (56). The information processing apparatus (2000) performs training of the recognizer (2020) using the label (30) and the structure information (40), which are acquired by inputting the training image (52) with respect to the recognizer (2020, and the training label (54) and the training structure information (56).
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.
Abstract:
The present invention provides a learning apparatus (10) including an acquisition unit (11) that acquires learning data associating a training image including a person with a correct answer label indicating a position of each person, a correct answer label indicating whether each of a plurality of keypoints of a body of each of the persons is visible in the training image, and a correct answer label indicating a position, within the training image, of the keypoint being visible in the training image among a plurality of the keypoints, and a learning unit (12) that learns, based on the learning data, an estimation model that estimates information indicating a position of each person, information indicating whether each of a plurality of the keypoints of each person included in a processing image is visible in the processing image, and information being related to a position of each of keypoints for computing a position, within the processing image, of the keypoint being visible in the processing image.