Abstract:
Monitoring target matching means 71 matches monitoring targets shown in video captured by one or more imaging devices, and identifies monitoring targets estimated to be the same monitoring target, as an identified monitoring target. Target object identifying means 72 identifies a desired target object from one or more identified monitoring targets captured, using imaging times of each of the one or more identified monitoring targets.
Abstract:
A state acquisition unit (2020) acquires a state of a monitoring target in a captured image captured by a camera (3040). A monitoring point acquisition unit (2040) acquires, from a monitoring point information storage unit (3020), a monitoring point corresponding to the state of the monitoring target acquired by the state acquisition unit (2020). The monitoring point indicates a position to be monitored in the captured image. A presentation unit (2060) presents the monitoring point on the captured image.
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.
Abstract:
Monitoring target matching means 71 matches monitoring targets shown in video captured by one or more imaging devices, and identifies monitoring targets estimated to be the same monitoring target, as an identified monitoring target. Target object identifying means 72 identifies a desired target object from one or more identified monitoring targets captured, using imaging times of each of the one or more identified monitoring targets.
Abstract:
Provided is an image processing apparatus (2000) including an index value calculation unit (2020) and a presentation unit (2040). The index value calculation unit (2020) acquires a plurality of images captured by a camera (3000) (captured images), and calculates an index value indicating the degree of change in the state of a monitoring target in the captured image, using the acquired captured image. The presentation unit (2040) presents an indication based on the index value calculated by the index value calculation unit (2020) on the captured image captured by the camera (3000).
Abstract:
Provided is an image recognition system that can easily perform image recognition on a side face of an item. An image recognition system according to one example embodiment of the present invention includes: a placement stage used for placing an item below an image capture device provided so as to perform capturing of a downward direction; a support structure configured to support the item at a predetermined angle relative to a top face of the placement stage; and an image recognition apparatus that identifies the item by performing image recognition on an image of the item acquired by the image capture device.
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.
Abstract:
A state acquisition unit (2020) acquires a state of a monitoring target in a captured image captured by a camera (3040). A monitoring point acquisition unit (2040) acquires, from a monitoring point information storage unit (3020), a monitoring point corresponding to the state of the monitoring target acquired by the state acquisition unit (2020). The monitoring point indicates a position to be monitored in the captured image. A presentation unit (2060) presents the monitoring point on the captured image.
Abstract:
Provided is an image processing apparatus (2000) including an index value calculation unit (2020) and a presentation unit (2040). The index value calculation unit (2020) acquires a plurality of images captured by a camera (3000) (captured images), and calculates an index value indicating the degree of change in the state of a monitoring target in the captured image, using the acquired captured image. The presentation unit (2040) presents an indication based on the index value calculated by the index value calculation unit (2020) on the captured image captured by the camera (3000).