Abstract:
An information processing apparatus includes a network learning portion that performs learning of an appearance/position recognition network by constraining first to third weights and using a learning image, wherein the appearance/position recognition network has a foreground layer including a position node, a background layer including a background node, and an image layer including a pixel node, and is a neural network in which the position node, the background node, and the pixel node are connected to each other, and wherein the first weight is a connection weight between the position node and the pixel node, the second weight is a connection weight between the position node and the background node, and the third weight is a connection weight between the background node and the pixel node.
Abstract:
An information processing apparatus includes a network learning portion that performs learning of an appearance/position recognition network by constraining first to third weights and using a learning image, wherein the appearance/position recognition network has a foreground layer including a position node, a background layer including a background node, and an image layer including a pixel node, and is a neural network in which the position node, the background node, and the pixel node are connected to each other, and wherein the first weight is a connection weight between the position node and the pixel node, the second weight is a connection weight between the position node and the background node, and the third weight is a connection weight between the background node and the pixel node.
Abstract:
An information input apparatus includes an observation unit that observes an environment including a user and one or more apparatuses to be controlled and includes a sensor; a learning unit that separates a foreground including the user and the one or more apparatuses to be controlled and a background including the environment except for the foreground from observation data obtained by the observation unit and learns three-dimensional models of the foreground and the background; a state estimation unit that estimates positions and postures of already modeled foregrounds in the environment; a user recognition unit that identifies fingers of the user from the foreground and recognizes a shape, position, and posture of the fingers; and an apparatus control unit that outputs a control command to the one or more apparatuses to be controlled on the basis of the recognized shape, position, and posture of the fingers.
Abstract:
An information processing device includes: a foreground state estimating unit configured to estimate a foreground state of an image using an actual image which is an image to be actually observed; and a visible model updating unit configured to update a background visible model which is visibility of the background of an image and a foreground visible model which is visibility of the foreground using an estimation result of the foreground state.
Abstract:
There is provided an information processing apparatus including a manipulation model learning unit configured to learn a manipulation model regarding manipulation of a first object by a second object, by use of an actual image that is an actually observed image including the first object and the second object, the manipulation model associating a position and a change in state of the second object, when a state of the second object changes at a position in an object reference coordinate system with the first object regarded as a reference, with a change in state of the first object caused by the change in state of the second object.
Abstract:
An information input apparatus includes an observation unit that observes an environment including a user and one or more apparatuses to be controlled and includes a sensor; a learning unit that separates a foreground including the user and the one or more apparatuses to be controlled and a background including the environment except for the foreground from observation data obtained by the observation unit and learns three-dimensional models of the foreground and the background; a state estimation unit that estimates positions and postures of already modeled foregrounds in the environment; a user recognition unit that identifies fingers of the user from the foreground and recognizes a shape, position, and posture of the fingers; and an apparatus control unit that outputs a control command to the one or more apparatuses to be controlled on the basis of the recognized shape, position, and posture of the fingers.
Abstract:
An information input apparatus includes an observation unit that observes an environment including a user and one or more apparatuses to be controlled and includes a sensor; a learning unit that separates a foreground including the user and the one or more apparatuses to be controlled and a background including the environment except for the foreground from observation data obtained by the observation unit and learns three-dimensional models of the foreground and the background; a state estimation unit that estimates positions and postures of already modeled foregrounds in the environment; a user recognition unit that identifies fingers of the user from the foreground and recognizes a shape, position, and posture of the fingers; and an apparatus control unit that outputs a control command to the one or more apparatuses to be controlled on the basis of the recognized shape, position, and posture of the fingers.