Abstract:
A mechanism is described for facilitating depth and motion estimation in machine learning environments, according to one embodiment. A method of embodiments, as described herein, includes receiving a frame associated with a scene captured by one or more cameras of a computing device; processing the frame using a deep recurrent neural network architecture, wherein processing includes simultaneously predicating values associated with multiple loss functions corresponding to the frame; and estimating depth and motion based the predicted values.
Abstract:
A perception device, including at least one image sensor configured to detect a plurality of images; an information estimator configured to estimate from each image of the plurality of images a depth estimate, a velocity estimate, an object classification estimate and an odometry estimate; a particle generator configured to generate a plurality of particles, wherein each particle of the plurality of particles comprises a position value determined from the depth estimate, a velocity value determined from the velocity estimate and a classification value determined from the classification estimate; an occupancy hypothesis determiner configured to determine an occupancy hypothesis of a predetermined region, wherein each particle of the plurality of particles contributes to the determination of the occupancy hypothesis.
Abstract:
Technologies for determining a distance of an object from a vehicle include a computing device to identify an object captured in a fisheye image generated by a fisheye camera of the vehicle. The computing device projects a contour of the identified object on a selected virtual plane that is located outside the vehicle and selected from a predefined set of virtual planes based on a location of the identified object relative to the vehicle. The computing device identifies a bottom of the projected contour on the selected virtual plane and determines an intersection point of an imaginary line with a ground plane coincident with a plane on which the vehicle is positioned. The imaginary line passes through each of the identified bottom of the projected contour and the fisheye camera. The computing device determines a location of the identified object relative to the vehicle based on the determined intersection point and the identified bottom of the projected contour.
Abstract:
A mechanism is described for facilitating depth and motion estimation in machine learning environments, according to one embodiment. A method of embodiments, as described herein, includes receiving a frame associated with a scene captured by one or more cameras of a computing device; processing the frame using a deep recurrent neural network architecture, wherein processing includes simultaneously predicating values associated with multiple loss functions corresponding to the frame; and estimating depth and motion based the predicted values.
Abstract:
According to various examples, a vehicle controller is described comprising a determiner configured to determine information about surroundings of a vehicle, the information about the surroundings comprising information about velocities of objects in the surroundings of the vehicle and a velocity controller configured to input the information about the surroundings of the vehicle and a specification of a path of the vehicle to a convolutional neural network, to determine a target velocity of the vehicle along the path based on an output of the convolutional neural network and to control the vehicle according to the determined target velocity.
Abstract:
A perception device including a receiver configured to receive sensor information including information about a location of one or more objects detected by a sensor; a memory configured to store an occupancy grid of a predetermined region, wherein the occupancy grid includes a plurality of grid cells, wherein each grid cell represents an area in the predetermined region, wherein at least one grid cell of the plurality of grid cells is associated with a respective single occupancy hypothesis; a single occupancy hypothesis determiner configured to determine a single occupancy hypothesis; wherein the single occupancy hypothesis includes degree of belief of occupancy of the grid cell depending on the sensor information; wherein a contribution of a sensor information value to the degree of belief of occupancy substantially decreases with an increase of a distance of the location of the object detected by the sensor from a center of the grid cell.
Abstract:
A method of determining a trajectory of motion of a vehicle in a predetermined region, wherein the predetermined region includes a plurality of sub-regions, the method being executed by one or more processors, the method including determining an occupancy hypothesis of the predetermined region, wherein the occupancy hypothesis indicates occupied sub-regions of the plurality of sub-regions and non-occupied sub-regions of the plurality of sub-regions; determining a utility value for each sub-region of the predetermined region; determining the trajectory of motion which crosses at least one sub-region of the non-occupied sub-regions, based on a function of the utility values of the least one sub-region of the non-occupied sub-regions crossed by the trajectory of motion and by maximizing a utility of motion of the vehicle, wherein the utility of motion of the vehicle is indicated by a function of the utility values of the sub-regions crossed by the trajectory of motion.
Abstract:
Techniques are provided for adaptive selection of feature keypoints of an image. An example system may include a contrast statistics calculation circuit configured to generate contrast measurements of regions of the image associated with each of the feature keypoints, and to calculate a mean and variance of the contrast measurements. The system may also include an edge statistics calculation circuit configured to generate ratios of principal curvatures of regions of the image associated with each of the feature keypoints, and to calculate a mean and variance of the ratios of principal curvatures. The system may further include a threshold calculation circuit configured to calculate thresholds based on the mean and variance of the contrast measurements and on the mean and variance of the ratios of principal curvatures; and a keypoint filter circuit configured to filter the set of feature keypoints based on the those thresholds.
Abstract:
Techniques are disclosed for improving classification performance in supervised learning. In accordance with some embodiments, a multiclass support vector machine (SVM) having three or more classes may be converted to a plurality of binary problems that then may be reduced via one or more reduced-set methods. The resultant reduced-set (RS) vectors may be combined together in one or more joint lists, along with the original support vectors (SVs) of the different binary classes. Each binary problem may be re-trained using the joint list(s) by applying a reduction factor (RF) parameter to reduce the total quantity of RS vectors. In re-training, different kernel methods can be combined, in accordance with some embodiments. Reduction may be performed until desired classification performance is achieved. The disclosed techniques can be used, for example, to improve classification speed, accuracy, class prioritization, or a combination thereof, in the SVM training phase, in accordance with some embodiments.
Abstract:
Methods, apparatuses, and systems may provide for using the motion of a vehicle to estimate the orientation of a camera system of a vehicle relative to the vehicle. Image data may be received from a plurality of cameras positioned on the vehicle, and a first constraint set may be determined for the plurality of cameras based on a plurality of feature points in a ground plane proximate to the vehicle. A second constraint set may be determined based on one or more borders of the vehicle. One or more of the cameras may be automatically calibrated based on the first constraint set and the second constraint set.