Abstract:
Methods and devices for actively modifying a field of view of an autonomous vehicle in view of constraints are disclosed. In one embodiment, an example method is disclosed that includes causing a sensor in an autonomous vehicle to sense information about an environment in a first field of view, where a portion of the environment is obscured in the first field of view. The example method further includes determining a desired field of view in which the portion of the environment is not obscured and, based on the desired field of view and a set of constraints for the vehicle, determining a second field of view in which the portion of the environment is less obscured than in the first field of view. The example method further includes modifying a position of the vehicle, thereby causing the sensor to sense information in the second field of view.
Abstract:
A method and apparatus are provided for determining one or more behavior models used by an autonomous vehicle to predict the behavior of detected objects. The autonomous vehicle may collect and record object behavior using one or more sensors. The autonomous vehicle may then communicate the recorded object behavior to a server operative to determine the behavior models. The server may determine the behavior models according to a given object classification, actions of interest performed by the object, and the object's perceived surroundings.
Abstract:
A system and method of displaying transitions between street level images is provided. In one aspect, the system and method creates a plurality of polygons that are both textured with images from a 2D street level image and associated with 3D positions, where the 3D positions correspond with the 3D positions of the objects contained in the image. These polygons, in turn, are rendered from different perspectives to convey the appearance of moving among the objects contained in the original image.
Abstract:
A method and apparatus are provided for optimizing one or more object detection parameters used by an autonomous vehicle to detect objects in images. The autonomous vehicle may capture the images using one or more sensors. The autonomous vehicle may then determine object labels and their corresponding object label parameters for the detected objects. The captured images and the object label parameters may be communicated to an object identification server. The object identification server may request that one or more reviewers identify objects in the captured images. The object identification server may then compare the identification of objects by reviewers with the identification of objects by the autonomous vehicle. Depending on the results of the comparison, the object identification server may recommend or perform the optimization of one or more of the object detection parameters.
Abstract:
The present invention relates to using image content to facilitate navigation in panoramic image data. In an embodiment, a computer-implemented method for navigating in panoramic image data includes: (1) determining an intersection of a ray and a virtual model, wherein the ray extends from a camera viewport of an image and the virtual model comprises a plurality of facade planes; (2) retrieving a panoramic image; (3) orienting the panoramic image to the intersection; and (4) displaying the oriented panoramic image.
Abstract:
A vehicle may operate in an autonomous mode in an environment during a test period. The vehicle may include at least one sensor coupled to the vehicle, configured to acquire sensor data during the test period. The sensor data may include data representative of a target object in the environment. The vehicle may operate the sensor to obtain the sensor data. The vehicle may define a movement of the vehicle, determine a predicted movement of the target object in the sensor data based on the defined movement, initiate the defined movement of the vehicle at an initiation time during the test period, complete the defined movement of the vehicle at a completion time during the test period, analyze the sensor data obtained during the test period, and determine a latency of the at least one sensor based on the analyzed data.
Abstract:
A system and method is provided of determining geographic positions. In one aspect, a user points the device at first and second positions on the surface of a geographic object. Based on the position of the device, the orientation of the device, and information identifying the geographic position of the surface of the object, a processor determines and displays the distance between the first and second positions.
Abstract:
Methods and systems for generating video from panoramic images using transition trees are provided. According to an embodiment, a method for generating a video from panoramic images may include receiving a transition tree corresponding to a current panoramic image from a server. The method may also include determining a path of the transition tree to a next panoramic image based on a user navigation request. The method may further include requesting and receiving a video chunk from the server for each edge of the determined path of the transition tree. The method may also include displaying the requested video chunks in sequence according to the transition tree. According to another embodiment, a system for generating a video from panoramic images may include a transition tree module and a video display module.
Abstract:
A vehicle configured to operate in an autonomous mode may engage in an obstacle evaluation technique that includes employing a sensor system to collect data relating to a plurality of obstacles, identifying from the plurality of obstacles an obstacle pair including a first obstacle and a second obstacle, engaging in an evaluation process by comparing the data collected for the first obstacle to the data collected for the second obstacle, and in response to engaging in the evaluation process, making a determination of whether the first obstacle and the second obstacle are two separate obstacles.
Abstract:
A system and method is provided of determining geographic positions. In one aspect, a user points the device at first and second positions on the surface of a geographic object. Based on the position of the device, the orientation of the device, and information identifying the geographic position of the surface of the object, a processor determines and displays the distance between the first and second positions.