Abstract:
Methods and systems for interacting with multiple three-dimensional (3D) object data models are provided. An example method may involve receiving a plurality of three-dimensional (3D) object data models that represent objects associated with a given category. The plurality of 3D object data models may include one or more annotated parts. The method may also involve receiving a candidate 3D object data model that is associated with the given category of objects. The candidate 3D object data model may include one or more un-annotated parts. The method may additionally involve, based on a shape and an appearance of the one or more annotated parts and spatial-layout information that defines how the one or more annotated parts are spatially arranged, applying an annotation of a respective part of a respective 3D object data model to a corresponding part of the candidate 3D object data model.
Abstract:
Methods, systems, and apparatus, including computer program products, for ranking search results for queries. The method includes calculating a visual similarity score for one or more pairs of images in a plurality of images based on visual features of images in each of the one or more pairs; building a graph of images by linking each of one or more images in the plurality of images to one or more nearest neighbor images based on the visual similarity scores; associating a respective score with each of one or more images in the graph based on data indicative of user behavior relative to the image as a search result for a query; and determining a new score for each of one or more images in the graph based on the respective score of the image, and the respective scores of one or more nearest neighbors to the image.
Abstract:
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for processing joint image-audio queries. In one aspect, a method includes receiving, from a client device, a joint image-audio query including query image data and query audio data. Query image feature data is determined from the query image data. Query audio feature data is determined from the audio data. The query image feature data and the query audio feature data are provided to a joint image-audio relevance model trained to generate relevance scores for a plurality of resources, each resource including resource image data defining a resource image for the resource and text data defining resource text for the resource. Each relevance score is a measure of the relevance of corresponding resource to the joint image-audio query. Data defining search results indicating the order of the resources is provided to the client device.