Abstract:
Systems and methods for describing video content establish video description records which include an object set, an object hierarchy and entity relation graphs. Video objects can include global ojects, segment objects and local objects. The video objects are further defined by a number of features organized in classes, which in turn are further defined by a number of feature descriptors. The relationships between and among the objects in the object set are defined by the object hierarchy and entity relation graphs. The video description records provide a standard vehicle for describing the content and context of video information for subsequent access and processing by computer applications such as search engines, filters, and archive systems.
Abstract:
Systems and methods for describing image content establish image description records which include an object set (24), an object hierarchy (26) and entity relation graphs (28). For image content, image objects can include global objects (O0 8) and local objects (O1 2 and O2 6). The image objects are further defined by a number of features of different classes (36, 38 and 40), which in turn are further defined by a number of feature descriptors. The relationships between and among the objects in the object set are defined by the object hierarchy (26) and entity relation graphs (28). The image description records provide a standard vehicle for describing the content and context of image information for subsequent access and processing by computer applications such as search engines, filters, and archive systems.
Abstract:
A computer system for automating the shifting of pixels within a digital video receives a first starting point through a user interface. The first starting point is received through a user selection of a first beginning portion of a video frame. The system then receives a first ending point through the user interface. The first ending point is received through a user selection of a first ending portion of the user interface. The system selects a first set of pixels that lie between the first starting point and the first ending point. The system then shifts the first set of pixels in the first direction, wherein shifting the first set of pixels comprises rendering and re-rendering in a loop the first set of pixels being shifted.
Abstract:
Systems and methods for augmenting and sharing data from audio/video (A/V) recording and communication devices are provided. In one embodiment, a method may include receiving, from an A/V recording and communication device, image data representing an image, analyzing the image data to determine one or more objects depicted by the image data, based on the analyzing the image data, determining at least one comment associated with the one or more objects; transmitting, to a first client device associated with the A/V recording and communication device, the image data representing the image and the at least one comment, receiving, from the first client device, consent for sharing the image data with a geographic area network, the consent indicating a selection of a comment from the at least one comment, and transmitting, to a second client device, the image data representing the image and the comment.
Abstract:
Various aspects of the subject technology relate to systems, methods, and machine-readable media for clustering search results based on image composition. A system may, for each image in a set of images, determine a compositional vector representing one or more objects and corresponding locations within the image using a trained computer-operated convolutional neural network. The system may provide each image through a clustering algorithm to produce one or more clusters based on compositional similarity. The system may provide images from the set of images clustered by composition, in which the images include a different listing of images for each of the one or more clusters. The system may provide a prioritized listing of images responsive to a user search query, in which the prioritized listing of images includes a different listing of images for each cluster of compositional similarity based on the metadata of each image associated with the cluster.
Abstract:
본 발명의 다양한 실시 예들은 복수의 이미지들에 대한 영상 인식을 수행하고, 상기 영상 인식의 결과에 적어도 일부 기초하여, 상기 복수의 이미지들 간의 유사도(similarity)를 계산하고, 상기 계산된 유사도에 적어도 일부 기초하여, 상기 복수의 이미지들 중 둘 이상의 이미지를 포함하는 그룹을 생성하고, 상기 그룹 내에 포함된 둘 이상의 이미지들 간의 유사도에 적어도 일부 기초하여, 디스플레이될 시퀀스를 결정하고, 상기 디스플레이 상에, 상기 시퀀스에 따라, 상기 둘 이상의 이미지들을 표시하도록 하는 인스트럭션들을 저장할 수 있다. 이외에도 다양한 다른 실시예들이 가능하다.
Abstract:
Some embodiments are directed to an unmanned vehicle for use with a companion unmanned vehicle. The unmanned vehicle can include a satellite navigation unit that is configured to receive a satellite signal indicative of a current position of the unmanned vehicle. The unmanned vehicle can also include an inertial navigation unit that is configured to determine the current position of the unmanned vehicle. The unmanned vehicle can also include a control unit disposed in communication with the satellite navigation unit and the inertial navigation unit. The control unit is configured to determine a planned position of the unmanned vehicle based on the planned path, compare the current position determined by the inertial navigation unit with the planned position based on the planned path, and control the movement of the unmanned vehicle based on at least the comparison between the current position and the planned position.
Abstract:
An unmanned vehicle for use with an entity physically spaced from the unmanned vehicle, the unmanned vehicle having objective parameters corresponding to controlled parameters of the entity. The unmanned vehicle comprises a transceiver that is configured to wirelessly receive an input signal from the entity, wherein the input signal is indicative of the controlled parameters of the entity. The unmanned vehicle further comprises a Phase-Locked Loop (PLL) circuit that is configured to generate a command signal based on a phase of the input signal and a phase of a reference signal, wherein the reference signal is indicative of the objective parameters of the unmanned vehicle. The transceiver is further configured to wirelessly transmit the command signal to the entity such that the entity controls the controlled parameters of the entity based on the command signal.
Abstract:
A system and method is provided for displaying surfaces of an object (240) from a vantage point (230) different from the vantage point from which imagery of the object was captured (210, 220). In some aspects, imagery (710) may be generated for display by combining visual characteristics from multiple source images (215, 225) and applying greater weight to the visual characteristics of some of the source images relative to the other source images. The weight may be based on the orientation of the surface (310) relative to the location from which the image was captured (320) and the location from which the object will be displayed (430).