Systems and methods for generating comic books from video and images

    公开(公告)号:US11532111B1

    公开(公告)日:2022-12-20

    申请号:US17344690

    申请日:2021-06-10

    Abstract: Techniques for a comic book feature are described herein. A visual data stream of a video may be parsed into a plurality of frames. Scene boundaries may be determined to generate a scene using the plurality of frames where a scene includes a subset of frames. A key frame may be determined for the scene using the subset of frames. An audio portion of an audio data stream of the video may be identified that maps to the subset of frames based on time information. The key frame may be converted to a comic image based on an algorithm. First dimensions and placement for a data object may be determined for the comic image. The data object may include the audio portion for the comic image. A comic panel may be generated for the comic image that incorporates the data object using the determined first dimensions and the placement.

    Navigation directly from perception data without pre-mapping

    公开(公告)号:US11175664B1

    公开(公告)日:2021-11-16

    申请号:US16248516

    申请日:2019-01-15

    Abstract: An autonomous delivery robot system to enable delivery of a product to a customer is described. One autonomous ground vehicle (AGV) includes a processing device that receives a delivery request comprising a route divided into multiple navigation segments and computes a navigable space from perception data stored in a perception map. The perception map is a robot-centered local map that stores the perception data indicative of the surroundings of the AGV. The processing device computes a cost inflation from the perception data stored in the perception map, determines a sub-goal that is on the navigable space and reachable by the AGV using the navigable space and the cost inflation, and determines a path to achieve the sub-goal using the using the navigable space and the cost inflation. The processing device controls one or more actuators to move along the path.

    Contrastive learning of scene representation guided by video similarities

    公开(公告)号:US12067779B1

    公开(公告)日:2024-08-20

    申请号:US17668014

    申请日:2022-02-09

    CPC classification number: G06V20/48 G06V10/774 G06V20/46

    Abstract: A plurality of similar video pairs may be determined based on one or more similarity information types. Each video pair of the plurality of similar video pairs may include a first respective video and a second respective video. For each video pair, one or more similar scene pairs may be determined. Each of the one or more similar scene pairs may include a respective first scene from the first respective video and a second respective scene from the second respective video. An encoder may be trained using a contrastive learning model that contrasts a plurality of similar scene pairs with a plurality of random scenes. The plurality of similar scene pairs may include the one or more scene pairs for each video pair. One or more scene features of one or more other scenes of one or more other videos may be determined using the encoder.

    Depth-guided structure-from-motion techniques

    公开(公告)号:US12046002B1

    公开(公告)日:2024-07-23

    申请号:US17684197

    申请日:2022-03-01

    CPC classification number: G06T7/73 G06T7/55

    Abstract: Systems, devices, and methods are provided for depth guided structure from motion. A system may obtain a plurality of image frames from a digital content item that corresponds to a scene and determine, based at least in part on a correspondence search, a set of 2-D keypoints for the plurality of image frames. A depth estimator may be used to determine a plurality of dense depth map for the plurality of image frames. The set of 2-D keypoints and the plurality of dense depth maps may be used to determine a corresponding set of depth priors. Initialization and/or depth-regularized optimization may be performed using the keypoints and depth priors.

    SYSTEMS AND METHODS OF OBSTACLE DETECTION FOR AUTOMATED DELIVERY APPARATUS

    公开(公告)号:US20210405638A1

    公开(公告)日:2021-12-30

    申请号:US16914189

    申请日:2020-06-26

    Abstract: The present disclosure generally relates to a system of a delivery device for combining sensor data from various types of sensors to generate a map that enables the delivery device to navigate from a first location to a second location to deliver an item to the second location. The system obtains data from RGB, LIDAR, and depth sensors and combines this sensor data according to various algorithms to detect objects in an environment of the delivery device, generate point cloud and pose information associated with the detected objects, and generates object boundary data for the detected objects. The system further identifies object states for the detected object and generates the map for the environment based on the detected object, the generated object proposal data, the labeled point cloud data, and the object states. The generated map may be provided to other systems to navigate the delivery device.

    DEPTH-GUIDED STRUCTURE-FROM-MOTION TECHNIQUES

    公开(公告)号:US20240346686A1

    公开(公告)日:2024-10-17

    申请号:US18749025

    申请日:2024-06-20

    CPC classification number: G06T7/73 G06T7/55

    Abstract: Systems, devices, and methods are provided for depth-guided structure from motion. A system may obtain a plurality of image frames from a digital content item that corresponds to a scene and determine, based at least in part on a correspondence search, a set of 2-D keypoints for the plurality of image frames. A depth estimator may be used to determine a plurality of dense depth map for the plurality of image frames. The set of 2-D keypoints and the plurality of dense depth maps may be used to determine a corresponding set of depth priors. Initialization and/or depth-regularized optimization may be performed using the keypoints and depth priors.

Patent Agency Ranking