Systems and methods for generating comic books from video and images

    公开(公告)号:US11532111B1

    公开(公告)日:2022-12-20

    申请号:US17344690

    申请日:2021-06-10

    Abstract: Techniques for a comic book feature are described herein. A visual data stream of a video may be parsed into a plurality of frames. Scene boundaries may be determined to generate a scene using the plurality of frames where a scene includes a subset of frames. A key frame may be determined for the scene using the subset of frames. An audio portion of an audio data stream of the video may be identified that maps to the subset of frames based on time information. The key frame may be converted to a comic image based on an algorithm. First dimensions and placement for a data object may be determined for the comic image. The data object may include the audio portion for the comic image. A comic panel may be generated for the comic image that incorporates the data object using the determined first dimensions and the placement.

    Language-agnostic subtitle drift detection and localization

    公开(公告)号:US10945041B1

    公开(公告)日:2021-03-09

    申请号:US16890940

    申请日:2020-06-02

    Abstract: Devices, systems, and methods are provided for language-agnostic subtitle drift detection and localization. A method may include extracting audio from video, dividing the audio into overlapping blocks, and determining the probabilities of overlapping portions of the blocks, the probabilities indicating a presence of voice data represented by the audio in the blocks. The method may generate machine blocks using overlapping portions of blocks where voice data is present, and may map the machine blocks to corresponding blocks indicating that subtitles are available for the video. For mapped blocks, the method may include determining features such as when subtitles are available without voice audio, when voice audio is available without subtitles, and when voice audio and subtitles both are available. Using the features, the method may include determining the probability that the video includes subtitle drift, and the method may include analyzing the video to localize where the subtitle drift occurs.

    DEPTH-GUIDED STRUCTURE-FROM-MOTION TECHNIQUES

    公开(公告)号:US20240346686A1

    公开(公告)日:2024-10-17

    申请号:US18749025

    申请日:2024-06-20

    CPC classification number: G06T7/73 G06T7/55

    Abstract: Systems, devices, and methods are provided for depth-guided structure from motion. A system may obtain a plurality of image frames from a digital content item that corresponds to a scene and determine, based at least in part on a correspondence search, a set of 2-D keypoints for the plurality of image frames. A depth estimator may be used to determine a plurality of dense depth map for the plurality of image frames. The set of 2-D keypoints and the plurality of dense depth maps may be used to determine a corresponding set of depth priors. Initialization and/or depth-regularized optimization may be performed using the keypoints and depth priors.

    Language agnostic drift correction
    27.
    发明授权

    公开(公告)号:US11900700B2

    公开(公告)日:2024-02-13

    申请号:US18175044

    申请日:2023-02-27

    Abstract: Systems, methods, and computer-readable media are disclosed for language-agnostic subtitle drift detection and correction. A method may include determining subtitles and/or captions from media content (e.g., videos), the subtitles and/or captions corresponding to dialog in the media content. The subtitles may be broken up into segments which may be analyzed to determine a likelihood of drift (e.g., a likelihood that the subtitles are out of synchronization with the dialog in the media content) for each segment. For segments with a high likelihood of drift, the subtitles may be incrementally adjusted to determine an adjustment that eliminates and/or reduces the amount of drift, and the drift in the segment may be corrected based on the drift amount detected. A linear regression model and/or human blocks determined by human operators may be used to otherwise optimize drift correction.

Patent Agency Ranking