OCCLUSION-AWARE FORWARD WARPING FOR VIDEO FRAME INTERPOLATION

    公开(公告)号:US20250139733A1

    公开(公告)日:2025-05-01

    申请号:US18499604

    申请日:2023-11-01

    Abstract: Systems and techniques described herein relate to generating an inter-frame from a first and second frame. An apparatus includes a memory storing a first frame and a second frame; and a processor coupled to the memory and configured to: estimate at least one optical flow between the first frame and the second frame; generate, based on the at least one optical flow, at least one occlusion mask; generate, based on the at least one optical flow and the at least one occlusion mask, at least one weighting mask; generate, based on the at least one optical flow and the at least one weighting mask, at least one inter-frame optical flow; generate, based on the at least one inter-frame optical flow and at least one of the first frame or the second frame, at least one warped frame; and generate, based on the at least one warped frame, an inter-frame.

    UNIFIED SIMULTANEOUS OPTICAL FLOW AND DEPTH ESTIMATION

    公开(公告)号:US20250095182A1

    公开(公告)日:2025-03-20

    申请号:US18468656

    申请日:2023-09-15

    Abstract: Techniques and systems are provided for image processing. For instance, a process can include correlating a first set of features from a first viewpoint with a second set of features from a second viewpoint at a first time period to generate a first disparity cost volume; correlating a third set of features from the first viewpoint at a second time period with the first set of features to generate a first optical flow cost volume; gating the first disparity cost volume to generate first intermediate disparity information; gating the first optical flow cost volume to generate first intermediate optical flow information; correlating the first set of features, the second set of features, and the first intermediate optical flow information to generate disparity information for output; and correlating the third set of features, the first set of features, and the first intermediate disparity information to generate optical flow information for output.

    HIERARCHICAL SUPERVISED TRAINING FOR NEURAL NETWORKS

    公开(公告)号:US20230004812A1

    公开(公告)日:2023-01-05

    申请号:US17808949

    申请日:2022-06-24

    Abstract: Certain aspects of the present disclosure provide techniques for training neural networks using hierarchical supervision. An example method generally includes training a neural network with a plurality of stages using a training data set and an initial number of classification clusters into which data in the training data set can be classified. A cluster-validation set performance metric is generated for each stage based on a reduced number of classification clusters relative to the initial number of classification clusters and a validation data set. A number of classification clusters to implement at each stage is selected based on the cluster-validation set performance metric and an angle selected relative to the cluster-validation set performance metric for a last stage of the neural network. The neural network is retrained based on the training data set and the selected number of classification clusters for each stage, and the trained neural network is deployed.

    DEPTH ESTIMATION BASED ON FEATURE RECONSTRUCTION WITH ADAPTIVE MASKING AND MOTION PREDICTION

    公开(公告)号:US20250148633A1

    公开(公告)日:2025-05-08

    申请号:US18666502

    申请日:2024-05-16

    Abstract: Systems and techniques are provided for generating depth information. For example, a process can include obtaining a first feature volume including visual features corresponding to each respective frame included in a first set of frames. A first query generator network can generate reconstruction features associated with a reconstructed feature volume corresponding to the first feature volume. Based on the first feature volume, a second query generator network can generate motion features associated with predicted future motion corresponding to the first feature volume. An initial depth prediction can be generated for each respective frame based on cross-attention between features of a depth prediction decoder, the reconstruction features, and the motion features. A refined depth prediction can be generated for each respective based on cross-attention between the initial depth prediction, the reconstruction features, and the motion features.

    MONOCULAR IMAGE DEPTH ESTIMATION WITH ATTENTION

    公开(公告)号:US20240303841A1

    公开(公告)日:2024-09-12

    申请号:US18538869

    申请日:2023-12-13

    CPC classification number: G06T7/50 G06T7/246 G06T11/60 G06V10/44 G06V10/62

    Abstract: Disclosed are systems and techniques for capturing images (e.g., using a monocular image sensor) and detecting depth information. According to some aspects, a computing system or device can generate a feature representation of a current image and update accumulated feature information for storage in a memory based on a feature representation of a previous image and optical flow information of the previous image. The accumulated feature information can include accumulated image feature information associated with a plurality of previous images and accumulated optical flow information associated of the plurality of previous images. The computing system or device can obtain information associated with relative motion of the current image based on the accumulated feature information and the feature representation of the current image. The computing system or device can estimate depth information for the current image based on the information associated with the relative motion and the accumulated feature information.

Patent Agency Ranking