SELF-SUPERVISED CROSS-VIDEO TEMPORAL DIFFERENCE LEARNING FOR UNSUPERVISED DOMAIN ADAPTATION

    公开(公告)号:US20210374481A1

    公开(公告)日:2021-12-02

    申请号:US17317202

    申请日:2021-05-11

    Abstract: A method is provided for Cross Video Temporal Difference (CVTD) learning. The method adapts a source domain video to a target domain video using a CVTD loss. The source domain video is annotated, and the target domain video is unannotated. The CVTD loss is computed by quantizing clips derived from the source and target domain videos by dividing the source domain video into source domain clips and the target domain video into target domain clips. The CVTD loss is further computed by sampling two clips from each of the source domain clips and the target domain clips to obtain four sampled clips including a first source domain clip, a second source domain clip, a first target domain clip, and a second target domain clip. The CVTD loss is computed as |(second source domain clip−first source domain clip)−(second target domain clip−first target domain clip)|.

    UNSUPERVISED DOMAIN ADAPTATION FOR VIDEO CLASSIFICATION

    公开(公告)号:US20200065617A1

    公开(公告)日:2020-02-27

    申请号:US16515593

    申请日:2019-07-18

    Abstract: A method is provided for unsupervised domain adaptation for video classification. The method learns a transformation for each target video clips taken from a set of target videos, responsive to original features extracted from the target video clips. The transformation corrects differences between a target domain corresponding to target video clips and a source domain corresponding to source video clips taken from a set of source videos. The method adapts the target to the source domain by applying the transformation to the original features extracted to obtain transformed features for the plurality of target video clips. The method converts the original and transformed features of same ones of the target video clips into a single classification feature for each of the target videos. The method classifies a new target video relative to the set of source videos using the single classification feature for each of the target videos.

    Attention and warping based domain adaptation for videos

    公开(公告)号:US11222210B2

    公开(公告)日:2022-01-11

    申请号:US16673156

    申请日:2019-11-04

    Abstract: A computer-implemented method is provided for domain adaptation between a source domain and a target domain. The method includes applying, by a hardware processor, an attention network to features extracted from images included in the source and target domains to provide attended features relating to a given task to be domain adapted between the source and target domains. The method further includes applying, by the hardware processor, a deformation network to at least some of the attended features to align the attended features between the source and target domains using warping to provide attended and warped features. The method also includes training, by the hardware processor, a target domain classifier using the images from the source domain. The method additionally includes classifying, by the hardware processor using the trained target domain classifier, at least one image from the target domain.

    Unsupervised domain adaptation for video classification

    公开(公告)号:US11301716B2

    公开(公告)日:2022-04-12

    申请号:US16515593

    申请日:2019-07-18

    Abstract: A method is provided for unsupervised domain adaptation for video classification. The method learns a transformation for each target video clips taken from a set of target videos, responsive to original features extracted from the target video clips. The transformation corrects differences between a target domain corresponding to target video clips and a source domain corresponding to source video clips taken from a set of source videos. The method adapts the target to the source domain by applying the transformation to the original features extracted to obtain transformed features for the plurality of target video clips. The method converts the original and transformed features of same ones of the target video clips into a single classification feature for each of the target videos. The method classifies a new target video relative to the set of source videos using the single classification feature for each of the target videos.

Patent Agency Ranking