Shuffle, attend, and adapt: video domain adaptation by clip order prediction and clip attention alignment
Abstract:
A method for performing video domain adaptation for human action recognition is presented. The method includes using annotated source data from a source video and unannotated target data from a target video in an unsupervised domain adaptation setting, identifying and aligning discriminative clips in the source and target videos via an attention mechanism, and learning spatial-background invariant human action representations by employing a self-supervised clip order prediction loss for both the annotated source data and the unannotated target data.
Information query
Patent Agency Ranking
0/0