1. Self-Supervised Representation Learning for Videos by Segmenting via Sampling Rate Order Prediction
- Author
-
Wenming Yang, Qicong Wang, Hongying Meng, Yan Huang, and Jing Huang
- Subjects
business.industry ,Computer science ,feature extraction ,streaming media ,Feature extraction ,motion segmentation ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,data mining ,Image segmentation ,Expression (mathematics) ,View synthesis ,Discriminative model ,task analysis ,Media Technology ,Feature (machine learning) ,Artificial intelligence ,Electrical and Electronic Engineering ,Representation (mathematics) ,business ,semantics ,image segmentation ,Feature learning - Abstract
Self-supervised representation learning for videos has been very attractive recently because these methods exploit the information inherently obtained from the video itself instead of annotated labels that is quite time-consuming. However, existing methods ignore the importance of global observation while performing spatio-temporal transformation perception, which highly limits the expression capabilities of the video representation. This paper proposes a novel pretext task that combines the temporal information perception of the video with the motion amplitude perception of moving objects to learn the spatio-temporal representation of the video. Specifically, given a video clip containing several video segments, each video segment is sampled by different sampling rates and the order of video segments is disrupted. Then, the network is used to regress the sampling rate of each video segment and classify the order of input video segments. In the pre-training stage, the network can learn rich spatio-temporal semantic information where content-related contrastive learning is introduced to make the learned video representation more discriminative. To alleviate the appearance dependency caused by contrastive learning, we design a novel and robust vector similarity measurement approach, which can take feature alignment into consideration. Moreover, a view synthesis framework is proposed to further improve the performance of contrastive learning by automatically generating reasonable transformed views. We conduct benchmark experiments with several 3D backbone networks on two datasets. The results show that our proposed method outperforms the existing state-of-the-art methods across the three backbones on two downstream tasks of human action recognition and video retrieval.
- Published
- 2022