Unsupervised Representation Learning by Sorting Sequence



Abstract
We present an unsupervised representation learning approach using videos without semantic labels. We leverage the temporal coherence as a supervisory signal by formulating representation learning as a sequence sorting task. We take temporally shuffled frames (\ie in non-chronological order) as inputs and train a convolutional neural network to sort the shuffled sequences. Similar to comparison-based sorting algorithms, we propose to extract features from all frame pairs and aggregate them to predict the correct order. As sorting shuffled image sequence requires an understanding of the statistical temporal structure of images, training with such a proxy task allows us to learn rich and generalizable visual representation. We validate the effectiveness of the learned representation using our method as pre-training on high-level recognition problems. The experimental results show that our method compares favorably against state-of-the-art methods on action recognition, image classification and object detection tasks.

Citation

Hsin-Ying Lee, Jia-Bin Huang, Maneesh Kumar Singh, and Ming-Hsuan Yang, "Unsupervised Representation Learning by Sorting Sequence", in IEEE International Conference on Computer Vision, 2017.


Bibtex
@inproceeding{OPN,
    author    = {Lee, Hsin-Ying and Huang, Jia-Bin and Singh, Maneesh Kumar and Yang, Ming-Hsuan}, 
    title     = {Unsupervised Representation Learning by Sorting Sequence}, 
    booktitle = {IEEE International Conference on Computer Vision},
    year      = {2017}
}
Network Architecture

Results
Performance on UCF-101 (3 splits)
Performance on HMDB-51 (3 splits)
Performance on Pascal VOC 2007
References