All Projects → laura-wang → video_repres_mas

laura-wang / video_repres_mas

Licence: MIT license
code for CVPR-2019 paper: Self-supervised Spatio-temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to video repres mas

temporal-ssl
Video Representation Learning by Recognizing Temporal Transformations. In ECCV, 2020.
Stars: ✭ 46 (-26.98%)
Mutual labels:  action-recognition, self-supervised-learning
video-clip-order-prediction
Self-supervised Spatiotemporal Learning via Video Clip Order Prediction
Stars: ✭ 85 (+34.92%)
Mutual labels:  self-supervised-learning, cvpr2019
ViCC
[WACV'22] Code repository for the paper "Self-supervised Video Representation Learning with Cross-Stream Prototypical Contrasting", https://arxiv.org/abs/2106.10137.
Stars: ✭ 33 (-47.62%)
Mutual labels:  action-recognition, self-supervised-learning
TA3N
[ICCV 2019 Oral] TA3N: https://github.com/cmhungsteve/TA3N (Most updated repo)
Stars: ✭ 45 (-28.57%)
Mutual labels:  action-recognition, cvpr2019
TCE
This repository contains the code implementation used in the paper Temporally Coherent Embeddings for Self-Supervised Video Representation Learning (TCE).
Stars: ✭ 51 (-19.05%)
Mutual labels:  action-recognition, self-supervised-learning
simsiam-cifar10
Code to train the SimSiam model on cifar10 using PyTorch
Stars: ✭ 33 (-47.62%)
Mutual labels:  self-supervised-learning
TCFPN-ISBA
Temporal Convolutional Feature Pyramid Network (TCFPN) & Iterative Soft Boundary Assignment (ISBA), CVPR '18
Stars: ✭ 40 (-36.51%)
Mutual labels:  action-recognition
SoCo
[NeurIPS 2021 Spotlight] Aligning Pretraining for Detection via Object-Level Contrastive Learning
Stars: ✭ 125 (+98.41%)
Mutual labels:  self-supervised-learning
esvit
EsViT: Efficient self-supervised Vision Transformers
Stars: ✭ 323 (+412.7%)
Mutual labels:  self-supervised-learning
MiniVox
Code for our ACML and INTERSPEECH papers: "Speaker Diarization as a Fully Online Bandit Learning Problem in MiniVox".
Stars: ✭ 15 (-76.19%)
Mutual labels:  self-supervised-learning
theWorldInSafety
Surveillance System Against Violence
Stars: ✭ 31 (-50.79%)
Mutual labels:  action-recognition
synse-zsl
Official PyTorch code for the ICIP 2021 paper 'Syntactically Guided Generative Embeddings For Zero Shot Skeleton Action Recognition'
Stars: ✭ 14 (-77.78%)
Mutual labels:  action-recognition
BossNAS
(ICCV 2021) BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search
Stars: ✭ 125 (+98.41%)
Mutual labels:  self-supervised-learning
BYOL
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Stars: ✭ 102 (+61.9%)
Mutual labels:  self-supervised-learning
newt
Natural World Tasks
Stars: ✭ 24 (-61.9%)
Mutual labels:  self-supervised-learning
G-SimCLR
This is the code base for paper "G-SimCLR : Self-Supervised Contrastive Learning with Guided Projection via Pseudo Labelling" by Souradip Chakraborty, Aritra Roy Gosthipaty and Sayak Paul.
Stars: ✭ 69 (+9.52%)
Mutual labels:  self-supervised-learning
GCL
List of Publications in Graph Contrastive Learning
Stars: ✭ 25 (-60.32%)
Mutual labels:  self-supervised-learning
adascan-public
Code for AdaScan: Adaptive Scan Pooling (CVPR 2017)
Stars: ✭ 43 (-31.75%)
Mutual labels:  action-recognition
Self-Supervised-Embedding-Fusion-Transformer
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
Stars: ✭ 57 (-9.52%)
Mutual labels:  self-supervised-learning
CVPR21 PASS
PyTorch implementation of our CVPR2021 (oral) paper "Prototype Augmentation and Self-Supervision for Incremental Learning"
Stars: ✭ 55 (-12.7%)
Mutual labels:  self-supervised-learning

Self-Supervised Spatio-Temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics

Tensorflow implementation of our CVPR 2019 paper Self-Supervised Spatio-Temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics.

Update

A journal (T-PAMI 2021) extension of this work can be found here, with extensive additional analysis and significant performance gain (~30%). The corresponding PyTorch implemetation is available here: https://github.com/laura-wang/video_repres_sts.

Overview

We realease partial of our training code on UCF101 dataset. It contains the self-supervised learning based on motion statistics (see more details in our paper).
The entire training protocol (both motion statistics and appearance statistics) is implemented in the pytorch version: https://github.com/laura-wang/video_repres_sts.

Requirements

  1. tensorflow >= 1.9.0
  2. Python 3
  3. cv2
  4. scipy

Data preparation

You can download the original UCF101 dataset from the official website. And then extarct RGB images from videos and finally extract optical flow data using TVL1 method. But I recommend you to direclty download the pre-processed RGB and optical flow data of UCF101 provided by feichtenhofer.

Train

Here we provide the first version of our training code with "placeholder" as data reading pipeline, so you don't need to write RGB/Optical flow data into tfrecord format. We also rewrite the training code using Dataset API, but currently we think the placeholder version is enough for you to get to understand motion statsitics.

Before python train.py, remember to set right dataset directory in the list file, and then you can play with the motion statistics!

Citation

If you find this repository useful in your research, please consider citing:

@inproceedings{wang2019self,
  title={Self-Supervised Spatio-Temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics},
  author={Wang, Jiangliu and Jiao, Jianbo and Bao, Linchao and He, Shengfeng and Liu, Yunhui and Liu, Wei},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={4006--4015},
  year={2019}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].