All Projects → ariG23498 → mae-scalable-vision-learners

ariG23498 / mae-scalable-vision-learners

Licence: MIT license
A TensorFlow 2.x implementation of Masked Autoencoders Are Scalable Vision Learners

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to mae-scalable-vision-learners

SimMIM
This is an official implementation for "SimMIM: A Simple Framework for Masked Image Modeling".
Stars: ✭ 717 (+1227.78%)
Mutual labels:  self-supervised-learning, masked-image-modeling
G-SimCLR
This is the code base for paper "G-SimCLR : Self-Supervised Contrastive Learning with Guided Projection via Pseudo Labelling" by Souradip Chakraborty, Aritra Roy Gosthipaty and Sayak Paul.
Stars: ✭ 69 (+27.78%)
Mutual labels:  self-supervised-learning, tensorflow2
mmselfsup
OpenMMLab Self-Supervised Learning Toolbox and Benchmark
Stars: ✭ 2,315 (+4187.04%)
Mutual labels:  self-supervised-learning
Awesome-Vision-Transformer-Collection
Variants of Vision Transformer and its downstream tasks
Stars: ✭ 124 (+129.63%)
Mutual labels:  self-supervised-learning
Awesome-Tensorflow2
基于Tensorflow2开发的优秀扩展包及项目
Stars: ✭ 45 (-16.67%)
Mutual labels:  tensorflow2
peax
Peax is a tool for interactive visual pattern search and exploration in epigenomic data based on unsupervised representation learning with autoencoders
Stars: ✭ 63 (+16.67%)
Mutual labels:  autoencoder
2D-and-3D-Deep-Autoencoder
Convolutional AutoEncoder application on MRI images
Stars: ✭ 57 (+5.56%)
Mutual labels:  autoencoder
tensorflow-tabnet
Improved TabNet for TensorFlow
Stars: ✭ 49 (-9.26%)
Mutual labels:  tensorflow2
pillar-motion
Self-Supervised Pillar Motion Learning for Autonomous Driving (CVPR 2021)
Stars: ✭ 98 (+81.48%)
Mutual labels:  self-supervised-learning
info-nce-pytorch
PyTorch implementation of the InfoNCE loss for self-supervised learning.
Stars: ✭ 160 (+196.3%)
Mutual labels:  self-supervised-learning
awesome-graph-self-supervised-learning-based-recommendation
A curated list of awesome graph & self-supervised-learning-based recommendation.
Stars: ✭ 37 (-31.48%)
Mutual labels:  self-supervised-learning
SESF-Fuse
SESF-Fuse: An Unsupervised Deep Model for Multi-Focus Image Fusion
Stars: ✭ 47 (-12.96%)
Mutual labels:  autoencoder
QuantumSpeech-QCNN
IEEE ICASSP 21 - Quantum Convolution Neural Networks for Speech Processing and Automatic Speech Recognition
Stars: ✭ 71 (+31.48%)
Mutual labels:  tensorflow2
imagenet-autoencoder
Autoencoder trained on ImageNet Using Torch 7
Stars: ✭ 18 (-66.67%)
Mutual labels:  autoencoder
face-mask-detection-tf2
A face mask detection using ssd with simplified Mobilenet and RFB or Pelee in Tensorflow 2.1. Training on your own dataset. Can be converted to kmodel and run on the edge device of k210
Stars: ✭ 72 (+33.33%)
Mutual labels:  tensorflow2
seq3
Source code for the NAACL 2019 paper "SEQ^3: Differentiable Sequence-to-Sequence-to-Sequence Autoencoder for Unsupervised Abstractive Sentence Compression"
Stars: ✭ 121 (+124.07%)
Mutual labels:  autoencoder
Continuous-Image-Autoencoder
Deep learning image autoencoder that not depends on image resolution
Stars: ✭ 20 (-62.96%)
Mutual labels:  autoencoder
Video-Compression-Net
A new approach to video compression by refining the shortcomings of conventional approach and substituting each traditional component with their neural network counterpart. Our proposed work consists of motion estimation, compression and compensation and residue compression, learned end-to-end to minimize the rate-distortion trade off. The whole…
Stars: ✭ 20 (-62.96%)
Mutual labels:  autoencoder
Reducing-the-Dimensionality-of-Data-with-Neural-Networks
Implementation of G. E. Hinton and R. R. Salakhutdinov's Reducing the Dimensionality of Data with Neural Networks (Tensorflow)
Stars: ✭ 34 (-37.04%)
Mutual labels:  autoencoder
datascienv
datascienv is package that helps you to setup your environment in single line of code with all dependency and it is also include pyforest that provide single line of import all required ml libraries
Stars: ✭ 53 (-1.85%)
Mutual labels:  tensorflow2

Masked Autoencoders Are Scalable Vision Learners

Open In Colab

A TensorFlow implementation of Masked Autoencoders Are Scalable Vision Learners [1]. Our implementation of the proposed method is available in mae-pretraining.ipynb notebook. It includes evaluation with linear probing as well. Furthermore, the notebook can be fully executed on Google Colab. Our main objective is to present the core idea of the proposed method in a minimal and readable manner. We have also prepared a blog for getting started with Masked Autoencoder easily.


With just 100 epochs of pre-training and a fairly lightweight and asymmetric Autoencoder architecture we achieve 49.33%% accuracy with linear probing on the CIFAR-10 dataset. Our training logs and encoder weights are released in Weights and Logs. For comparison, we took the encoder architecture and trained it from scratch (refer to regular-classification.ipynb) in a fully supervised manner. This gave us ~76% test top-1 accuracy.

We note that with further hyperparameter tuning and more epochs of pre-training, we can achieve a better performance with linear-probing. Below we present some more results:

Config Masking
proportion
LP
performance
Encoder weights
& logs
Encoder & decoder layers: 3 & 1
Batch size: 256
0.6 44.25% Link
Do 0.75 46.84% Link
Encoder & decoder layers: 6 & 2
Batch size: 256
0.75 48.16% Link
Encoder & decoder layers: 9 & 3
Batch size: 256
Weight deacy: 1e-5
0.75 49.33% Link

LP denotes linear-probing. Config is mostly based on what we define in the hyperparameters section of this notebook: mae-pretraining.ipynb.

Notes

Acknowledgements

References

[1] Masked Autoencoders Are Scalable Vision Learners; He et al.; arXiv 2021; https://arxiv.org/abs/2111.06377.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].