All Projects → FKD → Similar Projects or Alternatives

126 Open source projects that are alternatives of or similar to FKD

CLMR
Official PyTorch implementation of Contrastive Learning of Musical Representations
Stars: ✭ 216 (+340.82%)
Mutual labels:  self-supervised-learning
exponential-moving-average-normalization
PyTorch implementation of EMAN for self-supervised and semi-supervised learning: https://arxiv.org/abs/2101.08482
Stars: ✭ 76 (+55.1%)
Mutual labels:  self-supervised-learning
MoTIS
Mobile(iOS) Text-to-Image search powered by multimodal semantic representation models(e.g., OpenAI's CLIP). Accepted at NAACL 2022.
Stars: ✭ 60 (+22.45%)
Mutual labels:  knowledge-distillation
latent-pose-reenactment
The authors' implementation of the "Neural Head Reenactment with Latent Pose Descriptors" (CVPR 2020) paper.
Stars: ✭ 132 (+169.39%)
Mutual labels:  self-supervised-learning
S2-BNN
S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural Networks via Guided Distribution Calibration (CVPR 2021)
Stars: ✭ 53 (+8.16%)
Mutual labels:  self-supervised-learning
esvit
EsViT: Efficient self-supervised Vision Transformers
Stars: ✭ 323 (+559.18%)
Mutual labels:  self-supervised-learning
SAN
[ECCV 2020] Scale Adaptive Network: Learning to Learn Parameterized Classification Networks for Scalable Input Images
Stars: ✭ 41 (-16.33%)
Mutual labels:  knowledge-distillation
CVPR21 PASS
PyTorch implementation of our CVPR2021 (oral) paper "Prototype Augmentation and Self-Supervision for Incremental Learning"
Stars: ✭ 55 (+12.24%)
Mutual labels:  self-supervised-learning
LabelRelaxation-CVPR21
Official PyTorch Implementation of Embedding Transfer with Label Relaxation for Improved Metric Learning, CVPR 2021
Stars: ✭ 37 (-24.49%)
Mutual labels:  knowledge-distillation
info-nce-pytorch
PyTorch implementation of the InfoNCE loss for self-supervised learning.
Stars: ✭ 160 (+226.53%)
Mutual labels:  self-supervised-learning
self6dpp
Self6D++: Occlusion-Aware Self-Supervised Monocular 6D Object Pose Estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 2021.
Stars: ✭ 45 (-8.16%)
Mutual labels:  self-supervised-learning
SelfGNN
A PyTorch implementation of "SelfGNN: Self-supervised Graph Neural Networks without explicit negative sampling" paper, which appeared in The International Workshop on Self-Supervised Learning for the Web (SSL'21) @ the Web Conference 2021 (WWW'21).
Stars: ✭ 24 (-51.02%)
Mutual labels:  self-supervised-learning
SoCo
[NeurIPS 2021 Spotlight] Aligning Pretraining for Detection via Object-Level Contrastive Learning
Stars: ✭ 125 (+155.1%)
Mutual labels:  self-supervised-learning
SimSiam
Exploring Simple Siamese Representation Learning
Stars: ✭ 52 (+6.12%)
Mutual labels:  self-supervised-learning
simpleAICV-pytorch-ImageNet-COCO-training
SimpleAICV:pytorch training example on ImageNet(ILSVRC2012)/COCO2017/VOC2007+2012 datasets.Include ResNet/DarkNet/RetinaNet/FCOS/CenterNet/TTFNet/YOLOv3/YOLOv4/YOLOv5/YOLOX.
Stars: ✭ 276 (+463.27%)
Mutual labels:  distillation
ViCC
[WACV'22] Code repository for the paper "Self-supervised Video Representation Learning with Cross-Stream Prototypical Contrasting", https://arxiv.org/abs/2106.10137.
Stars: ✭ 33 (-32.65%)
Mutual labels:  self-supervised-learning
Zero-shot Knowledge Distillation Pytorch
ZSKD with PyTorch
Stars: ✭ 26 (-46.94%)
Mutual labels:  knowledge-distillation
barlowtwins
Implementation of Barlow Twins paper
Stars: ✭ 84 (+71.43%)
Mutual labels:  self-supervised-learning
Self-Supervised-Embedding-Fusion-Transformer
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
Stars: ✭ 57 (+16.33%)
Mutual labels:  self-supervised-learning
EDCC-Palmprint-Recognition
EDCC: An efficient and accurate algorithm for palmprint recognition.
Stars: ✭ 103 (+110.2%)
Mutual labels:  efficient-algorithm
awesome-graph-self-supervised-learning-based-recommendation
A curated list of awesome graph & self-supervised-learning-based recommendation.
Stars: ✭ 37 (-24.49%)
Mutual labels:  self-supervised-learning
CCL
PyTorch Implementation on Paper [CVPR2021]Distilling Audio-Visual Knowledge by Compositional Contrastive Learning
Stars: ✭ 76 (+55.1%)
Mutual labels:  distillation
BossNAS
(ICCV 2021) BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search
Stars: ✭ 125 (+155.1%)
Mutual labels:  self-supervised-learning
awesome-efficient-gnn
Code and resources on scalable and efficient Graph Neural Networks
Stars: ✭ 498 (+916.33%)
Mutual labels:  knowledge-distillation
MutualGuide
Localize to Classify and Classify to Localize: Mutual Guidance in Object Detection
Stars: ✭ 97 (+97.96%)
Mutual labels:  knowledge-distillation
TCE
This repository contains the code implementation used in the paper Temporally Coherent Embeddings for Self-Supervised Video Representation Learning (TCE).
Stars: ✭ 51 (+4.08%)
Mutual labels:  self-supervised-learning
Revisiting-Contrastive-SSL
Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
Stars: ✭ 81 (+65.31%)
Mutual labels:  self-supervised-learning
bert-squeeze
🛠️ Tools for Transformers compression using PyTorch Lightning ⚡
Stars: ✭ 56 (+14.29%)
Mutual labels:  distillation
awesome-graph-self-supervised-learning
Awesome Graph Self-Supervised Learning
Stars: ✭ 805 (+1542.86%)
Mutual labels:  self-supervised-learning
mmrazor
OpenMMLab Model Compression Toolbox and Benchmark.
Stars: ✭ 644 (+1214.29%)
Mutual labels:  knowledge-distillation
neural-compressor
Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
Stars: ✭ 666 (+1259.18%)
Mutual labels:  knowledge-distillation
GCL
List of Publications in Graph Contrastive Learning
Stars: ✭ 25 (-48.98%)
Mutual labels:  self-supervised-learning
lossyless
Generic image compressor for machine learning. Pytorch code for our paper "Lossy compression for lossless prediction".
Stars: ✭ 81 (+65.31%)
Mutual labels:  self-supervised-learning
G-SimCLR
This is the code base for paper "G-SimCLR : Self-Supervised Contrastive Learning with Guided Projection via Pseudo Labelling" by Souradip Chakraborty, Aritra Roy Gosthipaty and Sayak Paul.
Stars: ✭ 69 (+40.82%)
Mutual labels:  self-supervised-learning
DisCont
Code for the paper "DisCont: Self-Supervised Visual Attribute Disentanglement using Context Vectors".
Stars: ✭ 13 (-73.47%)
Mutual labels:  self-supervised-learning
mae-scalable-vision-learners
A TensorFlow 2.x implementation of Masked Autoencoders Are Scalable Vision Learners
Stars: ✭ 54 (+10.2%)
Mutual labels:  self-supervised-learning
sesemi
supervised and semi-supervised image classification with self-supervision (Keras)
Stars: ✭ 43 (-12.24%)
Mutual labels:  self-supervised-learning
EATNN
This is our implementation of EATNN: Efficient Adaptive Transfer Neural Network (SIGIR 2019)
Stars: ✭ 23 (-53.06%)
Mutual labels:  efficient-algorithm
LD
Localization Distillation for Dense Object Detection (CVPR 2022)
Stars: ✭ 271 (+453.06%)
Mutual labels:  knowledge-distillation
pillar-motion
Self-Supervised Pillar Motion Learning for Autonomous Driving (CVPR 2021)
Stars: ✭ 98 (+100%)
Mutual labels:  self-supervised-learning
MLIC-KD-WSD
Multi-Label Image Classification via Knowledge Distillation from Weakly-Supervised Detection (ACM MM 2018)
Stars: ✭ 58 (+18.37%)
Mutual labels:  knowledge-distillation
bert-AAD
Adversarial Adaptation with Distillation for BERT Unsupervised Domain Adaptation
Stars: ✭ 27 (-44.9%)
Mutual labels:  knowledge-distillation
Distill-BERT-Textgen
Research code for ACL 2020 paper: "Distilling Knowledge Learned in BERT for Text Generation".
Stars: ✭ 121 (+146.94%)
Mutual labels:  knowledge-distillation
Awesome-Vision-Transformer-Collection
Variants of Vision Transformer and its downstream tasks
Stars: ✭ 124 (+153.06%)
Mutual labels:  self-supervised-learning
Self-Supervised-GANs
Tensorflow Implementation for paper "self-supervised generative adversarial networks"
Stars: ✭ 34 (-30.61%)
Mutual labels:  self-supervised-learning
MSF
Official code for "Mean Shift for Self-Supervised Learning"
Stars: ✭ 42 (-14.29%)
Mutual labels:  self-supervised-learning
ccgl
TKDE 22. CCCL: Contrastive Cascade Graph Learning.
Stars: ✭ 20 (-59.18%)
Mutual labels:  self-supervised-learning
AB distillation
Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons (AAAI 2019)
Stars: ✭ 105 (+114.29%)
Mutual labels:  knowledge-distillation
VQ-APC
Vector Quantized Autoregressive Predictive Coding (VQ-APC)
Stars: ✭ 34 (-30.61%)
Mutual labels:  self-supervised-learning
BYOL
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Stars: ✭ 102 (+108.16%)
Mutual labels:  self-supervised-learning
video-pace
code for our ECCV-2020 paper: Self-supervised Video Representation Learning by Pace Prediction
Stars: ✭ 95 (+93.88%)
Mutual labels:  self-supervised-learning
roberta-wwm-base-distill
this is roberta wwm base distilled model which was distilled from roberta wwm by roberta wwm large
Stars: ✭ 61 (+24.49%)
Mutual labels:  distillation
libai
LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training
Stars: ✭ 284 (+479.59%)
Mutual labels:  self-supervised-learning
simsiam-cifar10
Code to train the SimSiam model on cifar10 using PyTorch
Stars: ✭ 33 (-32.65%)
Mutual labels:  self-supervised-learning
GeDML
Generalized Deep Metric Learning.
Stars: ✭ 30 (-38.78%)
Mutual labels:  self-supervised-learning
video repres mas
code for CVPR-2019 paper: Self-supervised Spatio-temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics
Stars: ✭ 63 (+28.57%)
Mutual labels:  self-supervised-learning
MiniVox
Code for our ACML and INTERSPEECH papers: "Speaker Diarization as a Fully Online Bandit Learning Problem in MiniVox".
Stars: ✭ 15 (-69.39%)
Mutual labels:  self-supervised-learning
EfficientIR
人工智障本地图片检索工具 | An EfficientNet based image retrieval tool
Stars: ✭ 64 (+30.61%)
Mutual labels:  efficientnet-pytorch
newt
Natural World Tasks
Stars: ✭ 24 (-51.02%)
Mutual labels:  self-supervised-learning
mmselfsup
OpenMMLab Self-Supervised Learning Toolbox and Benchmark
Stars: ✭ 2,315 (+4624.49%)
Mutual labels:  self-supervised-learning
1-60 of 126 similar projects