All Projects → hushell → sib_meta_learn

hushell / sib_meta_learn

Licence: other
Code of Empirical Bayes Transductive Meta-Learning with Synthetic Gradients

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to sib meta learn

simple-cnaps
Source codes for "Improved Few-Shot Visual Classification" (CVPR 2020), "Enhancing Few-Shot Image Classification with Unlabelled Examples" (WACV 2022), and "Beyond Simple Meta-Learning: Multi-Purpose Models for Multi-Domain, Active and Continual Few-Shot Learning" (Neural Networks 2022 - in submission)
Stars: ✭ 88 (+57.14%)
Mutual labels:  meta-learning, few-shot-learning, mini-imagenet
LearningToCompare-Tensorflow
Tensorflow implementation for paper: Learning to Compare: Relation Network for Few-Shot Learning.
Stars: ✭ 17 (-69.64%)
Mutual labels:  meta-learning, few-shot-learning
LibFewShot
LibFewShot: A Comprehensive Library for Few-shot Learning.
Stars: ✭ 629 (+1023.21%)
Mutual labels:  meta-learning, few-shot-learning
awesome-few-shot-meta-learning
awesome few shot / meta learning papers
Stars: ✭ 44 (-21.43%)
Mutual labels:  meta-learning, few-shot-learning
Awesome-Few-shot
Awesome Few-shot learning
Stars: ✭ 50 (-10.71%)
Mutual labels:  meta-learning, few-shot-learning
Transferlearning
Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习
Stars: ✭ 8,481 (+15044.64%)
Mutual labels:  meta-learning, few-shot-learning
FSL-Mate
FSL-Mate: A collection of resources for few-shot learning (FSL).
Stars: ✭ 1,346 (+2303.57%)
Mutual labels:  meta-learning, few-shot-learning
Meta-TTS
Official repository of https://arxiv.org/abs/2111.04040v1
Stars: ✭ 69 (+23.21%)
Mutual labels:  meta-learning, few-shot-learning
CDFSL-ATA
[IJCAI 2021] Cross-Domain Few-Shot Classification via Adversarial Task Augmentation
Stars: ✭ 21 (-62.5%)
Mutual labels:  meta-learning, few-shot-learning
MeTAL
Official PyTorch implementation of "Meta-Learning with Task-Adaptive Loss Function for Few-Shot Learning" (ICCV2021 Oral)
Stars: ✭ 24 (-57.14%)
Mutual labels:  meta-learning, few-shot-learning
Learning-To-Compare-For-Text
Learning To Compare For Text , Few shot learning in text classification
Stars: ✭ 38 (-32.14%)
Mutual labels:  meta-learning, few-shot-learning
Meta Learning Papers
Meta Learning / Learning to Learn / One Shot Learning / Few Shot Learning
Stars: ✭ 2,420 (+4221.43%)
Mutual labels:  meta-learning, few-shot-learning
FUSION
PyTorch code for NeurIPSW 2020 paper (4th Workshop on Meta-Learning) "Few-Shot Unsupervised Continual Learning through Meta-Examples"
Stars: ✭ 18 (-67.86%)
Mutual labels:  meta-learning, few-shot-learning
attMPTI
[CVPR 2021] Few-shot 3D Point Cloud Semantic Segmentation
Stars: ✭ 118 (+110.71%)
Mutual labels:  few-shot-learning
pykale
Knowledge-Aware machine LEarning (KALE): accessible machine learning from multiple sources for interdisciplinary research, part of the 🔥PyTorch ecosystem
Stars: ✭ 381 (+580.36%)
Mutual labels:  meta-learning
FOCAL-ICLR
Code for FOCAL Paper Published at ICLR 2021
Stars: ✭ 35 (-37.5%)
Mutual labels:  meta-learning
tensorflow-maml
TensorFlow 2.0 implementation of MAML.
Stars: ✭ 79 (+41.07%)
Mutual labels:  meta-learning
sinkhorn-label-allocation
Sinkhorn Label Allocation is a label assignment method for semi-supervised self-training algorithms. The SLA algorithm is described in full in this ICML 2021 paper: https://arxiv.org/abs/2102.08622.
Stars: ✭ 49 (-12.5%)
Mutual labels:  few-shot-learning
few shot dialogue generation
Dialogue Knowledge Transfer Networks (DiKTNet)
Stars: ✭ 24 (-57.14%)
Mutual labels:  few-shot-learning
HiCE
Code for ACL'19 "Few-Shot Representation Learning for Out-Of-Vocabulary Words"
Stars: ✭ 56 (+0%)
Mutual labels:  few-shot-learning

[Update on 31/03/2020] This repository has been merged to Xfer.

[ICLR 2020] Synthetic information bottleneck for transductive meta-learning

This repo contains the implementation of the synthetic information bottleneck algorithm for few-shot classification on Mini-ImageNet, which is used in our ICLR 2020 paper Empirical Bayes Transductive Meta-Learning with Synthetic Gradients.

If our code is helpful for your research, please consider citing:

@inproceedings{
    Hu2020Empirical,
    title={Empirical Bayes Transductive Meta-Learning with Synthetic Gradients},
    author={Shell Xu Hu and Pablo Moreno and Yang Xiao and Xi Shen and Guillaume Obozinski and Neil Lawrence and Andreas Damianou},
    booktitle={International Conference on Learning Representations (ICLR)},
    year={2020},
    url={https://openreview.net/forum?id=Hkg-xgrYvH}
}

Authors of the code

Shell Xu Hu, Xi Shen and Yang Xiao

Dependencies

The code is tested under Pytorch > 1.0 + Python 3.6 environment with extra packages:

pip install -r requirements.txt

How to use the code on Mini-ImageNet?

Step 0: Download Mini-ImageNet dataset

cd data
bash download_miniimagenet.sh 
cd ..

Step 1 (optional): train a WRN-28-10 feature network (aka backbone)

The weights of the feature network is downloaded in step 0, but you may also train from scracth by running

python main_feat.py --outDir miniImageNet_WRN_60Epoch --cuda --dataset miniImageNet --nbEpoch 60

Step 2: Meta-training on Mini-ImageNet, e.g., 5-way-1-shot:

python main.py --config config/miniImageNet_1shot.yaml --seed 100 --gpu 0

Step 3: Meta-testing on Mini-ImageNet with a checkpoint:

python main.py --config config/miniImageNet_1shot.yaml --seed 100 --gpu 0 --ckpt cache/miniImageNet_1shot_K3_seed100/outputs_xx.xxx/netSIBBestxx.xxx.pth

Mini-ImageNet Results (LAST ckpt)

Setup 5-way-1-shot 5-way-5-shot
SIB (K=3) 70.700% ± 0.585% 80.045% ± 0.363%
SIB (K=5) 70.494 ± 0.619% 80.192% ± 0.372%

CIFAR-FS Results (LAST ckpt)

Setup 5-way-1-shot 5-way-5-shot
SIB (K=3) 79.763% ± 0.577% 85.721% ± 0.369%
SIB (K=5) 79.627 ± 0.593% 85.590% ± 0.375%
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].