All Projects → sseung0703 → Knowledge_distillation_via_TF2.0

sseung0703 / Knowledge_distillation_via_TF2.0

Licence: other
The codes for recent knowledge distillation algorithms and benchmark results via TF2.0 low-level API

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Knowledge distillation via TF2.0

CARLA
CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms
Stars: ✭ 166 (+90.8%)
Mutual labels:  benchmark, tensorflow2
glassbench
A micro-benchmark framework to use with cargo bench
Stars: ✭ 29 (-66.67%)
Mutual labels:  benchmark
benchbox
🧀 The Benchmark Testing Box
Stars: ✭ 19 (-78.16%)
Mutual labels:  benchmark
BIRL
BIRL: Benchmark on Image Registration methods with Landmark validations
Stars: ✭ 66 (-24.14%)
Mutual labels:  benchmark
benchmark-malloc
Trace memory allocations and collect stats
Stars: ✭ 18 (-79.31%)
Mutual labels:  benchmark
iohk-monitoring-framework
This framework provides logging, benchmarking and monitoring.
Stars: ✭ 27 (-68.97%)
Mutual labels:  benchmark
fizzboom
Benchmark to compare async web server + interpreter + web client implementations across various languages
Stars: ✭ 46 (-47.13%)
Mutual labels:  benchmark
FewCLUE
FewCLUE 小样本学习测评基准,中文版
Stars: ✭ 251 (+188.51%)
Mutual labels:  benchmark
IGUANA
IGUANA is a benchmark execution framework for querying HTTP endpoints and CLI Applications such as Triple Stores. Contact: [email protected]
Stars: ✭ 22 (-74.71%)
Mutual labels:  benchmark
MVP Benchmark
MVP Benchmark for Multi-View Partial Point Cloud Completion and Registration
Stars: ✭ 74 (-14.94%)
Mutual labels:  benchmark
Meta-SelfLearning
Meta Self-learning for Multi-Source Domain Adaptation: A Benchmark
Stars: ✭ 157 (+80.46%)
Mutual labels:  benchmark
react-native-css-in-js-benchmarks
CSS in JS Benchmarks for React Native
Stars: ✭ 46 (-47.13%)
Mutual labels:  benchmark
json-serialization-benchmarking
Miscellaneous benchmarks for JSON serialization on JVM/Android
Stars: ✭ 48 (-44.83%)
Mutual labels:  benchmark
SLE-GAN
Towards Faster and Stabilized GAN Training for High-fidelity Few-shot Image Synthesis
Stars: ✭ 53 (-39.08%)
Mutual labels:  tensorflow2
MaskedFaceRepresentation
Masked face recognition focuses on identifying people using their facial features while they are wearing masks. We introduce benchmarks on face verification based on masked face images for the development of COVID-safe protocols in airports.
Stars: ✭ 17 (-80.46%)
Mutual labels:  benchmark
model optimizer
Model optimizer used in Adlik.
Stars: ✭ 22 (-74.71%)
Mutual labels:  knowledge-distillation
hyperspectral-soilmoisture-dataset
Hyperspectral and soil-moisture data from a field campaign based on a soil sample. Karlsruhe (Germany), 2017.
Stars: ✭ 23 (-73.56%)
Mutual labels:  benchmark
autobench
Benchmark your application on CI
Stars: ✭ 16 (-81.61%)
Mutual labels:  benchmark
tls-perf
TLS handshakes benchnarking tool
Stars: ✭ 18 (-79.31%)
Mutual labels:  benchmark
criterion-compare-action
⚡️📊 Compare the performance of Rust project branches
Stars: ✭ 16 (-81.61%)
Mutual labels:  benchmark

Knowledge_distillation_via_TF2.0

  • Now, I'm fixing all the issues and refining the codes. It will be easier to understand how each KD works than before.
  • Algorithms are already implemented again, but they should be checked more with hyperparameter tuning.
    • the algorithms which have experimental results have been confirmed.
  • This Repo. will be upgraded version of my previous benchmark Repo. (link)

Implemented Knowledge Distillation Methods

Defined knowledge by the neural response of the hidden layer or the output layer of the network

Experimental Results

  • I use WResNet-40-4 and WResNet-16-4 as the teacher and the student network, respectively.
  • All the algorithm is trained in the sample configuration, which is described in "train_w_distillation.py", and only each algorithm's hyper-parameters are tuned. I tried only several times to get acceptable performance, which means that my experimental results are perhaps not optimal.
  • Although some of the algorithms used soft-logits parallelly in the paper, I used only the proposed knowledge distillation algorithm to make a fair comparison.
  • Initialization-based methods give a far higher performance in the start point but a poor performance in the last point due to overfitting. Therefore, initialized students must have a regularization algorithm, such as Soft-logits.

Training/Validation accuracy

Full Dataset 50% Dataset 25% Dataset 10% Dataset
Methods Accuracy Last Accuracy Last Accuracy Last Accuracy
Teacher 78.59 - - -
Student 76.25 - - -
Soft_logits 76.57 - - -
FitNet 75.78 - - -
AT 78.14 - - -
FSP 76.08 - - -
DML - - - -
KD_SVD - - - -
FT 77.30 - - -
AB 76.52 - - -
RKD 77.69 - - -
VID - - - -
MHGD - - - -
CO 78.54 - - -

Plan to do

  • Check all the algorithms.
  • do experiments.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].