All Projects → txsun1997 → awesome-early-exiting

txsun1997 / awesome-early-exiting

Licence: MIT license
A curated list of Early Exiting papers, benchmarks, and misc.

Projects that are alternatives of or similar to awesome-early-exiting

Personal-Emotional-Stylized-Dialog
A Paper List for Personalized, Emotional, and stylized Dialog
Stars: ✭ 112 (+133.33%)
Mutual labels:  paper-list
Emotion-Cause-Analysis-Papers
Collection of papers on Emotion Cause Analysis
Stars: ✭ 63 (+31.25%)
Mutual labels:  paper-list
Nrlpapers
Must-read papers on network representation learning (NRL) / network embedding (NE)
Stars: ✭ 2,503 (+5114.58%)
Mutual labels:  paper-list
Gnnpapers
Must-read papers on graph neural networks (GNN)
Stars: ✭ 12,293 (+25510.42%)
Mutual labels:  paper-list
Traffic-Prediction-Open-Code-Summary
Summary of open source code for deep learning models in the field of traffic prediction
Stars: ✭ 58 (+20.83%)
Mutual labels:  paper-list
DeepTimeSeriesModel
A paper list for Time series modelling, including prediciton and anomaly detection
Stars: ✭ 78 (+62.5%)
Mutual labels:  paper-list
Literatures-on-GNN-Acceleration
A reading list for deep graph learning acceleration.
Stars: ✭ 50 (+4.17%)
Mutual labels:  paper-list
Natural-language-understanding-papers
NLU: domain-intent-slot; text2SQL
Stars: ✭ 77 (+60.42%)
Mutual labels:  paper-list
Knowledge Graph Wander
A collection of papers, codes, projects, tutorials ... for Knowledge Graph and other NLP methods
Stars: ✭ 26 (-45.83%)
Mutual labels:  paper-list
causal-ml
Must-read papers and resources related to causal inference and machine (deep) learning
Stars: ✭ 387 (+706.25%)
Mutual labels:  paper-list
awesome-point-cloud-deep-learning
Paper list of deep learning on point clouds.
Stars: ✭ 39 (-18.75%)
Mutual labels:  paper-list

Awesome Early Exiting Papers

A curated list of Early Exiting papers, benchmarks, and misc. Currently, the resources listed in this repo are mainly in the field of natural language processing. Adding papers of early exiting in other fields (e.g. computer vision) is also welcome. (This repo is constantly updated.)

Early Exiting is an efficient technique that trains a deep model with multiple injected internal classifiers (exits) such that test samples can selectively exit instead of passing through the entire model.

Contents

Introduction

Early exiting methods usually add internal classifiers to different layers of a model. By training these internal classifiers with the ground truth, the model has a chance to predict the correct label and exit earlier during inference. Current early exiting methods can be divided into two branches: 1. Dynamic methods and 2. Static methods.

Dynamic Early Exiting methods typically have two steps: (a) Training the internal classifiers on downstream tasks to make them capable of making predictions, (b) Designing an exiting strategy to decide whether to exit early or continue to the next layer.

Static Early Exiting methods assign each test sample to a specific layer by learning the difficulty of the samples or heuristically pre-defining the assignment of samples.

Benchmarks

ELUE Benchmark

Towards Efficient NLP: A Standard Evaluation and A Strong Baseline. NAACL 2022.

Xiangyang Liu*, Tianxiang Sun*, Junliang He, Lingling Wu, Xinyu Zhang, Hao Jiang, Zhao Cao, Xuanjing Huang, Xipeng Qiu. [pdf][website]

Papers

Surveys

  1. Dynamic Neural Networks: A Survey. Preprint Feb 2021.

    Yizeng Han, Gao Huang, Shiji Song, Le Yang, Honghui Wang, Yulin Wang. [pdf]

  2. A Survey on Dynamic Neural Networks for Natural Language Processing. Preprint Feb 2022.

    Canwen Xu, Julian McAuley. [pdf]

  3. Split Computing and Early Exiting for Deep Learning Applications: Survey and Research Challenges. Preprint Mar 2021.

    Yoshitomo Matsubara, Marco Levorato, Francesco Restuccia. [pdf]

  4. Adaptive Inference through Early-Exit Networks: Design, Challenges and Directions. EMDL 2021.

    Stefanos Laskaridis, Alexandros Kouris, Nicholas D. Lane. [pdf]

Dynamic Methods

  1. DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference. ACL 2020.

    Ji Xin, Raphael Tang, Jaejun Lee, Yaoliang Yu, and Jimmy Lin. [pdf]

  2. The Right Tool for the Job: Matching Model and Instance Complexities. ACL 2020.

    Roy Schwartz, Gabriel Stanovsky, Swabha Swayamdipta, Jesse Dodge, and Noah A. Smith. [pdf]

  3. FastBERT: a Self-distilling BERT with Adaptive Inference Time. ACL 2020.

    Weijie Liu, Peng Zhou, Zhiruo Wang, Zhe Zhao, Haotang Deng, and Qi Ju. [pdf]

  4. Early Exiting BERT for Efficient Document Ranking. ACL 2020.

    Ji Xin, Rodrigo Nogueira, Yaoliang Yu, Jimmy Lin. [pdf]

  5. BERT Loses Patience: Fast and Robust Inference with Early Exit. NeurIPS 2020.

    Wangchunshu Zhou, Canwen X u, Tao Ge, Julian McAuley, Ke Xu, Furu Wei. [pdf]

  6. DynaBERT: Dynamic BERT with Adaptive Width and Depth. NeurlPS 2020.

    Lu Hou, Zhiqi Huang, Lifeng Shang, Xin Jiang, Xiao Chen, Qun Liu. [pdf]

  7. A Global Past-Future Early Exit Method for Accelerating Inference of Pre-trained Language Models. NAACL 2021.

    Kaiyuan Liao, Yi Zhang, Xuancheng Ren, Qi Su, Xu Sun, Bin He. [pdf]

  8. RomeBERT: Robust Training of Multi-Exit BERT. Preprint Jan 2021.

    Shijie Geng, Peng Gao, Zuohui Fu, Yongfeng Zhang. [pdf]

  9. BERxiT: Early Exiting for BERT with Better Fine-Tuning and Extension to Regression. EACL 2021.

    Ji Xin, Raphael Tang, Yaoliang Yu, Jimmy Lin. [pdf]

  10. Accelerating BERT Inference for Sequence Labeling via Early-Exit. ACL 2021.

    Xiaonan Li, Yunfan Shao, Tianxiang Sun, Hang Yan, Xipeng Qiu, Xuanjing Huang. [pdf]

  11. LeeBERT: Learned Early Exit for BERT with Cross-Level Optimization. ACL 2021.

    Wei Zhu. [pdf]

  12. TR-BERT: Dynamic Token Reduction for Accelerating BERT Inference. ACL 2021.

    Deming Ye, Yankai Lin, Yufei Huang, Maosong Sun. [pdf]

  13. EBERT: Efficient BERT Inference with Dynamic Structured Pruning. ACL 2021.

    Zejian Liu, Fanrong Li, Gang Li, Jian Cheng. [pdf]

  14. Class Means as an Early Exit Decision Mechanism. Preprint Mar 2021.

    Alperen Gormez, Erdem Koyuncu. [pdf]

  15. Early Exiting with Ensemble Internal Classifiers. Preprint May 2021.

    Tianxiang Sun, Yunhua Zhou, Xiangyang Liu, Xinyu Zhang, Hao Jiang, Zhao Cao, Xuanjing Huang, Xipeng Qiu. [pdf]

  16. ELBERT: Fast Albert with Confidence-Window Based Early Exit. Preprint Jul 2021.

    Keli Xie, Siyuan Lu, Meiqi Wang, Zhongfeng Wang. [pdf]

  17. CascadeBERT: Accelerating Inference of Pre-trained Language Models via Calibrated Complete Models Cascade. EMNLP 2021.

    Lei Li, Yankai Lin, Deli Chen, Shuhuai Ren, Peng Li, Jie Zhou, Xu Sun. [pdf]

  18. Consistent Accelerated Inference via Confident Adaptive Transformers. EMNLP 2021.

    Tal Schuster, Adam Fisch, Tommi Jaakkola, Regina Barzilay. [pdf]

  19. DACT-BERT: Differentiable Adaptive Computation Time for an Efficient BERT Inference. Preprint Sep 2021.

    Cristóbal Eyzaguirre, Felipe del Río, Vladimir Araujo, Álvaro Soto. [pdf]

  20. Towards Efficient NLP: A Standard Evaluation and A Strong Baseline. NAACL 2022.

    Xiangyang Liu*, Tianxiang Sun*, Junliang He, Lingling Wu, Xinyu Zhang, Hao Jiang, Zhao Cao, Xuanjing Huang, Xipeng Qiu. [pdf]

Static Methods

  1. Depth-Adaptive Transformer. ICLR 2020.

    Maha Elbayad, Jiatao Gu, Edouard Grave, Michael Auli. [pdf]

  2. Faster Depth-Adaptive Transformers. AAAI 2021.

    Yijin Liu, Fandong Meng, Jie Zhou, Yufeng Chen, Jinan Xu. [pdf]

  3. A Simple Hash-Based Early Exiting Approach For Language Understanding and Generation. Findings of ACL 2022.

    Tianxiang Sun, Xiangyang Liu, Wei Zhu, Zhichao Geng, Lingling Wu, Yilong He, Yuan Ni, Guotong Xie, Xuanjing Huang, Xipeng Qiu [pdf]

Contributing

👍🎉 First off, thanks for taking the time to contribute! 🎉👍

Steps to contribute:

  • Make your awesome changes
  • Submit pull request; if you add a new entry, please give a very brief explanation why you think it should be added.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].