All Projects → OATML → Bdl Benchmarks

OATML / Bdl Benchmarks

Licence: apache-2.0
Bayesian Deep Learning Benchmarks

Projects that are alternatives of or similar to Bdl Benchmarks

Datasets For Recommender Systems
This is a repository of a topic-centric public data sources in high quality for Recommender Systems (RS)
Stars: ✭ 564 (-2.42%)
Mutual labels:  jupyter-notebook
Data Science Notes
数据科学的笔记以及资料搜集
Stars: ✭ 6,072 (+950.52%)
Mutual labels:  jupyter-notebook
Gym Trading
Environment for reinforcement-learning algorithmic trading models
Stars: ✭ 574 (-0.69%)
Mutual labels:  jupyter-notebook
Bert score
BERT score for text generation
Stars: ✭ 568 (-1.73%)
Mutual labels:  jupyter-notebook
Cs231
My corrections for the Standford class assingments CS231n - Convolutional Neural Networks for Visual Recognition
Stars: ✭ 570 (-1.38%)
Mutual labels:  jupyter-notebook
Wgan Tensorflow
a tensorflow implementation of WGAN
Stars: ✭ 572 (-1.04%)
Mutual labels:  jupyter-notebook
Machine Learning Specialization
Stars: ✭ 566 (-2.08%)
Mutual labels:  jupyter-notebook
Business Machine Learning
A curated list of practical business machine learning (BML) and business data science (BDS) applications for Accounting, Customer, Employee, Legal, Management and Operations (by @firmai)
Stars: ✭ 575 (-0.52%)
Mutual labels:  jupyter-notebook
Cleverhans
An adversarial example library for constructing attacks, building defenses, and benchmarking both
Stars: ✭ 5,356 (+826.64%)
Mutual labels:  jupyter-notebook
Ml Design Patterns
Source code accompanying O'Reilly book: Machine Learning Design Patterns
Stars: ✭ 566 (-2.08%)
Mutual labels:  jupyter-notebook
Tutorials
CatBoost tutorials repository
Stars: ✭ 563 (-2.6%)
Mutual labels:  jupyter-notebook
Dogs vs cats
猫狗大战
Stars: ✭ 570 (-1.38%)
Mutual labels:  jupyter-notebook
Codereader
跟大咖一起读源码
Stars: ✭ 572 (-1.04%)
Mutual labels:  jupyter-notebook
Python For Data Analysis
An introduction to data science using Python and Pandas with Jupyter notebooks
Stars: ✭ 564 (-2.42%)
Mutual labels:  jupyter-notebook
Agegenderdeeplearning
Stars: ✭ 575 (-0.52%)
Mutual labels:  jupyter-notebook
Ttur
Two time-scale update rule for training GANs
Stars: ✭ 567 (-1.9%)
Mutual labels:  jupyter-notebook
Deeplearning.ai
Some work of Andrew Ng's course on Coursera
Stars: ✭ 572 (-1.04%)
Mutual labels:  jupyter-notebook
Functional Zoo
PyTorch and Tensorflow functional model definitions
Stars: ✭ 577 (-0.17%)
Mutual labels:  jupyter-notebook
T2t Vit
Stars: ✭ 575 (-0.52%)
Mutual labels:  jupyter-notebook
Eng Edu
Stars: ✭ 573 (-0.87%)
Mutual labels:  jupyter-notebook

Bayesian Deep Learning Benchmarks

This repository is no longer being updated.

Please refer to the Diabetic Retinopathy Detection implementation in Google's 'uncertainty-baselines' repo for up-to-date baseline implementations.

Overview

In order to make real-world difference with Bayesian Deep Learning (BDL) tools, the tools must scale to real-world settings. And for that we, the research community, must be able to evaluate our inference tools (and iterate quickly) with real-world benchmark tasks. We should be able to do this without necessarily worrying about application-specific domain knowledge, like the expertise often required in medical applications for example. We require benchmarks to test for inference robustness, performance, and accuracy, in addition to cost and effort of development. These benchmarks should be at a variety of scales, ranging from toy MNIST-scale benchmarks for fast development cycles, to large data benchmarks which are truthful to real-world applications, capturing their constraints.

Our BDL benchmarks should

  • provide a transparent, modular and consistent interface for the evaluation of deep probabilistic models on a variety of downstream tasks;
  • rely on expert-driven metrics of uncertainty quality (actual applications making use of BDL uncertainty in the real-world), but abstract away the expert-knowledge and eliminate the boilerplate steps necessary for running experiments on real-world datasets;
  • make it easy to compare the performance of new models against well tuned baselines, models that have been well-adopted by the machine learning community, under a fair and realistic setting (e.g., computational resources, model sizes, datasets);
  • provide reference implementations of baseline models (e.g., Monte Carlo Dropout Inference, Mean Field Variational Inference, Deep Ensembles), enabling rapid prototyping and easy development of new tools;
  • be independent of specific deep learning frameworks (e.g., not depend on TensorFlow, PyTorch, etc.), and integrate with the SciPy ecosystem (i.e., NumPy, Pandas, Matplotlib). Benchmarks are framework-agnostic, while baselines are framework-dependent.

In this repo we strive to provide such well-needed benchmarks for the BDL community, and collect and maintain new baselines and benchmarks contributed by the community. A colab notebook demonstrating the MNIST-like workflow of our benchmarks is available here.

We highly encourage you to contribute your models as new baselines for others to compete against, as well as contribute new benchmarks for others to evaluate their models on!

List of Benchmarks

Bayesian Deep Learning Benchmarks (BDL Benchmarks or bdlb for short), is an open-source framework that aims to bridge the gap between the design of deep probabilistic machine learning models and their application to real-world problems. Our currently supported benchmarks are:

Installation

BDL Benchmarks is shipped as a PyPI package (Python3 compatible) installable as:

pip3 install git+https://github.com/OATML/bdl-benchmarks.git

The data downloading and preparation is benchmark-specific, and you can follow the relevant guides at baselines/<benchmark>/README.md (e.g. baselines/diabetic_retinopathy_diagnosis/README.md).

Examples

For example, the Diabetic Retinopathy Diagnosis benchmark comes with several baselines, including MC Dropout, MFVI, Deep Ensembles, and more. These models are trained with images of blood vessels in the eye:

The models try to predict diabetic retinopathy, and use their uncertainty for prescreening (sending patients the model is uncertain about to an expert for further examination). When you implement a new model, you can easily benchmark your model against existing baseline results provided in the repo, and generate plots using expert metrics (such as the AUC of retained data when referring 50% most uncertain patients to an expert):

You can even play with a colab notebook to see the workflow of the benchmark, and contribute your model for others to benchmark against.

Cite as

Please cite individual benchmarks when you use these, as well as the baselines you compare against. For the Diabetic Retinopathy Diagnosis benchmark please see here.

Acknowledgements

The repository is developed and maintained by the Oxford Applied and Theoretical Machine Learning group, with sponsorship from:

Contact Us

Email us for questions at [email protected], or submit any issues to improve the framework.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].