All Projects → ModelOriented → Dalex

ModelOriented / Dalex

Licence: gpl-3.0
moDel Agnostic Language for Exploration and eXplanation

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Dalex

Mlj.jl
A Julia machine learning framework
Stars: ✭ 982 (+23.52%)
Mutual labels:  data-science, predictive-modeling
Mlr
Machine Learning in R
Stars: ✭ 1,542 (+93.96%)
Mutual labels:  data-science, predictive-modeling
Modelstudio
📍 Interactive Studio for Explanatory Model Analysis
Stars: ✭ 163 (-79.5%)
Mutual labels:  interpretability, predictive-modeling
Data Science Wg
SF Brigade's Data Science Working Group.
Stars: ✭ 135 (-83.02%)
Mutual labels:  data-science, predictive-modeling
Imodels
Interpretable ML package 🔍 for concise, transparent, and accurate predictive modeling (sklearn-compatible).
Stars: ✭ 194 (-75.6%)
Mutual labels:  data-science, interpretability
Mli Resources
H2O.ai Machine Learning Interpretability Resources
Stars: ✭ 428 (-46.16%)
Mutual labels:  data-science, interpretability
Breakdown
Model Agnostics breakDown plots
Stars: ✭ 93 (-88.3%)
Mutual labels:  data-science, interpretability
Awesome Machine Learning Interpretability
A curated list of awesome machine learning interpretability resources.
Stars: ✭ 2,404 (+202.39%)
Mutual labels:  data-science, interpretability
Data Science Live Book
An open source book to learn data science, data analysis and machine learning, suitable for all ages!
Stars: ✭ 193 (-75.72%)
Mutual labels:  data-science, predictive-modeling
Facet
Human-explainable AI.
Stars: ✭ 269 (-66.16%)
Mutual labels:  data-science, interpretability
Interpretable machine learning with python
Examples of techniques for training interpretable ML models, explaining ML models, and debugging ML models for accuracy, discrimination, and security.
Stars: ✭ 530 (-33.33%)
Mutual labels:  data-science, interpretability
Mit 15 003 Data Science Tools
Study guides for MIT's 15.003 Data Science Tools
Stars: ✭ 743 (-6.54%)
Mutual labels:  data-science
Statistical Rethinking With Python And Pymc3
Python/PyMC3 port of the examples in " Statistical Rethinking A Bayesian Course with Examples in R and Stan" by Richard McElreath
Stars: ✭ 713 (-10.31%)
Mutual labels:  data-science
Reflow
A language and runtime for distributed, incremental data processing in the cloud
Stars: ✭ 706 (-11.19%)
Mutual labels:  data-science
Cookbook 2nd
IPython Cookbook, Second Edition, by Cyrille Rossant, Packt Publishing 2018
Stars: ✭ 704 (-11.45%)
Mutual labels:  data-science
Pycall.rb
Calling Python functions from the Ruby language
Stars: ✭ 773 (-2.77%)
Mutual labels:  data-science
Rows
A common, beautiful interface to tabular data, no matter the format
Stars: ✭ 739 (-7.04%)
Mutual labels:  data-science
Orchest
A new kind of IDE for Data Science.
Stars: ✭ 694 (-12.7%)
Mutual labels:  data-science
H1st
The AI Application Platform We All Need. Human AND Machine Intelligence. Based on experience building AI solutions at Panasonic: robotics predictive maintenance, cold-chain energy optimization, Gigafactory battery mfg, avionics, automotive cybersecurity, and more.
Stars: ✭ 697 (-12.33%)
Mutual labels:  data-science
Sciblog support
Support content for my blog
Stars: ✭ 694 (-12.7%)
Mutual labels:  data-science

moDel Agnostic Language for Exploration and eXplanation

R build status Coverage Status CRAN_Status_Badge Total Downloads DrWhy-eXtrAI

Python-check Supported Python versions PyPI version Downloads

Overview

Unverified black box model is the path to the failure. Opaqueness leads to distrust. Distrust leads to ignoration. Ignoration leads to rejection.

The DALEX package xrays any model and helps to explore and explain its behaviour, helps to understand how complex models are working. The main function explain() creates a wrapper around a predictive model. Wrapped models may then be explored and compared with a collection of local and global explainers. Recent developents from the area of Interpretable Machine Learning/eXplainable Artificial Intelligence.

The philosophy behind DALEX explanations is described in the Explanatory Model Analysis e-book. The DALEX package is a part of DrWhy.AI universe.

If you work with scikit-learn, keras, H2O, tidymodels, xgboost, mlr or mlr3, you may be interested in the DALEXtra package. It is an extension pack for DALEX with easy to use connectors to models created in these libraries.

Installation

The easiest way to get the R version of DALEX is to install it from CRAN

install.packages("DALEX")

The Python version of dalex is available on PyPI

pip install dalex -U

Learn more

Machine Learning models are widely used and have various applications in classification or regression tasks. Due to increasing computational power, availability of new data sources and new methods, ML models are more and more complex. Models created with techniques like boosting, bagging of neural networks are true black boxes. It is hard to trace the link between input variables and model outcomes. They are use because of high performance, but lack of interpretability is one of their weakest sides.

In many applications we need to know, understand or prove how input variables are used in the model and what impact do they have on final model prediction. DALEX is a set of tools that help to understand how complex models are working.

Resources

R package

Python package

Talks about DALEX

Why

76 years ago Isaac Asimov devised Three Laws of Robotics: 1) a robot may not injure a human being, 2) a robot must obey the orders given it by human beings and 3) A robot must protect its own existence. These laws impact discussion around Ethics of AI. Today’s robots, like cleaning robots, robotic pets or autonomous cars are far from being conscious enough to be under Asimov’s ethics.

Today we are surrounded by complex predictive algorithms used for decision making. Machine learning models are used in health care, politics, education, judiciary and many other areas. Black box predictive models have far larger influence on our lives than physical robots. Yet, applications of such models are left unregulated despite many examples of their potential harmfulness. See Weapons of Math Destruction by Cathy O'Neil for an excellent overview of potential problems.

It's clear that we need to control algorithms that may affect us. Such control is in our civic rights. Here we propose three requirements that any predictive model should fulfill.

  • Prediction's justifications. For every prediction of a model one should be able to understand which variables affect the prediction and how strongly. Variable attribution to final prediction.
  • Prediction's speculations. For every prediction of a model one should be able to understand how the model prediction would change if input variables were changed. Hypothesizing about what-if scenarios.
  • Prediction's validations For every prediction of a model one should be able to verify how strong are evidences that confirm this particular prediction.

There are two ways to comply with these requirements. One is to use only models that fulfill these conditions by design. White-box models like linear regression or decision trees. In many cases the price for transparency is lower performance. The other way is to use approximated explainers – techniques that find only approximated answers, but work for any black box model. Here we present such techniques.

Acknowledgments

Work on this package was financially supported by the NCN Opus grant 2016/21/B/ST6/02176 and NCN Opus grant 2017/27/B/ST6/0130.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].