All Projects → reiinakano → Xcessiv

reiinakano / Xcessiv

Licence: apache-2.0
A web-based application for quick, scalable, and automated hyperparameter tuning and stacked ensembling in Python.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Xcessiv

Autogluon
AutoGluon: AutoML for Text, Image, and Tabular Data
Stars: ✭ 3,920 (+212.35%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, ensemble-learning, automated-machine-learning
Hyperactive
A hyperparameter optimization and data collection toolbox for convenient and fast prototyping of machine-learning models.
Stars: ✭ 182 (-85.5%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, automated-machine-learning
Auto ml
[UNMAINTAINED] Automated machine learning for analytics & production
Stars: ✭ 1,559 (+24.22%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, automated-machine-learning
Mljar Supervised
Automated Machine Learning Pipeline with Feature Engineering and Hyper-Parameters Tuning 🚀
Stars: ✭ 961 (-23.43%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, automated-machine-learning
Tpot
A Python Automated Machine Learning tool that optimizes machine learning pipelines using genetic programming.
Stars: ✭ 8,378 (+567.57%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, automated-machine-learning
Lale
Library for Semi-Automated Data Science
Stars: ✭ 198 (-84.22%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization, automated-machine-learning
Nni
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+752.43%)
Mutual labels:  data-science, hyperparameter-optimization, automated-machine-learning
Dat8
General Assembly's 2015 Data Science course in Washington, DC
Stars: ✭ 1,516 (+20.8%)
Mutual labels:  data-science, scikit-learn, ensemble-learning
Auptimizer
An automatic ML model optimization tool.
Stars: ✭ 166 (-86.77%)
Mutual labels:  data-science, hyperparameter-optimization, automated-machine-learning
mindware
An efficient open-source AutoML system for automating machine learning lifecycle, including feature engineering, neural architecture search, and hyper-parameter tuning.
Stars: ✭ 34 (-97.29%)
Mutual labels:  hyperparameter-optimization, ensemble-learning, automated-machine-learning
Machinejs
[UNMAINTAINED] Automated machine learning- just give it a data file! Check out the production-ready version of this project at ClimbsRocks/auto_ml
Stars: ✭ 412 (-67.17%)
Mutual labels:  data-science, scikit-learn, automated-machine-learning
Featuretools
An open source python library for automated feature engineering
Stars: ✭ 5,891 (+369.4%)
Mutual labels:  data-science, scikit-learn, automated-machine-learning
Auto Sklearn
Automated Machine Learning with scikit-learn
Stars: ✭ 5,916 (+371.39%)
Mutual labels:  scikit-learn, hyperparameter-optimization, automated-machine-learning
Hyperparameter hunter
Easy hyperparameter optimization and automatic result saving across machine learning algorithms and libraries
Stars: ✭ 648 (-48.37%)
Mutual labels:  data-science, scikit-learn, hyperparameter-optimization
Crime Analysis
Association Rule Mining from Spatial Data for Crime Analysis
Stars: ✭ 20 (-98.41%)
Mutual labels:  data-science, scikit-learn
Autodl
Automated Deep Learning without ANY human intervention. 1'st Solution for AutoDL [email protected]
Stars: ✭ 854 (-31.95%)
Mutual labels:  data-science, automated-machine-learning
Python for ml
brief introduction to Python for machine learning
Stars: ✭ 29 (-97.69%)
Mutual labels:  data-science, scikit-learn
Model Describer
model-describer : Making machine learning interpretable to humans
Stars: ✭ 22 (-98.25%)
Mutual labels:  data-science, scikit-learn
Mlcourse.ai
Open Machine Learning Course
Stars: ✭ 7,963 (+534.5%)
Mutual labels:  data-science, scikit-learn
Machinelearningcourse
A collection of notebooks of my Machine Learning class written in python 3
Stars: ✭ 35 (-97.21%)
Mutual labels:  data-science, scikit-learn

Xcessiv

PyPI license PyPI Build Status

Xcessiv is a tool to help you create the biggest, craziest, and most excessive stacked ensembles you can think of.

Stacked ensembles are simple in theory. You combine the predictions of smaller models and feed those into another model. However, in practice, implementing them can be a major headache.

Xcessiv holds your hand through all the implementation details of creating and optimizing stacked ensembles so you're free to fully define only the things you care about.

The Xcessiv process

Define your base learners and performance metrics

define_base_learner

Keep track of hundreds of different model-hyperparameter combinations

list_base_learner

Effortlessly choose your base learners and create an ensemble with the click of a button

ensemble

Features

  • Fully define your data source, cross-validation process, relevant metrics, and base learners with Python code
  • Any model following the Scikit-learn API can be used as a base learner
  • Task queue based architecture lets you take full advantage of multiple cores and embarrassingly parallel hyperparameter searches
  • Direct integration with TPOT for automated pipeline construction
  • Automated hyperparameter search through Bayesian optimization
  • Easy management and comparison of hundreds of different model-hyperparameter combinations
  • Automatic saving of generated secondary meta-features
  • Stacked ensemble creation in a few clicks
  • Automated ensemble construction through greedy forward model selection
  • Export your stacked ensemble as a standalone Python file to support multiple levels of stacking

Installation and Documentation

You can find installation instructions and detailed documentation hosted here.

FAQ

Where does Xcessiv fit in the machine learning process?

Xcessiv fits in the model building part of the process after data preparation and feature engineering. At this point, there is no universally acknowledged way of determining which algorithm will work best for a particular dataset (see No Free Lunch Theorem), and while heuristic optimization methods do exist, things often break down into trial and error as you try to find the best model-hyperparameter combinations.

Stacking is an almost surefire method to improve performance beyond that of any single model, however, the complexity of proper implementation often makes it impractical to apply them in practice outside of Kaggle competitions. Xcessiv aims to make the construction of stacked ensembles as painless as possible and lower the barrier for entry.

I don't care about fancy stacked ensembles and what not, should I still use Xcessiv?

Absolutely! Even without the ensembling functionality, the sheer amount of utility provided by keeping track of the performance of hundreds, and even thousands of ML models and hyperparameter combinations is a huge boon.

How does Xcessiv generate meta-features for stacking?

You can choose whether to generate meta-features through cross-validation (stacked generalization) or with a holdout set (blending). You can read about these two methods and a lot more about stacked ensembles in the Kaggle Ensembling Guide. It's a great article and provides most of the inspiration for this project.

Contributing

Xcessiv is in its very early stages and needs the open-source community to guide it along.

There are many ways to contribute to Xcessiv. You could report a bug, suggest a feature, submit a pull request, improve documentation, and many more.

If you would like to contribute something, please visit our Contributor Guidelines.

Project Status

Xcessiv is currently in alpha and is unstable. Future versions are not guaranteed to be backwards-compatible with current project files.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].