All Projects → project-codeflare → codeflare

project-codeflare / codeflare

Licence: Apache-2.0 license
Simplifying the definition and execution, scaling and deployment of pipelines on the cloud.

Programming Languages

Jupyter Notebook
11667 projects
python
139335 projects - #7 most used programming language
Dockerfile
14818 projects

Projects that are alternatives of or similar to codeflare

Automl alex
State-of-the art Automated Machine Learning python library for Tabular Data
Stars: ✭ 132 (-19.02%)
Mutual labels:  sklearn, hyperparameter-optimization, automl
Ray
An open source framework that provides a simple, universal API for building distributed applications. Ray is packaged with RLlib, a scalable reinforcement learning library, and Tune, a scalable hyperparameter tuning library.
Stars: ✭ 18,547 (+11278.53%)
Mutual labels:  hyperparameter-optimization, ray, automl
Mlmodels
mlmodels : Machine Learning and Deep Learning Model ZOO for Pytorch, Tensorflow, Keras, Gluon models...
Stars: ✭ 145 (-11.04%)
Mutual labels:  sklearn, hyperparameter-optimization, automl
Hyperparameter hunter
Easy hyperparameter optimization and automatic result saving across machine learning algorithms and libraries
Stars: ✭ 648 (+297.55%)
Mutual labels:  sklearn, hyperparameter-optimization
Mlprimitives
Primitives for machine learning and data science.
Stars: ✭ 46 (-71.78%)
Mutual labels:  pipelines, automl
Transmogrifai
TransmogrifAI (pronounced trăns-mŏgˈrə-fī) is an AutoML library for building modular, reusable, strongly typed machine learning workflows on Apache Spark with minimal hand-tuning
Stars: ✭ 2,084 (+1178.53%)
Mutual labels:  pipelines, automl
Onepanel
The open and extensible integrated development environment (IDE) for computer vision with built-in modules for model building, automated labeling, data processing, model training, hyperparameter tuning and workflow orchestration.
Stars: ✭ 428 (+162.58%)
Mutual labels:  pipelines, workflows
Auto ts
Automatically build ARIMA, SARIMAX, VAR, FB Prophet and XGBoost Models on Time Series data sets with a Single Line of Code. Now updated with Dask to handle millions of rows.
Stars: ✭ 195 (+19.63%)
Mutual labels:  sklearn, automl
Igel
a delightful machine learning tool that allows you to train, test, and use models without writing code
Stars: ✭ 2,956 (+1713.5%)
Mutual labels:  sklearn, automl
dolphinnext
A graphical user interface for distributed data processing of high throughput genomics
Stars: ✭ 92 (-43.56%)
Mutual labels:  pipelines, workflows
foreshadow
An automatic machine learning system
Stars: ✭ 29 (-82.21%)
Mutual labels:  sklearn, automl
cli
Polyaxon Core Client & CLI to streamline MLOps
Stars: ✭ 18 (-88.96%)
Mutual labels:  hyperparameter-optimization, workflows
ultraopt
Distributed Asynchronous Hyperparameter Optimization better than HyperOpt. 比HyperOpt更强的分布式异步超参优化库。
Stars: ✭ 93 (-42.94%)
Mutual labels:  hyperparameter-optimization, automl
Argo Events
Event-driven workflow automation framework
Stars: ✭ 821 (+403.68%)
Mutual labels:  pipelines, workflows
Hungabunga
HungaBunga: Brute-Force all sklearn models with all parameters using .fit .predict!
Stars: ✭ 614 (+276.69%)
Mutual labels:  sklearn, automl
Rain
Framework for large distributed pipelines
Stars: ✭ 645 (+295.71%)
Mutual labels:  pipelines, workflows
maggy
Distribution transparent Machine Learning experiments on Apache Spark
Stars: ✭ 83 (-49.08%)
Mutual labels:  hyperparameter-optimization, automl
modules
Repository to host tool-specific module files for the Nextflow DSL2 community!
Stars: ✭ 94 (-42.33%)
Mutual labels:  pipelines, workflows
Hypernets
A General Automated Machine Learning framework to simplify the development of End-to-end AutoML toolkits in specific domains.
Stars: ✭ 221 (+35.58%)
Mutual labels:  hyperparameter-optimization, automl
torchx
TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and support for E2E production ML pipelines when you're ready.
Stars: ✭ 165 (+1.23%)
Mutual labels:  pipelines, ray

License Build Status PyPI Downloads Documentation Status GitHub

UPDATE
CodeFlare is evolving! Check our updates for CodeFlare Pipelines and related contributions to Ray Workflows under Ray project.

Scale complex AI/ML pipelines anywhere

CodeFlare is a framework to simplify the integration, scaling and acceleration of complex multi-step analytics and machine learning pipelines on the cloud.

Its main features are:

  • Pipeline execution and scaling: CodeFlare Pipelines faciltates the definition and parallel execution of pipelines. It unifies pipeline workflows across multiple frameworks while providing nearly optimal scale-out parallelism on pipelined computations.
  • Deploy and integrate anywhere: CodeFlare simplifies deployment and integration by enabling a serverless user experience with the integration with Red Hat OpenShift and IBM Cloud Code Engine and providing adapters and connectors to make it simple to load data and connect to data services.

Release status

This project is under active development. See the Documentation for design descriptions and the latest version of the APIs.

Quick start

Run in your laptop

Instaling locally

CodeFlare can be installed from PyPI.

Prerequisites:

We recommend installing Python 3.8.6 using pyenv. You can find here recommended steps to set up the Python environment.

Install from PyPI:

pip3 install --upgrade pip          # CodeFlare requires pip >21.0
pip3 install --upgrade codeflare

Alternatively, you can also build locally with:

git clone https://github.com/project-codeflare/codeflare.git
cd codeflare
pip3 install --upgrade pip
pip3 install .

Using Docker

You can try CodeFlare by running the docker image from Docker Hub:

  • projectcodeflare/codeflare:latest has the latest released version installed.

The command below starts the most recent development build in a clean environment:

docker run --rm -it -p 8888:8888 projectcodeflare/codeflare:latest

It should produce an output similar to the one below, where you can then find the URL to run CodeFlare from a Jupyter notebook in your local browser.

[I <time_stamp> ServerApp] Jupyter Server <version> is running at:
...
[I <time_stamp> ServerApp]     http://127.0.0.1:8888/lab

Using Binder service

You can try out some of CodeFlare features using the My Binder service.

Click on the link below to try CodeFlare, on a sandbox environment, without having to install anything.

Binder

Pipeline execution and scaling

UPDATE
As of January 2022, this feature is now built on Ray Workflows with parts of it in Ray core and the rest in a DAG contribution repository. Please follow these links to contribute to CodeFlare Pipelines.

CodeFlare Pipelines reimagined pipelines to provide a more intuitive API for the data scientist to create AI/ML pipelines, data workflows, pre-processing, post-processing tasks, and many more which can scale from a laptop to a cluster seamlessly.

See the API documentation here, and reference use case documentation in the Examples section.

A set of reference examples are provided as executable notebooks.

To run examples, if you haven't done so yet, clone the CodeFlare project with:

git clone https://github.com/project-codeflare/codeflare.git

Example notebooks require JupyterLab, which can be installed with:

pip3 install --upgrade jupyterlab

Use the command below to run locally:

jupyter-lab codeflare/notebooks/<example_notebook>

The step above should automatically open a browser window and connect to a running Jupyter server.

If you are using any one of the recommended cloud based deployments (see below), examples are found in the codeflare/notebooks directory in the container image. The examples can be executed directly from the Jupyter environment.

As a first example of the API usage, see the sample pipeline.

For an example of how CodeFlare Pipelines can be used to scale out common machine learning problems, see the grid search example. It shows how hyperparameter optimization for a reference pipeline can be scaled and accelerated with both task and data parallelism.

Deploy and integrate anywhere

Unleash the power of pipelines by seamlessly scaling on the cloud. CodeFlare can be deployed on any Kubernetes-based platform, including IBM Cloud Code Engine and Red Hat OpenShift Container Platform.

  • IBM Cloud Code Engine for detailed instructions on how to run CodeFlare on a serverless platform.
  • Red Hat OpenShift for detailed instructions on how to run CodeFlare on OpenShift Container Platform.

Contributing

Join us in making CodeFlare Better! We encourage you to take a look at our Contributing page.

Blog

CodeFlare related blogs are published on our Medium publication.

License

CodeFlare is an open-source project with an Apache 2.0 license.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].