All Projects → plynx-team → Plynx

plynx-team / Plynx

Licence: apache-2.0
PLynx is a domain agnostic platform for managing reproducible experiments and data-oriented workflows.

Programming Languages

javascript
184084 projects - #8 most used programming language

Projects that are alternatives of or similar to Plynx

Drake
An R-focused pipeline toolkit for reproducibility and high-performance computing
Stars: ✭ 1,301 (+577.6%)
Mutual labels:  data-science, reproducibility, workflow
Drake Examples
Example workflows for the drake R package
Stars: ✭ 57 (-70.31%)
Mutual labels:  data-science, reproducibility, workflow
Bit
A tool for component-driven application development.
Stars: ✭ 14,443 (+7422.4%)
Mutual labels:  collaboration, workflow, distributed
Dvc
🦉Data Version Control | Git for Data & Models | ML Experiments Management
Stars: ✭ 9,004 (+4589.58%)
Mutual labels:  collaboration, data-science, reproducibility
Targets
Function-oriented Make-like declarative workflows for R
Stars: ✭ 293 (+52.6%)
Mutual labels:  data-science, reproducibility, workflow
Open Solution Value Prediction
Open solution to the Santander Value Prediction Challenge 🐠
Stars: ✭ 34 (-82.29%)
Mutual labels:  data-science, reproducibility
Vds
Verteego Data Suite
Stars: ✭ 9 (-95.31%)
Mutual labels:  data-science, workflow
Attaca
Robust, distributed version control for large files.
Stars: ✭ 41 (-78.65%)
Mutual labels:  data-science, distributed
Mlbox
MLBox is a powerful Automated Machine Learning python library.
Stars: ✭ 1,199 (+524.48%)
Mutual labels:  data-science, distributed
Flyte
Accelerate your ML and Data workflows to production. Flyte is a production grade orchestration system for your Data and ML workloads. It has been battle tested at Lyft, Spotify, freenome and others and truly open-source.
Stars: ✭ 1,242 (+546.88%)
Mutual labels:  data-science, workflow
Maze
Maze Applied Reinforcement Learning Framework
Stars: ✭ 85 (-55.73%)
Mutual labels:  data-science, distributed
Plz
Say the magic word 😸
Stars: ✭ 31 (-83.85%)
Mutual labels:  experiments, reproducibility
Steppy Toolkit
Curated set of transformers that make your work with steppy faster and more effective 🔭
Stars: ✭ 21 (-89.06%)
Mutual labels:  data-science, reproducibility
Steppy
Lightweight, Python library for fast and reproducible experimentation 🔬
Stars: ✭ 119 (-38.02%)
Mutual labels:  data-science, reproducibility
Nni
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+5471.88%)
Mutual labels:  data-science, distributed
Cape Python
Collaborate on privacy-preserving policy for data science projects in Pandas and Apache Spark
Stars: ✭ 125 (-34.9%)
Mutual labels:  collaboration, data-science
Titanoboa
Titanoboa makes complex workflows easy. It is a low-code workflow orchestration platform for JVM - distributed, highly scalable and fault tolerant.
Stars: ✭ 787 (+309.9%)
Mutual labels:  distributed, workflow
Renku
The Renku Project provides a platform and tools for reproducible and collaborative data analysis.
Stars: ✭ 141 (-26.56%)
Mutual labels:  collaboration, reproducibility
H1st
The AI Application Platform We All Need. Human AND Machine Intelligence. Based on experience building AI solutions at Panasonic: robotics predictive maintenance, cold-chain energy optimization, Gigafactory battery mfg, avionics, automotive cybersecurity, and more.
Stars: ✭ 697 (+263.02%)
Mutual labels:  collaboration, data-science
Prefect
The easiest way to automate your data
Stars: ✭ 7,956 (+4043.75%)
Mutual labels:  data-science, workflow

Plynx

Plynx

Website shields.io Website shields.io Website shields.io

Website and demo: plynx.com.

Docs: docs.

PLynx is a domain agnostic platform for managing reproducible experiments and data-oriented workflows.

Features

Workflow Editor

Interactive User Interface. You can clone successful experiment and reuse it or create one from scratch. PLynx manages history of the experiments and they can be reproduced.

interactive graph editor

Operations editor

Operations can be customized independently from the platform. Users can define their own Operations or reuse existing ones.

online editor

Monitor progress

Track the progress of the experiment. Each of intermediate operations produce results that you can inspect.

monitor progress

Preview the results

View the results right in the browser.

results preview

Scalable architecture

Execution Engine is based on scalable pub/sub model. Each Worker performs their jobs independently from each other and can publish subtasks back to the queue. Executers are plugins themselves and can support multiple scenarios from "compile to binary code" to "deploy and serve" to "run in a cluster using as many distributed workers as possible".

Scalable architecture

Requirements

Plugins work on python3. User Interface is based on React. PLynx is using MongoDB as a primary metadata storage. In order to meet diverse data storage requirements, its own storage plugins to store the artifacts. It supports multiple data storages such as AWS S3, Google Cloud Storage and traditional filesystems.

In order to reduce complexity we recommend to install docker and run make command to start local cluster.

Get started

Usage

Make sure you install docker first. Get started with Docker

tl;dr

git clone https://github.com/plynx-team/plynx.git   # Clone the repo

cd plynx

cp template_config.yaml config.yaml                 # Make a copy of a config
make up                                             # to start production services

Then go to http://localhost:3001

By default it will start the following services:

  • MongoDB instance
  • PLynx User Interface
  • Backend
  • Several workers

Other make commands:

  • make build - build all docker images.
  • make run_tests - build docker images and run the tests.
  • make up - run the services locally.
  • make dev - run developer version of PLynx.

Config

Most of the parameters can be set in command line when the services are called. For example:

$ plynx exec --help
usage: -c exec [-h] [-v] -f FILENAME [--storage-prefix STORAGE_PREFIX]

optional arguments:
  -h, --help            show this help message and exit
  -v, --verbose         Set logging output more verbose
  -f FILENAME, --filename FILENAME
                        Path to file
  --storage-prefix STORAGE_PREFIX
                        Storage prefix

But we recommend to store the config in a separate file.

Plynx config location is can be set in env variable PLYNX_CONFIG_PATH. Default value is ./config.yaml.

External links

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].