All Projects β†’ Cloud-CV β†’ Evalai

Cloud-CV / Evalai

Licence: other
☁️ πŸš€ πŸ“Š πŸ“ˆ Evaluating state of the art in AI

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Evalai

Avalanche
Avalanche: a End-to-End Library for Continual Learning.
Stars: ✭ 151 (-86.11%)
Mutual labels:  artificial-intelligence, evaluation, reproducible-research
Pycm
Multi-class confusion matrix library in Python
Stars: ✭ 1,076 (-1.01%)
Mutual labels:  artificial-intelligence, ai, evaluation
Enmf
This is our implementation of ENMF: Efficient Neural Matrix Factorization (TOIS. 38, 2020). This also provides a fair evaluation of existing state-of-the-art recommendation models.
Stars: ✭ 96 (-91.17%)
Mutual labels:  evaluation, reproducible-research, reproducibility
open-solution-googleai-object-detection
Open solution to the Google AI Object Detection Challenge 🍁
Stars: ✭ 46 (-95.77%)
Mutual labels:  challenge, reproducible-research, reproducibility
Grand Challenge.org
A platform for end-to-end development of machine learning solutions in biomedical imaging
Stars: ✭ 89 (-91.81%)
Mutual labels:  ai, django, reproducible-research
Xai
XAI - An eXplainability toolbox for machine learning
Stars: ✭ 596 (-45.17%)
Mutual labels:  artificial-intelligence, ai, evaluation
Awesome Ai Books
Some awesome AI related books and pdfs for learning and downloading, also apply some playground models for learning
Stars: ✭ 855 (-21.34%)
Mutual labels:  artificial-intelligence, ai
Steppy Toolkit
Curated set of transformers that make your work with steppy faster and more effective πŸ”­
Stars: ✭ 21 (-98.07%)
Mutual labels:  reproducible-research, reproducibility
Ai Programmer
Using artificial intelligence and genetic algorithms to automatically write programs. Tutorial: http://www.primaryobjects.com/cms/article149
Stars: ✭ 948 (-12.79%)
Mutual labels:  artificial-intelligence, ai
Letslearnai.github.io
Lets Learn AI
Stars: ✭ 33 (-96.96%)
Mutual labels:  artificial-intelligence, ai
Goapy
Goal-Oriented Action Planning implementation in Python
Stars: ✭ 33 (-96.96%)
Mutual labels:  artificial-intelligence, ai
Advanced Directives With Angular Js
Advanced Directives with Angular JS - Code for the Screencast
Stars: ✭ 33 (-96.96%)
Mutual labels:  challenge, angularjs
Autodl
Automated Deep Learning without ANY human intervention. 1'st Solution for AutoDL [emailΒ protected]
Stars: ✭ 854 (-21.44%)
Mutual labels:  artificial-intelligence, ai
Machine Learning Experiments
πŸ€– Interactive Machine Learning experiments: πŸ‹οΈmodels training + 🎨models demo
Stars: ✭ 841 (-22.63%)
Mutual labels:  artificial-intelligence, ai
Machine Learning Open Source
Monthly Series - Machine Learning Top 10 Open Source Projects
Stars: ✭ 943 (-13.25%)
Mutual labels:  artificial-intelligence, ai
Sliding puzzle
Swift implementation of the Sliding Puzzle game with Iterative Deepening A* AI Solver.
Stars: ✭ 25 (-97.7%)
Mutual labels:  artificial-intelligence, ai
Tetrisai
The old school Tetris game in addition with an AI that learns evolutionary how to play this game
Stars: ✭ 22 (-97.98%)
Mutual labels:  artificial-intelligence, ai
Quant Finance Resources
Courses, Articles and many more which can help beginners or professionals.
Stars: ✭ 36 (-96.69%)
Mutual labels:  artificial-intelligence, ai
Dvc
πŸ¦‰Data Version Control | Git for Data & Models | ML Experiments Management
Stars: ✭ 9,004 (+728.33%)
Mutual labels:  ai, reproducibility
Clarifai Apple Sdk
Artificial Intelligence with a Vision
Stars: ✭ 46 (-95.77%)
Mutual labels:  artificial-intelligence, ai


Join the chat at https://gitter.im/Cloud-CV/EvalAI Build Status codecov Coverage Status Code style: black Requirements Status Code Climate Documentation Status Twitter Follow

EvalAI is an open source platform for evaluating and comparing machine learning (ML) and artificial intelligence (AI) algorithms at scale.

In recent years, it has become increasingly difficult to compare an algorithm solving a given task with other existing approaches. These comparisons suffer from minor differences in algorithm implementation, use of non-standard dataset splits and different evaluation metrics. By providing a central leaderboard and submission interface, we make it easier for researchers to reproduce the results mentioned in the paper and perform reliable & accurate quantitative analysis. By providing swift and robust backends based on map-reduce frameworks that speed up evaluation on the fly, EvalAI aims to make it easier for researchers to reproduce results from technical papers and perform reliable and accurate analyses.

Features

  • Custom evaluation protocols and phases: We allow creation of an arbitrary number of evaluation phases and dataset splits, compatibility using any programming language, and organizing results in both public and private leaderboards.

  • Remote evaluation: Certain large-scale challenges need special compute capabilities for evaluation. If the challenge needs extra computational power, challenge organizers can easily add their own cluster of worker nodes to process participant submissions while we take care of hosting the challenge, handling user submissions, and maintaining the leaderboard.

  • Evaluation inside environments: EvalAI lets participants submit code for their agent in the form of docker images which are evaluated against test environments on the evaluation server. During evaluation, the worker fetches the image, test environment, and the model snapshot and spins up a new container to perform evaluation.

  • CLI support: evalai-cli is designed to extend the functionality of the EvalAI web application to your command line to make the platform more accessible and terminal-friendly.

  • Portability: EvalAI is designed with keeping in mind scalability and portability of such a system from the very inception of the idea. Most of the components rely heavily on open-source technologies – Docker, Django, Node.js, and PostgreSQL.

  • Faster evaluation: We warm-up the worker nodes at start-up by importing the challenge code and pre-loading the dataset in memory. We also split the dataset into small chunks that are simultaneously evaluated on multiple cores. These simple tricks result in faster evaluation and reduces the evaluation time by an order of magnitude in some cases.

Platform Comparison

Features OpenML TopCoder Kaggle CrowdAI ParlAI Codalab EvalAI
AI challenge hosting βœ–οΈ βœ… βœ… βœ… βœ–οΈ βœ… βœ…
Custom metrics βœ–οΈ βœ–οΈ βœ–οΈ βœ… βœ… βœ… βœ…
Multiple phases/splits βœ–οΈ βœ–οΈ βœ–οΈ βœ… βœ–οΈ βœ… βœ…
Open source βœ… βœ–οΈ βœ–οΈ βœ… βœ… βœ… βœ…
Remote evaluation βœ–οΈ βœ–οΈ βœ–οΈ βœ–οΈ βœ… βœ… βœ…
Human evaluation βœ–οΈ βœ–οΈ βœ–οΈ βœ–οΈ βœ… βœ–οΈ βœ…
Evaluation in Environments βœ–οΈ βœ–οΈ βœ–οΈ βœ… βœ–οΈ βœ–οΈ βœ…

Goal

Our ultimate goal is to build a centralized platform to host, participate and collaborate in AI challenges organized around the globe and we hope to help in benchmarking progress in AI.

Installation instructions

Setting up EvalAI on your local machine is really easy. You can setup EvalAI using docker: The steps are:

  1. Install docker and docker-compose on your machine.

  2. Get the source code on to your machine via git.

    git clone https://github.com/Cloud-CV/EvalAI.git evalai && cd evalai
    
  3. Build and run the Docker containers. This might take a while.

    docker-compose up --build
    
  4. That's it. Open web browser and hit the URL http://127.0.0.1:8888. Three users will be created by default which are listed below -

    SUPERUSER- username: admin password: password
    HOST USER- username: host password: password
    PARTICIPANT USER- username: participant password: password

If you are facing any issue during installation, please see our common errors during installation page.

Citing EvalAI

If you are using EvalAI for hosting challenges, please cite the following technical report:

@article{EvalAI,
    title   =  {EvalAI: Towards Better Evaluation Systems for AI Agents},
    author  =  {Deshraj Yadav and Rishabh Jain and Harsh Agrawal and Prithvijit
                Chattopadhyay and Taranjeet Singh and Akash Jain and Shiv Baran
                Singh and Stefan Lee and Dhruv Batra},
    year    =  {2019},
    volume  =  arXiv:1902.03570
}

Team

EvalAI is currently maintained by Rishabh Jain, Deshraj Yadav, and Ram Ramrakhya. A non-exhaustive list of other major contributors includes: Akash Jain, Taranjeet Singh, Shiv Baran Singh, Harsh Agarwal, Prithvijit Chattopadhyay, Devi Parikh and Dhruv Batra.

Contribution guidelines

If you are interested in contributing to EvalAI, follow our contribution guidelines.

Contributors

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].