All Projects → abhinavsagar → ICC-2019-WC-prediction

abhinavsagar / ICC-2019-WC-prediction

Licence: MIT license
Predicting the winner of 2019 cricket world cup using random forest algorithm

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to ICC-2019-WC-prediction

Trajectory-Analysis-and-Classification-in-Python-Pandas-and-Scikit-Learn
Formed trajectories of sets of points.Experimented on finding similarities between trajectories based on DTW (Dynamic Time Warping) and LCSS (Longest Common SubSequence) algorithms.Modeled trajectories as strings based on a Grid representation.Benchmarked KNN, Random Forest, Logistic Regression classification algorithms to classify efficiently t…
Stars: ✭ 41 (+0%)
Mutual labels:  random-forest, scikit-learn, logistic-regression
AIML-Projects
Projects I completed as a part of Great Learning's PGP - Artificial Intelligence and Machine Learning
Stars: ✭ 85 (+107.32%)
Mutual labels:  random-forest, logistic-regression, support-vector-machines
100 Days Of Ml Code
100 Days of ML Coding
Stars: ✭ 33,641 (+81951.22%)
Mutual labels:  scikit-learn, logistic-regression, support-vector-machines
Machine Learning With Python
Python code for common Machine Learning Algorithms
Stars: ✭ 3,334 (+8031.71%)
Mutual labels:  random-forest, logistic-regression, knn-classification
Tensorflow Ml Nlp
텐서플로우와 머신러닝으로 시작하는 자연어처리(로지스틱회귀부터 트랜스포머 챗봇까지)
Stars: ✭ 176 (+329.27%)
Mutual labels:  random-forest, logistic-regression
Machine Learning Is All You Need
🔥🌟《Machine Learning 格物志》: ML + DL + RL basic codes and notes by sklearn, PyTorch, TensorFlow, Keras & the most important, from scratch!💪 This repository is ALL You Need!
Stars: ✭ 173 (+321.95%)
Mutual labels:  random-forest, logistic-regression
Orange3
🍊 📊 💡 Orange: Interactive data analysis
Stars: ✭ 3,152 (+7587.8%)
Mutual labels:  random-forest, scikit-learn
Isl Python
Solutions to labs and excercises from An Introduction to Statistical Learning, as Jupyter Notebooks.
Stars: ✭ 108 (+163.41%)
Mutual labels:  random-forest, logistic-regression
STOCK-RETURN-PREDICTION-USING-KNN-SVM-GUASSIAN-PROCESS-ADABOOST-TREE-REGRESSION-AND-QDA
Forecast stock prices using machine learning approach. A time series analysis. Employ the Use of Predictive Modeling in Machine Learning to Forecast Stock Return. Approach Used by Hedge Funds to Select Tradeable Stocks
Stars: ✭ 94 (+129.27%)
Mutual labels:  random-forest, logistic-regression
click-through-rate-prediction
📈 Click-Through Rate Prediction using Logistic Regression and Tree Algorithms
Stars: ✭ 60 (+46.34%)
Mutual labels:  random-forest, logistic-regression
Machine-Learning-Models
In This repository I made some simple to complex methods in machine learning. Here I try to build template style code.
Stars: ✭ 30 (-26.83%)
Mutual labels:  random-forest, logistic-regression
Machine Learning Models
Decision Trees, Random Forest, Dynamic Time Warping, Naive Bayes, KNN, Linear Regression, Logistic Regression, Mixture Of Gaussian, Neural Network, PCA, SVD, Gaussian Naive Bayes, Fitting Data to Gaussian, K-Means
Stars: ✭ 160 (+290.24%)
Mutual labels:  random-forest, logistic-regression
Emlearn
Machine Learning inference engine for Microcontrollers and Embedded devices
Stars: ✭ 154 (+275.61%)
Mutual labels:  random-forest, scikit-learn
Textclassification
several methods for text classification
Stars: ✭ 180 (+339.02%)
Mutual labels:  random-forest, logistic-regression
Machine Learning With Python
Practice and tutorial-style notebooks covering wide variety of machine learning techniques
Stars: ✭ 2,197 (+5258.54%)
Mutual labels:  random-forest, scikit-learn
R-stats-machine-learning
Misc Statistics and Machine Learning codes in R
Stars: ✭ 33 (-19.51%)
Mutual labels:  random-forest, support-vector-machines
dlime experiments
In this work, we propose a deterministic version of Local Interpretable Model Agnostic Explanations (LIME) and the experimental results on three different medical datasets shows the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME).
Stars: ✭ 21 (-48.78%)
Mutual labels:  random-forest, scikit-learn
text-classification-cn
中文文本分类实践,基于搜狗新闻语料库,采用传统机器学习方法以及预训练模型等方法
Stars: ✭ 81 (+97.56%)
Mutual labels:  scikit-learn, logistic-regression
25daysinmachinelearning
I will update this repository to learn Machine learning with python with statistics content and materials
Stars: ✭ 53 (+29.27%)
Mutual labels:  random-forest, logistic-regression
Text classification
Text Classification Algorithms: A Survey
Stars: ✭ 1,276 (+3012.2%)
Mutual labels:  random-forest, logistic-regression

ICC 2019 WC prediction

Predicting the winner of 2019 cricket world cup using random forest algorithm

Citing

If you find this code useful in your research, please consider citing the blog:

@misc{sagaricc,
  Author = {Abhinav Sagar},
  Title = {ICC 2019 Cricket World Cup Prediction using Machine Learning},
  Year = {2019},
  Journal = {Towards Data Science},
}

WC Logo

I used Machine Learning to make a model using scikit-learn, pandas, numpy, seaborn and matplotlib to predict the results of ICC 2019 Cricket World Cup.

Check out the corresponding medium blog post https://towardsdatascience.com/icc-2019-cricket-world-cup-prediction-using-machine-learning-7c42d848ace1.

Goals

  • Use Machine Learning to predict the winner of ICC 2019 Cricket World Cup.

  • Predict the outcome of individual matches for the entire competition.

  • Run simulation of the next matches i.e semi finals and finals.

These goals present a unique real-world Machine Learning prediction problem and involve solving various Machine Learning tasks: data wrangling, feature extraction and outcome prediction.

Data

I used data sets from Kaggle - Results of the matches since 1975 and 2017. I didn't get the data for 2018 and 2019 so this model might not be that accurate but still I believe this gives a fairly good intuition. Also I removed all the data from 1975 to 2010 since what happened way back in the past will have much less weight than the recent results. For the rest of data files I used the crickbuzz website.

Environment and tools

  1. Jupyter Notebook
  2. Numpy
  3. Pandas
  4. Seaborn
  5. Matplotlib
  6. Scikit-learn

I used the ICC ranking as of MAY 2019 dataset and a dataset containing the fixture of the group stages of the tournament. I compared Support Vector Machines, Logistic Regression, Random Forest and K-Nearest Neighbours model.

Random Forest was the winner with a training accuracy of 70 % and test accuracy of 67.5%.

Installation

pip install -r requirements.txt

jupyter notebook

According to this model England is likely to win this World Cup.

Areas of further Improvement

  1. Dataset - to improve dataset you could use 2018 and 2019 years into account by scraping them from the ESPN website and also possibly use the players data to assess the quality of each team player.

  2. A confusion matrix would be great to analyse which games the model got wrong.

  3. We could ensemble that is we could try stacking more models together to improve the accuracy.

References

  1. https://arxiv.org/abs/1806.03208

License

MIT License

Copyright (c) 2019 Abhinav Sagar

Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].