All Projects β†’ moebg β†’ click-through-rate-prediction

moebg / click-through-rate-prediction

Licence: MIT license
πŸ“ˆ Click-Through Rate Prediction using Logistic Regression and Tree Algorithms

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to click-through-rate-prediction

AIML-Projects
Projects I completed as a part of Great Learning's PGP - Artificial Intelligence and Machine Learning
Stars: ✭ 85 (+41.67%)
Mutual labels:  random-forest, logistic-regression, decision-tree-classifier
Machine Learning Models
Decision Trees, Random Forest, Dynamic Time Warping, Naive Bayes, KNN, Linear Regression, Logistic Regression, Mixture Of Gaussian, Neural Network, PCA, SVD, Gaussian Naive Bayes, Fitting Data to Gaussian, K-Means
Stars: ✭ 160 (+166.67%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Machine Learning Is All You Need
πŸ”₯πŸŒŸγ€ŠMachine Learning 格物志》: ML + DL + RL basic codes and notes by sklearn, PyTorch, TensorFlow, Keras & the most important, from scratch!πŸ’ͺ This repository is ALL You Need!
Stars: ✭ 173 (+188.33%)
Mutual labels:  random-forest, logistic-regression, decision-trees
supervised-machine-learning
This repo contains regression and classification projects. Examples: development of predictive models for comments on social media websites; building classifiers to predict outcomes in sports competitions; churn analysis; prediction of clicks on online ads; analysis of the opioids crisis and an analysis of retail store expansion strategies using…
Stars: ✭ 34 (-43.33%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Amazon-Fine-Food-Review
Machine learning algorithm such as KNN,Naive Bayes,Logistic Regression,SVM,Decision Trees,Random Forest,k means and Truncated SVD on amazon fine food review
Stars: ✭ 28 (-53.33%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Machine Learning With Python
Python code for common Machine Learning Algorithms
Stars: ✭ 3,334 (+5456.67%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Breast-Cancer-Scikitlearn
simple tutorial on Machine Learning with Scikitlearn
Stars: ✭ 33 (-45%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Text classification
Text Classification Algorithms: A Survey
Stars: ✭ 1,276 (+2026.67%)
Mutual labels:  random-forest, logistic-regression, decision-trees
Machine Learning With Python
Practice and tutorial-style notebooks covering wide variety of machine learning techniques
Stars: ✭ 2,197 (+3561.67%)
Mutual labels:  random-forest, decision-trees
Chefboost
A Lightweight Decision Tree Framework supporting regular algorithms: ID3, C4,5, CART, CHAID and Regression Trees; some advanced techniques: Gradient Boosting (GBDT, GBRT, GBM), Random Forest and Adaboost w/categorical features support for Python
Stars: ✭ 176 (+193.33%)
Mutual labels:  random-forest, decision-trees
Textclassification
several methods for text classification
Stars: ✭ 180 (+200%)
Mutual labels:  random-forest, logistic-regression
STOCK-RETURN-PREDICTION-USING-KNN-SVM-GUASSIAN-PROCESS-ADABOOST-TREE-REGRESSION-AND-QDA
Forecast stock prices using machine learning approach. A time series analysis. Employ the Use of Predictive Modeling in Machine Learning to Forecast Stock Return. Approach Used by Hedge Funds to Select Tradeable Stocks
Stars: ✭ 94 (+56.67%)
Mutual labels:  random-forest, logistic-regression
Awesome Decision Tree Papers
A collection of research papers on decision, classification and regression trees with implementations.
Stars: ✭ 1,908 (+3080%)
Mutual labels:  random-forest, decision-tree-classifier
Orange3
🍊 πŸ“Š πŸ’‘ Orange: Interactive data analysis
Stars: ✭ 3,152 (+5153.33%)
Mutual labels:  random-forest, decision-trees
Trajectory-Analysis-and-Classification-in-Python-Pandas-and-Scikit-Learn
Formed trajectories of sets of points.Experimented on finding similarities between trajectories based on DTW (Dynamic Time Warping) and LCSS (Longest Common SubSequence) algorithms.Modeled trajectories as strings based on a Grid representation.Benchmarked KNN, Random Forest, Logistic Regression classification algorithms to classify efficiently t…
Stars: ✭ 41 (-31.67%)
Mutual labels:  random-forest, logistic-regression
Tensorflow Ml Nlp
ν…μ„œν”Œλ‘œμš°μ™€ λ¨Έμ‹ λŸ¬λ‹μœΌλ‘œ μ‹œμž‘ν•˜λŠ” μžμ—°μ–΄μ²˜λ¦¬(λ‘œμ§€μŠ€ν‹±νšŒκ·€λΆ€ν„° 트랜슀포머 μ±—λ΄‡κΉŒμ§€)
Stars: ✭ 176 (+193.33%)
Mutual labels:  random-forest, logistic-regression
Machine Learning In R
Workshop (6 hours): preprocessing, cross-validation, lasso, decision trees, random forest, xgboost, superlearner ensembles
Stars: ✭ 144 (+140%)
Mutual labels:  random-forest, decision-trees
R-stats-machine-learning
Misc Statistics and Machine Learning codes in R
Stars: ✭ 33 (-45%)
Mutual labels:  random-forest, decision-trees
Statistical-Learning-using-R
This is a Statistical Learning application which will consist of various Machine Learning algorithms and their implementation in R done by me and their in depth interpretation.Documents and reports related to the below mentioned techniques can be found on my Rpubs profile.
Stars: ✭ 27 (-55%)
Mutual labels:  logistic-regression, decision-trees
Machine-Learning-Models
In This repository I made some simple to complex methods in machine learning. Here I try to build template style code.
Stars: ✭ 30 (-50%)
Mutual labels:  random-forest, logistic-regression

πŸ“ˆ Click-Through Rate Prediction CTR Icon

In online advertising, CTR is an important metric to measure an ad's performance. In this project we use a dataset from the Click-Through Rate Prediction competiton on Kaggle and evaluate methods for CTR prediction.

In the demo below, I've trained the models already and have them stored in the 'models' directory in the repo. In order to retrain the models, go into ctr_prediction.py and set the load_model parameter to False in order to retrain on your machine.

Contents

Overview

Digital marketers use online advertising (such as Adwords, Facebook ads, etc.) to display ads to users while they are browsing online. In order to deliver and see how well their ad performs, advertisers use click-through rate (CTR) to measure how well their ad is doing. CTR is the ratio of clicks on a specific ad to the total number of views. The higher the CTR, the better.

Problem: Given a user U and a page P the user is visiting, what is the probability of the user U to click the displayed ad.

For the above problem, machine learning is heavily used to in order to improve the quality of ads that are delivered to users. Here are some of the features we take into consideration when trying to predict the probability a user will click on an ad:

  • Ad information and content (format, text information, position, topic/content)
  • Publisher information and content of page (website type, topic, context, etc.)
  • User information (age, income, search history, gender, location, interests, etc.)

Dataset

The dataset used for this project is the Avazu CTR Dataset which contains 11 days of click-through data. 10 days of click-through data is used for training, and 1 day of ads to test the model. The dataset contains many features including site domains, categories, user device/ip/location, banner positon, and label of whether the ad was clicked or not.

Learning Algorithms Used

  • Decision Trees
  • Random Forest
  • SGD-Based Logistic Regression
  • Logistic Regression with Online Learning

Evaluation Methods

To evaluate the performance of each model, we used the ROC AUC Score. Learn more about ROC curves and AUC here.

Results

Usage

  1. Clone repo to local directory (make sure to have Python 3.4+ installed)
  2. cd src
  3. python ctr_prediction.py

Dependencies

  • Scikit-Learn
  • Numpy

Motivation

As someone who has worked in digital marketing in the past and have used products such as Google Adwords and Facebook Ads, I found this project to be an exploration of learning about some of the components that go into building ad products.

References

Attribution

Contributing

Contributions are always welcome! For bug reports or requests please submit an issue.

License

MIT

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].