All Projects → philtabor → Youtube Code Repository

philtabor / Youtube Code Repository

Repository for most of the code from my YouTube channel

Projects that are alternatives of or similar to Youtube Code Repository

2048 Deep Reinforcement Learning
Trained A Convolutional Neural Network To Play 2048 using Deep-Reinforcement Learning
Stars: ✭ 169 (-46.69%)
Mutual labels:  jupyter-notebook, reinforcement-learning, convolutional-neural-networks
Tensorflow Book
Accompanying source code for Machine Learning with TensorFlow. Refer to the book for step-by-step explanations.
Stars: ✭ 4,448 (+1303.15%)
Mutual labels:  jupyter-notebook, reinforcement-learning, convolutional-neural-networks
Easy Tensorflow
Simple and comprehensive tutorials in TensorFlow
Stars: ✭ 2,871 (+805.68%)
Mutual labels:  reinforcement-learning, convolutional-neural-networks
Pytorch Image Classification
Tutorials on how to implement a few key architectures for image classification using PyTorch and TorchVision.
Stars: ✭ 272 (-14.2%)
Mutual labels:  jupyter-notebook, convolutional-neural-networks
Dinoruntutorial
Accompanying code for Paperspace tutorial "Build an AI to play Dino Run"
Stars: ✭ 285 (-10.09%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Rad
RAD: Reinforcement Learning with Augmented Data
Stars: ✭ 268 (-15.46%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Popular Rl Algorithms
PyTorch implementation of Soft Actor-Critic (SAC), Twin Delayed DDPG (TD3), Actor-Critic (AC/A2C), Proximal Policy Optimization (PPO), QT-Opt, PointNet..
Stars: ✭ 266 (-16.09%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Trading Bot
Stock Trading Bot using Deep Q-Learning
Stars: ✭ 273 (-13.88%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Stanford Cs231
Resources for students in the Udacity's Machine Learning Engineer Nanodegree to work through Stanford's Convolutional Neural Networks for Visual Recognition course (CS231n).
Stars: ✭ 249 (-21.45%)
Mutual labels:  jupyter-notebook, convolutional-neural-networks
Gdrl
Grokking Deep Reinforcement Learning
Stars: ✭ 304 (-4.1%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Baby Steps Of Rl Ja
Pythonで学ぶ強化学習 -入門から実践まで- サンプルコード
Stars: ✭ 302 (-4.73%)
Mutual labels:  jupyter-notebook, reinforcement-learning
A3c trading
Trading with recurrent actor-critic reinforcement learning
Stars: ✭ 305 (-3.79%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Deeplearning.ai Assignments
Stars: ✭ 268 (-15.46%)
Mutual labels:  jupyter-notebook, convolutional-neural-networks
Vel
Velocity in deep-learning research
Stars: ✭ 267 (-15.77%)
Mutual labels:  reinforcement-learning, convolutional-neural-networks
Drq
DrQ: Data regularized Q
Stars: ✭ 268 (-15.46%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Grad Cam Tensorflow
tensorflow implementation of Grad-CAM (CNN visualization)
Stars: ✭ 261 (-17.67%)
Mutual labels:  jupyter-notebook, convolutional-neural-networks
Rlquant
Applying Reinforcement Learning in Quantitative Trading
Stars: ✭ 271 (-14.51%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Reinforcement Learning
Learn Deep Reinforcement Learning in 60 days! Lectures & Code in Python. Reinforcement Learning + Deep Learning
Stars: ✭ 3,329 (+950.16%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Aleph star
Reinforcement learning with A* and a deep heuristic
Stars: ✭ 235 (-25.87%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Deeppicar
Deep Learning Autonomous Car based on Raspberry Pi, SunFounder PiCar-V Kit, TensorFlow, and Google's EdgeTPU Co-Processor
Stars: ✭ 242 (-23.66%)
Mutual labels:  jupyter-notebook, convolutional-neural-networks

Youtube-Code-Repository

Repository for all the code from my youtube channel You can find me at https://youtube.com/MachineLearningWithPhil

Kaggle/Venus-Volcanoes

My crude implementation of a convolutional neural network to perform image classification on data gathered
by the Magellan spacecraft. The data is horribly skewed, as most images do not contain a volcano.
This means we'll have to do some creative data engineering for our model training.
Please note that in the test set, 84.1% of the data is "no volcano", and our model returns
an accuracy of around 88%, which is better than a model that outputs straight 0s for predictions.

You can check out the video for this at https://youtu.be/Ki-xOKydQrY
You can find the data for this project at https://www.kaggle.com/fmena14/volcanoesvenus/home

ReinforcementLearning/DeepQLearning

My implementation of the Deep Q learning algorithm in PyTorch. Here we teach the algorithm to play the game of space invaders. I haven't had enough time to train this model yet, as it takes quite some time even on my 1080Ti / i7 7820k @ 4.4 GHz. I'll train longer and provide a video on how well it does, at a later time.

The blog post talking about how Deep Q learning works can be found at http://www.neuralnet.ai/coding-a-deep-q-network-in-pytorch/
Video for this is at https://www.youtube.com/watch?v=RfNxXlO6BiA&t=2s

CNN.py

Simple implementation of a convolutional neural network in TensorFlow, version 1.5.
Video tutorial on this code can be found here https://youtu.be/azFyHS0odcM
Achieves accuracy of 98% after 10 epochs of training
Requires data from http://yann.lecun.com/exdb/mnist/

ReinforcementLearning/blackJack-no-es.py

Implementation of Monte Carlo control without exploring starts in the blackjack environment from the OpenAI gym.
Video tutorial on this code can be found at https://youtu.be/e8ofon3sg8E
Algorithm trains for 1,000,000 games and produces a win rate of around 42%, loss rate of 52% and draw rate of 6%

ReinforcementLearning/blackJack-off-policy.py

Implementation of off policy Monte Carlo control in the blackjack environment from the OpenAI gym.
Video tutorial on this code can be found at https://youtu.be/TvO0Sa-6UVc
Algorithm trains for 1,000,000 games and produces a win rate of around 29%, loss rate of 66% and draw rate of 5%

ReinforcementLearning/cartpole_qlearning.py

Implementation of the Q learning algorithm for the cart pole problem. Code is based on the course by lazy programmer,
which you can find here here
Video tutorial on this code can be found at https://youtu.be/ViwBAK8Hd7Q

ReinforcementLearning/doubleQLearning.py

Implementation of the double Q learning algorithm in the cart pole environment. This is based on my course on
reinforcement learning, which you can find at this repo
Video tutorial on this code can be found https://youtu.be/Q99bEPStnxk

ReinforcementLearning/sarsa.py

Implementation of the SARSA algorithm in the cart pole environment. This is based on my course on reinforcement learning,
which can be found here
Video tutorial on this code can be found at https://youtu.be/P9XezMuPfLE

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].