All Projects → JasonYao81000 → Mlds2018spring

JasonYao81000 / Mlds2018spring

Machine Learning and having it Deep and Structured (MLDS) in 2018 spring

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Mlds2018spring

ADL2019
Applied Deep Learning (2019 Spring) @ NTU
Stars: ✭ 20 (-83.87%)
Mutual labels:  generative-adversarial-network, gan, policy-gradient, image-generation, actor-critic
Tensorflow Reinforce
Implementations of Reinforcement Learning Models in Tensorflow
Stars: ✭ 480 (+287.1%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic, deep-q-network
Pytorch Rl
This repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (+217.74%)
Mutual labels:  reinforcement-learning, gan, generative-adversarial-network, policy-gradient
Seq2seq Chatbot For Keras
This repository contains a new generative model of chatbot based on seq2seq modeling.
Stars: ✭ 322 (+159.68%)
Mutual labels:  chatbot, gan, generative-adversarial-network, seq2seq
Reinforcement Learning
Minimal and Clean Reinforcement Learning Examples
Stars: ✭ 2,863 (+2208.87%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic, deep-q-network
Reinforcement Learning With Tensorflow
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
Stars: ✭ 6,948 (+5503.23%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic, deep-q-network
Seqgan
A simplified PyTorch implementation of "SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient." (Yu, Lantao, et al.)
Stars: ✭ 502 (+304.84%)
Mutual labels:  gan, generative-adversarial-network, policy-gradient
Hidt
Official repository for the paper "High-Resolution Daytime Translation Without Domain Labels" (CVPR2020, Oral)
Stars: ✭ 513 (+313.71%)
Mutual labels:  gan, generative-adversarial-network, image-generation
Exposure
Learning infinite-resolution image processing with GAN and RL from unpaired image datasets, using a differentiable photo editing model.
Stars: ✭ 605 (+387.9%)
Mutual labels:  reinforcement-learning, gan, generative-adversarial-network
Rlseq2seq
Deep Reinforcement Learning For Sequence to Sequence Models
Stars: ✭ 683 (+450.81%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic
Hands On Reinforcement Learning With Python
Master Reinforcement and Deep Reinforcement Learning using OpenAI Gym and TensorFlow
Stars: ✭ 640 (+416.13%)
Mutual labels:  reinforcement-learning, policy-gradient, deep-q-network
Deeprl Tutorials
Contains high quality implementations of Deep Reinforcement Learning algorithms written in PyTorch
Stars: ✭ 748 (+503.23%)
Mutual labels:  reinforcement-learning, actor-critic, deep-q-network
Reinforcement learning tutorial with demo
Reinforcement Learning Tutorial with Demo: DP (Policy and Value Iteration), Monte Carlo, TD Learning (SARSA, QLearning), Function Approximation, Policy Gradient, DQN, Imitation, Meta Learning, Papers, Courses, etc..
Stars: ✭ 442 (+256.45%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic
Pytorch Rl
Tutorials for reinforcement learning in PyTorch and Gym by implementing a few of the popular algorithms. [IN PROGRESS]
Stars: ✭ 121 (-2.42%)
Mutual labels:  reinforcement-learning, policy-gradient, actor-critic
Tensorflow Tutorial
Tensorflow tutorial from basic to hard, 莫烦Python 中文AI教学
Stars: ✭ 4,122 (+3224.19%)
Mutual labels:  gan, generative-adversarial-network, deep-q-network
Pytorch Rl
PyTorch implementation of Deep Reinforcement Learning: Policy Gradient methods (TRPO, PPO, A2C) and Generative Adversarial Imitation Learning (GAIL). Fast Fisher vector product TRPO.
Stars: ✭ 658 (+430.65%)
Mutual labels:  reinforcement-learning, generative-adversarial-network, policy-gradient
Pix2pix
Image-to-image translation with conditional adversarial nets
Stars: ✭ 8,765 (+6968.55%)
Mutual labels:  gan, generative-adversarial-network, image-generation
Tensorlayer
Deep Learning and Reinforcement Learning Library for Scientists and Engineers 🔥
Stars: ✭ 6,796 (+5380.65%)
Mutual labels:  chatbot, reinforcement-learning, gan
Dcgan Tensorflow
A Tensorflow implementation of Deep Convolutional Generative Adversarial Networks trained on Fashion-MNIST, CIFAR-10, etc.
Stars: ✭ 70 (-43.55%)
Mutual labels:  gan, generative-adversarial-network, image-generation
Text summurization abstractive methods
Multiple implementations for abstractive text summurization , using google colab
Stars: ✭ 359 (+189.52%)
Mutual labels:  reinforcement-learning, seq2seq, policy-gradient

MLDS2018SPRING

Machine Learning and having it deep and structured (MLDS) at NTU 2018 Spring.

This course has four homeworks, group by group. The four homeworks are as follows:

  1. Deep Learning Theory
  2. Sequence-to-sequence Model
  3. Deep Generative Model
  4. Deep Reinforcement Learning

Browse this course website for more details.

Table of Contents

  1. Deep Learning Theory
  2. Sequence-to-sequence Model
  3. Deep Generative Model
  4. Deep Reinforcement Learning

Results of Four Homeworks

1. Deep Learning Theory

1.1 Deep vs Shallow

1.2 Optimization

1.3 Generalization

2. Sequence-to-sequence Model

2.1 Video caption generation

2.2 Chat-bot

3. Deep Generative Model

3.1 Image Generation

  • README
  • Image Generation: 100% (25/25) Pass Baseline
./gan-baseline/baseline_result_gan.png

3.2 Text-to-Image Generation

  • README
  • Text-to-Image Generation: 100% (25/25) Pass Baseline
Testing Tags ./gan-baseline/baseline_result_cgan.png
blue hair blue eyes


blue hair green eyes


blue hair red eyes


green hair blue eyes


green hair red eyes

3.3 Style Transfer

4. Deep Reinforcement Learning

4.1 Policy Gradient

  • README
  • Policy Gradient: Mean Rewards in 30 Episodes = 16.466666666666665

4.2 Deep Q Learning

  • README
  • Deep Q Learning: Mean Rewards in 100 Episodes = 73.16

4.3 Actor-Critic

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].