All Projects → m5823779 → DDPG

m5823779 / DDPG

Licence: other
End to End Mobile Robot Navigation using DDPG (Continuous Control with Deep Reinforcement Learning) based on Tensorflow + Gazebo

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to DDPG

motion-planner-reinforcement-learning
End to end motion planner using Deep Deterministic Policy Gradient (DDPG) in gazebo
Stars: ✭ 99 (+141.46%)
Mutual labels:  navigation, deep-reinforcement-learning, gazebo, ddpg
drl grasping
Deep Reinforcement Learning for Robotic Grasping from Octrees
Stars: ✭ 160 (+290.24%)
Mutual labels:  deep-reinforcement-learning, ros, gazebo
Autonomous-Drifting
Autonomous Drifting using Reinforcement Learning
Stars: ✭ 83 (+102.44%)
Mutual labels:  deep-reinforcement-learning, ros, gazebo
wpr simulation
No description or website provided.
Stars: ✭ 24 (-41.46%)
Mutual labels:  navigation, ros, gazebo
Smacc
An Event-Driven, Asynchronous, Behavioral State Machine Library for real-time ROS (Robotic Operating System) applications written in C++
Stars: ✭ 129 (+214.63%)
Mutual labels:  navigation, ros
Mesh navigation
ROS Mesh Navigation Bundle
Stars: ✭ 114 (+178.05%)
Mutual labels:  navigation, ros
Self Driving Golf Cart
Be Driven 🚘
Stars: ✭ 147 (+258.54%)
Mutual labels:  navigation, ros
Deep-Reinforcement-Learning-With-Python
Master classic RL, deep RL, distributional RL, inverse RL, and more using OpenAI Gym and TensorFlow with extensive Math
Stars: ✭ 222 (+441.46%)
Mutual labels:  deep-reinforcement-learning, ddpg
Navigation
ROS Navigation stack. Code for finding where the robot is and how it can get somewhere else.
Stars: ✭ 1,248 (+2943.9%)
Mutual labels:  navigation, ros
Spatio temporal voxel layer
A new voxel layer leveraging modern 3D graphics tools to modernize navigation environmental representations
Stars: ✭ 246 (+500%)
Mutual labels:  navigation, ros
Rainy
☔ Deep RL agents with PyTorch☔
Stars: ✭ 39 (-4.88%)
Mutual labels:  deep-reinforcement-learning, ddpg
Navbot
Using RGB Image as Visual Input for Mapless Robot Navigation
Stars: ✭ 111 (+170.73%)
Mutual labels:  navigation, ros
Turtlebot3 simulations
Simulations for TurtleBot3
Stars: ✭ 104 (+153.66%)
Mutual labels:  navigation, ros
Mpc ros
Nonlinear Model Predictive Control on Differential Wheeled Mobile Robot using ROS
Stars: ✭ 140 (+241.46%)
Mutual labels:  navigation, ros
Mrpt navigation
ROS nodes wrapping core MRPT functionality: localization, autonomous navigation, rawlogs, etc.
Stars: ✭ 90 (+119.51%)
Mutual labels:  navigation, ros
pytorch-distributed
Ape-X DQN & DDPG with pytorch & tensorboard
Stars: ✭ 98 (+139.02%)
Mutual labels:  deep-reinforcement-learning, ddpg
deeprl-continuous-control
Learning Continuous Control in Deep Reinforcement Learning
Stars: ✭ 14 (-65.85%)
Mutual labels:  deep-reinforcement-learning, ddpg
Deep-Reinforcement-Learning-for-Automated-Stock-Trading-Ensemble-Strategy-ICAIF-2020
Live Trading. Please star.
Stars: ✭ 1,251 (+2951.22%)
Mutual labels:  deep-reinforcement-learning, ddpg
AI
使用深度强化学习解决视觉跟踪和视觉导航问题
Stars: ✭ 16 (-60.98%)
Mutual labels:  navigation, deep-reinforcement-learning
Tianbot racecar
DISCONTINUED - MIGRATED TO TIANRACER - A Low cost Autonomous Driving Car Educational and Competition Kit
Stars: ✭ 26 (-36.59%)
Mutual labels:  navigation, ros

DDPG (Notice: Stop updating!!! https://github.com/m5823779/MotionPlannerUsingDDPG to view the final version)

End to End Mobile Robot Navigation using DDPG (Continuous Control with Deep Reinforcement Learning) based on Tensorflow + Gazebo

(Notice: Stop updating!!! https://github.com/m5823779/MotionPlannerUsingDDPG to view the final version)

Goal: Let robot(turtlebot) navigate to the target(enter green circle)

Input: 10 laser finding Output: 2 action (linear velocity [0 - 1] / angular velocity [-1 - 1]) (action_dimension = 2)

Algorithm: DDPG (Actor with batch normlization Critic without batch normlization) Training env: gazebo

Source code: https://github.com/floodsung/DDPG

Testing result:

Following video is my testing result when action dimension = 1 (only control angular velocity / linear velocity = o.5 m/s)

result is good enough image

Following video is my testing result when action dimension = 2 (control both linear velocity and angular velocity)

result is quite bad output will saturate image

Following vedio is output First is linear velocity, second is angular velocity the output is always converge to 0

image

Problem:

When action dimension = 2 action will be saturate(can't navigation)

"Have anyone meet this problem and already solved it?"

reference:

https://arxiv.org/pdf/1703.00420.pdf

https://github.com/floodsung/DDPG

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].