All Projects → shinkansan → 2019-UGRP-DPoom

shinkansan / 2019-UGRP-DPoom

Licence: Apache-2.0 License
2019 DGIST DPoom project under UGRP : SBC and RGB-D camera based full autonomous driving system for mobile robot with indoor SLAM

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to 2019-UGRP-DPoom

roskinectic src
This ROS kinectic workspace src folder, which was created on Ubuntu 16.04. Here I worked on ROS1 projects like 2d & 3D SLAM, Motion Planning, SWARM of drone, RL based drone, Surveilling Robot etc.
Stars: ✭ 44 (+25.71%)
Mutual labels:  ros, slam, ros-kinetic, rtabmap
FusionAD
An open source autonomous driving stack by San Jose State University Autonomous Driving Team
Stars: ✭ 30 (-14.29%)
Mutual labels:  robotics, ros, autonomous-driving, ros-kinetic
Awesome Robotic Tooling
Tooling for professional robotic development in C++ and Python with a touch of ROS, autonomous driving and aerospace.
Stars: ✭ 1,876 (+5260%)
Mutual labels:  robotics, ros, slam, autonomous-driving
Pythonrobotics
Python sample codes for robotics algorithms.
Stars: ✭ 13,934 (+39711.43%)
Mutual labels:  robotics, path-planning, slam, autonomous-driving
Orb slam 2 ros
A ROS implementation of ORB_SLAM2
Stars: ✭ 294 (+740%)
Mutual labels:  robotics, ros, slam
Se2lam
(ICRA 2019) Visual-Odometric On-SE(2) Localization and Mapping
Stars: ✭ 285 (+714.29%)
Mutual labels:  robotics, ros, slam
conde simulator
Autonomous Driving Simulator for the Portuguese Robotics Open
Stars: ✭ 31 (-11.43%)
Mutual labels:  ros, autonomous-driving, ros-kinetic
Mrpt
⚡️ The Mobile Robot Programming Toolkit (MRPT)
Stars: ✭ 1,190 (+3300%)
Mutual labels:  robotics, slam, autonomous-driving
Autorally
Software for the AutoRally platform
Stars: ✭ 595 (+1600%)
Mutual labels:  robotics, ros, autonomous-driving
Evo
Python package for the evaluation of odometry and SLAM
Stars: ✭ 1,373 (+3822.86%)
Mutual labels:  robotics, ros, slam
Cleanit
Open-source Autonomy Software in Rust-lang with gRPC for the Roomba series robot vacuum cleaners. Under development.
Stars: ✭ 125 (+257.14%)
Mutual labels:  robotics, ros, slam
Camlasercalibratool
Extrinsic Calibration of a Camera and 2d Laser
Stars: ✭ 277 (+691.43%)
Mutual labels:  robotics, ros, slam
Se2clam
SE(2)-Constrained Localization and Mapping by Fusing Odometry and Vision (IEEE Transactions on Cybernetics 2019)
Stars: ✭ 116 (+231.43%)
Mutual labels:  robotics, ros, slam
direct lidar odometry
Direct LiDAR Odometry: Fast Localization with Dense Point Clouds
Stars: ✭ 202 (+477.14%)
Mutual labels:  robotics, ros, slam
Cherry Autonomous Racecar
Implementation of the CNN from End to End Learning for Self-Driving Cars on a Nvidia Jetson TX1 using Tensorflow and ROS
Stars: ✭ 294 (+740%)
Mutual labels:  robotics, ros, autonomous-driving
erdos
Dataflow system for building self-driving car and robotics applications.
Stars: ✭ 135 (+285.71%)
Mutual labels:  robotics, ros, robot-framework
Deepseqslam
The Official Deep Learning Framework for Route-based Place Recognition
Stars: ✭ 49 (+40%)
Mutual labels:  robotics, slam, autonomous-driving
SLAM AND PATH PLANNING ALGORITHMS
This repository contains the solutions to all the exercises for the MOOC about SLAM and PATH-PLANNING algorithms given by professor Claus Brenner at Leibniz University. This repository also contains my personal notes, most of them in PDF format, and many vector graphics created by myself to illustrate the theoretical concepts. Hope you enjoy it! :)
Stars: ✭ 107 (+205.71%)
Mutual labels:  robotics, path-planning, slam
ROS-GPS
GPS Localization with ROS, OSM and rviz
Stars: ✭ 19 (-45.71%)
Mutual labels:  robotics, ros, ros-kinetic
JuliaAutonomy
Julia sample codes for Autonomy, Robotics and Self-Driving Algorithms.
Stars: ✭ 21 (-40%)
Mutual labels:  path-planning, slam, autonomous-driving

Indoor Mobile Robot Fully Autonomous Driving Project - DPoom build_status

For further information, please visit our page

Click below for introduction video.

Introduction Video

Single board computer Lattepanda Alpha based indoor mobile robot DPoom, with fully autonomous driving system using single RGB-D camera Intel Realsense D435i.

Keywords: autonomy, autonomous driving system, mobile robot, SLAM, ROS, RGB-D, Low-end, global path planning, motion planning, ground segmentaion, navigation, path tracking, control, Human-Robot Interaction

Installation

Follow installation guide here

How to Run

Robot control

The robot control two dynamixel motors via ROS and OpenCR. For providing a easy way of robot control, we built a driving control package name easyGo. See Control Package Page.

SLAM

Mapping should be preceded before deploying robots. See SLAM Page.

Global path planning

Using pcd map, the robot can plan global path using our FMM based modified A*. See GPP Page.

Motion planning

The robot can follow the generated path by motion planner. Our motion planner is using our real-time ground segmentation method named MORP. See Motion Planning (MORP) Page.

Full autonomous driving

RGB-D localization, global path planning and motion planning are integrated in one python script. Just run integration.py. For details, see Integration Page

Human-Computer Interaction

See Human-Computer Interaction Page.

Related Repositories

DPoom_Gazebo

DPoom is also availble in ROS Gazebo simulation with equivalent codes. To simulate DPoom in Gazebo: DPoom_gazebo

Gazebo-CrowdNav

Our navigation method can be simulated in Gazebo. Current state-of-the-art navigation approches based on DRL (CADRL, LSTM-RL, SARL) are available with DPoom platform. To evalute navigation performance with DPoom in Gazebo: Gazebo-CrowdNav

Project Info

2019/12/5 We opened our github repos to public!!.

a a

Purpose of this project

These days mobile robots are rapidly developing in the industry. However, there are still some problems for practical application such as expensive hardware and high power consumption. In this study, we propose a navigation system, which can be operated on a low-end computer with an RGB-D camera, and a mobile robot platform to operate integrated autonomous driving system. The proposed system does not require LiDARS or GPUs. Our raw depth image ground segmentation extracts a traversability map for safe driving of the low-body mobile robots. It is designed to guarantee real-time performance on a low-cost commercial single board computer with integrated SLAM, global path planning, and motion planning. Running sensor data processing and other autonomous driving functions simultaneously, our navigation method performs fast at 18Hz refresh rate for control command, while the others have slower refresh rates. Our method outperforms current state-of-the-art navigation approaches as shown in 3D simulation tests. In addition, we demonstrate the applicability of our mobile robot system by successfully autonomous driving in a residential lobby.

Our team

Taekyung Kim / DGIST Class of 2020 @ktk1501
Seunghyun Lim / DGIST Class of 2020 @SeunghyunLim
Gwanjun Shin / DGIST undergraduate @shinkansan
Geonhee Sim / DGIST undergraduate @jane79

Platform info

SOFTWARE

  • Ubuntu 16.04 LTS
  • ROS Kinetic
  • Python 3.6
  • Python 2.7 on Robot Platform
  • Realsense SDK 2
  • Tensorflow 1.8.0
  • OpenCV

HARDWARE

  • Platform Computing Unit : Lattepanda Alpha 864
  • Intel Realsense Camera D435i
  • Turtlebot3 waffle pi
  • Outer HW printed by 3D-Printer

Honors

DPoom won the 2019 Samsung Open Source Conference (SOSCON) Robot Competition. Media coverage

Paper info

Our paper had submitted to IROS 2021

Build Status Badge by shield.io

This work is partially supported by DGIST UGRP.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].