All Projects → tier4 → Autowarearchitectureproposal.proj

tier4 / Autowarearchitectureproposal.proj

Licence: apache-2.0
This is the source code of the feasibility study for Autoware architecture proposal.

Programming Languages

shell
77523 projects

Projects that are alternatives of or similar to Autowarearchitectureproposal.proj

Apollo
An open autonomous driving platform
Stars: ✭ 19,814 (+4928.93%)
Mutual labels:  autonomous-driving, autonomous-vehicles
Hybridastartrailer
A path planning algorithm based on Hybrid A* for trailer truck
Stars: ✭ 279 (-29.19%)
Mutual labels:  autonomous-driving, autonomous-vehicles
copilot
Lane and obstacle detection for active assistance during driving. Uses windowed sweep for lane detection. Combination of object tracking and YOLO for obstacles. Determines lane change, relative velocity and time to collision
Stars: ✭ 95 (-75.89%)
Mutual labels:  autonomous-driving, autonomous-vehicles
highway-path-planning
My path-planning pipeline to navigate a car safely around a virtual highway with other traffic.
Stars: ✭ 39 (-90.1%)
Mutual labels:  autonomous-driving, autonomous-vehicles
WIMP
[arXiv] What-If Motion Prediction for Autonomous Driving ❓🚗💨
Stars: ✭ 80 (-79.7%)
Mutual labels:  autonomous-driving, autonomous-vehicles
SelfDrivingRCCar
Autonomous RC Car using Neural Networks, Python and Open CV
Stars: ✭ 102 (-74.11%)
Mutual labels:  autonomous-driving, autonomous-vehicles
ar-tu-do
ROS & Gazebo project for 1/10th scale self-driving race cars
Stars: ✭ 65 (-83.5%)
Mutual labels:  autonomous-driving, autonomous-vehicles
glcapsnet
Global-Local Capsule Network (GLCapsNet) is a capsule-based architecture able to provide context-based eye fixation prediction for several autonomous driving scenarios, while offering interpretability both globally and locally.
Stars: ✭ 33 (-91.62%)
Mutual labels:  autonomous-driving, autonomous-vehicles
FusionAD
An open source autonomous driving stack by San Jose State University Autonomous Driving Team
Stars: ✭ 30 (-92.39%)
Mutual labels:  autonomous-driving, autonomous-vehicles
Awesome-3D-Object-Detection-for-Autonomous-Driving
Papers on 3D Object Detection for Autonomous Driving
Stars: ✭ 52 (-86.8%)
Mutual labels:  autonomous-driving, autonomous-vehicles
pyMHT
Track oriented, multi target, multi hypothesis tracker
Stars: ✭ 66 (-83.25%)
Mutual labels:  autonomous-driving, autonomous-vehicles
Error-State-Extended-Kalman-Filter
Vehicle State Estimation using Error-State Extended Kalman Filter
Stars: ✭ 100 (-74.62%)
Mutual labels:  autonomous-driving, autonomous-vehicles
Visualizing-lidar-data
Visualizing lidar data using Uber Autonomous Visualization System (AVS) and Jupyter Notebook Application
Stars: ✭ 75 (-80.96%)
Mutual labels:  autonomous-driving, autonomous-vehicles
PyLidar3
PyLidar3 is python 3 package to get data from Lidar devices from various manufacturers.
Stars: ✭ 35 (-91.12%)
Mutual labels:  autonomous-driving, autonomous-vehicles
dig-into-apollo
Apollo notes (Apollo学习笔记) - Apollo learning notes for beginners.
Stars: ✭ 1,786 (+353.3%)
Mutual labels:  autonomous-driving, autonomous-vehicles
OpenHDMap
An open HD map production process for autonomous car simulation
Stars: ✭ 152 (-61.42%)
Mutual labels:  autonomous-driving, autonomous-vehicles
JuliaAutonomy
Julia sample codes for Autonomy, Robotics and Self-Driving Algorithms.
Stars: ✭ 21 (-94.67%)
Mutual labels:  autonomous-driving, autonomous-vehicles
SelfDrivingCarsControlDesign
Self Driving Cars Longitudinal and Lateral Control Design
Stars: ✭ 96 (-75.63%)
Mutual labels:  autonomous-driving, autonomous-vehicles
BtcDet
Behind the Curtain: Learning Occluded Shapes for 3D Object Detection
Stars: ✭ 104 (-73.6%)
Mutual labels:  autonomous-driving, autonomous-vehicles
sparse-scene-flow
This repo contains C++ code for sparse scene flow method.
Stars: ✭ 23 (-94.16%)
Mutual labels:  autonomous-driving, autonomous-vehicles

Autoware (Architecture Proposal)

meta-repository for Autoware architecture proposal version

autoware

What's this

This is the source code of the feasibility study for Autoware architecture proposal.

WARNING: This source is solely for demonstrating an architecture proposal. It should not be used to drive cars.

NOTE: The features in AutowareArchitectureProposal.iv will be merged into Autoware.Auto.

Architecture overview is here.

How to setup

Requirements

Hardware

  • x86 CPU (8 or more cores)
  • 16 GB or more of memory
  • Nvidia GPU (4GB or more of memory) *optional
    • When you use following packages, GPU is mandatory:
      • lidar_apollo_instance_segmentation
      • traffic_light_ssd_fine_detector
      • cnn_classifier

If cuda or tensorRT is already installed, it is recommended to remove it.

Software

  • Ubuntu 18.04
  • Nvidia driver

If cuda or tensorRT is already installed, it is recommended to remove it.

How to setup

  1. Set up the repository

If ROS hasn't been installed yet in PC, at first run commands of 1.2 ~ 1.4, described in ROS wiki.

sudo sh -c 'echo "deb http://packages.ros.org/ros/ubuntu $(lsb_release -sc) main" > /etc/apt/sources.list.d/ros-latest.list'
sudo apt-key adv --keyserver 'hkp://keyserver.ubuntu.com:80' --recv-key C1CF6E31E6BADE8868B172B4F42ED6FBAB17C654
sudo apt update

Set up the repository

sudo apt install -y python3-vcstool
git clone [email protected]:tier4/AutowareArchitectureProposal.git
cd AutowareArchitectureProposal
mkdir -p src
vcs import src < autoware.proj.repos
  1. Run the setup script
./setup_ubuntu18.04.sh

In this step, the following software are installed.
Please confirm their licenses before using them.

  1. Build the source
source ~/.bashrc
colcon build --cmake-args -DCMAKE_BUILD_TYPE=Release --catkin-skip-building-tests

How to configure

Set hardware configuration

Prepare launch files and vehicle_description according to the sensor configuration of your hardware.
The following are the samples.

How to run

Supported Simulations

sim

Quick Start

Rosbag

  1. Download sample map from here.

  2. Download sample rosbag from here.

  3. Launch Autoware

cd AutowareArchitectureProposal
source install/setup.bash
roslaunch autoware_launch logging_simulator.launch map_path:=[path] vehicle_model:=lexus sensor_model:=aip_xx1 rosbag:=true

* Absolute path is required for map_path.

  1. Play rosbag
rosbag play --clock [rosbag file] -r 0.2
Note
  • sample map : © 2020 TierIV inc.
  • rosbag : © 2020 TierIV inc.
    • Image data are removed due to privacy concerns.
      • Cannot run traffic light recognition
      • Decreased accuracy of object detection

Planning Simulator

  1. Download sample map from here.

  2. Launch Autoware

cd AutowareArchitectureProposal
source install/setup.bash
roslaunch autoware_launch planning_simulator.launch map_path:=[path] vehicle_model:=lexus sensor_model:=aip_xx1

* Absolute path is required for map_path.

  1. Set initial pose
  2. Set goal pose
  3. Push engage button. autoware_web_controller
Note
  • sample map : © 2020 TierIV inc.

Running With AutowareAuto

We are planning propose the architecture and reference implementation to AutowareAuto. For the time being, use ros_bridge if you wish to use this repository with AutowareAuto modules. You would have to do the message type conversions in order to communicate between AutowareAuto and AutowareArchitectureProposal modules until the architecture is aligned.

For setting up AutowareAuto, please follow the instruction in: https://gitlab.com/autowarefoundation/autoware.auto/AutowareAuto

For setting up ros_bridge, please follow the instruction in: https://github.com/ros2/ros1_bridge

Tutorial in detail

See here. for more information.

References

Videos

Credits

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].