All Projects → aravindr93 → trajopt

aravindr93 / trajopt

Licence: MIT License
Trajectory optimization algorithms for robotic control.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to trajopt

PnC
Planning and Control Algorithms for Robotics
Stars: ✭ 22 (-70.27%)
Mutual labels:  robotics, trajectory-optimization
TORA.jl
Trajectory Optimization for Robot Arms
Stars: ✭ 27 (-63.51%)
Mutual labels:  robotics, trajectory-optimization
Pytorch Rl
This repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (+432.43%)
Mutual labels:  robotics, mujoco
cito
A Contact-Implicit Trajectory Optimization Package
Stars: ✭ 20 (-72.97%)
Mutual labels:  trajectory-optimization, mujoco
kinpy
Simple kinematics calculation toolkit for robotics
Stars: ✭ 48 (-35.14%)
Mutual labels:  robotics, mujoco
Mjrl
Reinforcement learning algorithms for MuJoCo tasks
Stars: ✭ 162 (+118.92%)
Mutual labels:  robotics, mujoco
MuJoCo RL UR5
A MuJoCo/Gym environment for robot control using Reinforcement Learning. The task of agents in this environment is pixel-wise prediction of grasp success chances.
Stars: ✭ 108 (+45.95%)
Mutual labels:  robotics, mujoco
RoboticsAcademy
Learn Robotics with JdeRobot
Stars: ✭ 160 (+116.22%)
Mutual labels:  robotics
trac ik
ROS 2 port of `trac_ik`, an alternative Inverse Kinematics solver to the popular inverse Jacobian methods in KDL.
Stars: ✭ 14 (-81.08%)
Mutual labels:  robotics
SLAM AND PATH PLANNING ALGORITHMS
This repository contains the solutions to all the exercises for the MOOC about SLAM and PATH-PLANNING algorithms given by professor Claus Brenner at Leibniz University. This repository also contains my personal notes, most of them in PDF format, and many vector graphics created by myself to illustrate the theoretical concepts. Hope you enjoy it! :)
Stars: ✭ 107 (+44.59%)
Mutual labels:  robotics
neonavigation
A 2-D/3-DOF seamless global/local mobile robot motion planner package for ROS
Stars: ✭ 199 (+168.92%)
Mutual labels:  robotics
brax
Massively parallel rigidbody physics simulation on accelerator hardware.
Stars: ✭ 1,208 (+1532.43%)
Mutual labels:  robotics
erwhi-hedgehog
Erwhi Hedgehog main repository
Stars: ✭ 31 (-58.11%)
Mutual labels:  robotics
RustRobotics
Rust implementation of PythonRobotics such as EKF, DWA, Pure Pursuit, LQR.
Stars: ✭ 40 (-45.95%)
Mutual labels:  robotics
ROS-GPS
GPS Localization with ROS, OSM and rviz
Stars: ✭ 19 (-74.32%)
Mutual labels:  robotics
AI-Lab
🔬 Absolutely comfort lab for me to work around with my own AIs and to empirically observe how powerful and impactful these technologies are. I do love these technologies!
Stars: ✭ 19 (-74.32%)
Mutual labels:  robotics
robo-playground
Games and examples built for RoboMaster EP with RoboMasterPy | 与你的大疆机甲大师愉快玩耍,基于RoboMasterPy构建
Stars: ✭ 33 (-55.41%)
Mutual labels:  robotics
MeshCatMechanisms.jl
3D Visualization of mechanisms and URDFs using MeshCat.jl and RigidBodyDynamics.jl
Stars: ✭ 36 (-51.35%)
Mutual labels:  robotics
robocup-firmware
Georgia Tech RoboJackets Firmware for the RoboCup Small Size League
Stars: ✭ 22 (-70.27%)
Mutual labels:  robotics
Tinymovr
Compact brushless motor controller with integrated absolute encoder and CAN Bus.
Stars: ✭ 90 (+21.62%)
Mutual labels:  robotics

Trajectory Optimization Algorithms

This package contains trajectory optimization algorithms meant predominantly for continuous control taks (simulated with MuJoCo).

Installation

The main package dependencies are MuJoCo and mjrl. See setup-instructions to get a working conda environment and setup dependencies.

After mujoco_py has been installed, the package can be used by either adding to path as:

export PYTHONPATH=<path/to/trajopt>$PYTHONPATH

or through the pip install module

$ cd trajopt
$ pip install -e .

The tricky part of the installation is likely to be mujoco_py. Please see instructions and known issues for help.

API and example usage

The algorithms assume an environment abstraction similar to OpenAI gym, but requires two additional functions to be able to run the algorithms provided here.

  • get_env_state() should return a dictionary with all the information required to reconstruct the scene and dynamics. For most use cases, this can just be the qpos and qvel. However, in some cases, additional information may be required to construct scene and dynamics. For example, in multi-goal RL, we can represent virtual goals using sites as opposed to real joints.
  • set_env_state(state_dict) should take in a dictionary, and use the contents of the dictionary to recreate the scene specified by the dictionary. The example reacher environment has an illustrative.

Example Usage

See this directory for illustrative examples: trajopt/examples.

Bibliography

If you find the package useful, please cite the following paper.

@INPROCEEDINGS{Lowrey-ICLR-19,
    AUTHOR    = {Kendall Lowrey AND Aravind Rajeswaran AND Sham Kakade AND 
                 Emanuel Todorov AND Igor Mordatch},
    TITLE     = "{Plan Online, Learn Offline: Efficient Learning and Exploration via Model-Based Control}",
    BOOKTITLE = {ICLR},
    YEAR      = {2019},
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].