README.md
![image](docs/_static/img/banner.png)
### Streamlining the Transfer of Simulated Robot Learning to the Real-World
[![license](https://img.shields.io/badge/License-Apache_2.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
[![codestyle](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black)
[![Documentation Status](https://readthedocs.org/projects/eagerx/badge/?version=master)](https://eagerx.readthedocs.io/en/master/?badge=master)
[![Continuous Integration](https://github.com/eager-dev/eagerx/actions/workflows/ci.yml/badge.svg?branch=master)](https://github.com/eager-dev/eagerx/actions/workflows/ci.yml)
[![Maintainability](https://api.codeclimate.com/v1/badges/3146dce3dd4c3537834c/maintainability)](https://codeclimate.com/github/eager-dev/eagerx/maintainability)
[![Test Coverage](https://api.codeclimate.com/v1/badges/3146dce3dd4c3537834c/test_coverage)](https://codeclimate.com/github/eager-dev/eagerx/test_coverage)
What is EAGERx
--------------
You can use EAGERx (*Engine Agnostic Graph Environments for Robotics*) to easily define new ([Gymnasium compatible](https://gymnasium.farama.org/)) environments with modular robot definitions.
It enables users to:
- Define environments as graphs of nodes
- Visualize these graph environments interactively in a GUI
- Use a single graph environment both in reality and with various simulators
EAGERx explicitly addresses the differences in learning between simulation and reality, with native support for essential features such as:
- Safety layers and various other state, action and time-scale abstractions
- Delay simulation & domain randomization
- Real-world reset routines
- Synchronized parallel computation within a single environment
[Full documentation and tutorials available
here](https://eagerx.readthedocs.io/en/master/).
<p align="center">
<img src="docs/_static/gif/pendulum_sim.gif" width="22.8%" alt="pendulum_sim"/>
<img src="docs/_static/gif/pendulum_real.gif" width="22.8%" alt="pendulum_real"/>
<img src="docs/_static/gif/box_pushing_pybullet.gif" width="22.8%" alt="box_sim"/>
<img src="docs/_static/gif/box_pushing_real.gif" width="22.8%" alt="box_real"/>
</p>
<p align="center">
<img src="docs/_static/gif/drone_landing.gif" width="66%" alt="cf_real"/>
</p>
**Sim2Real:** Policies trained in simulation and zero-shot evaluated on real systems using EAGERx.
In the top left the successful transfer of a policy for the classic pendulum swing-up problem is shown and in the top right for a box-pushing task.
Below that a policy to land a quadrotor on a moving inclined platform is shown.
<p align="center">
<img src="docs/_static/gif/all.gif" width="66%" />
</p>
**Modular:** The modular design of EAGERx allows users to create complex environments easily through composition.
<img align="middle" src="docs/_static/img/gui.svg" width="100%" />
**GUI:** Users can visualize their graph environment.
Here we visualize the graph environment that we built in [this tutorial](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/icra/advanced_usage.ipynb).
See the [documentation](https://eagerx.readthedocs.io/en/master/guide/getting_started/index.html#extras-gui) for more information.
<p align="center">
<img src="docs/_static/gif/rqt_plot.GIF" width="50%" />
</p>
**Live plotting:** In robotics it is crucial to monitor the robot\'s behavior during the learning process.
Luckily, inter-node communication within EAGERx can be listened to externally, so that any relevant information stream can be trivially monitored on-demand.
See the [documentation](https://eagerx.readthedocs.io/en/master/guide/getting_started/index.html#extras-training-visualization) for more information.
<p align="center">
<img src="docs/_static/gif/use_case.gif" width="35.5%" alt="use_case"/>
<img src="docs/_static/gif/swim_sim.gif" width="25%" alt="swim_sim"/>
<img src="docs/_static/gif/swim_real.gif" width="25%" alt="swim_real"/>
</p>
**Applications beyond RL:** The modular design and unified software pipeline of the framework have utility beyond reinforcement learning.
We explored two such instances: interactive language-conditioned imitation learning (left) and classical control with deep learning based perception in a swimming pool environment (right).
Installation
------------
You can do a minimal installation of `EAGERx` with:
```bash
pip3 install eagerx
```
We provide other options (Docker, Conda) for installing EAGERx in the [documentation](https://eagerx.readthedocs.io/en/master/guide/getting_started/index.html#installing-eagerx).
Tutorials
---------
The following tutorials are currently available in the form of Google
Colabs:
**Introduction to EAGERx**
- [Tutorial 1: Getting
started](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/icra/getting_started.ipynb)
- [Tutorial 2: Advanced
usage](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/icra/advanced_usage.ipynb)
The solutions are available
[here](https://github.com/eager-dev/eagerx_tutorials/tree/master/tutorials/icra/solutions/).
**Developer tutorials**
- [Tutorial 1: Environment Creation and Training with
EAGERx](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/1_environment_creation.ipynb)
- [Tutorial 2: Reset and Step
Function](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/2_reset_and_step.ipynb)
- [Tutorial 3: Space and
Processors](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/3_space_and_processors.ipynb)
- [Tutorial 4: Nodes and Graph
Validity](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/4_nodes.ipynb)
- [Tutorial 5: Adding Engine Support for an
Object](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/5_engine_implementation.ipynb)
- [Tutorial 6: Defining a new
Object](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/6_objects.ipynb)
- [Tutorial 7: More Informative
Rendering](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/7_rendering.ipynb)
- [Tutorial 8: Reset
Routines](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/8_reset_routine.ipynb)
The solutions are available
[here](https://github.com/eager-dev/eagerx_tutorials/tree/master/tutorials/pendulum/solutions/).
For more information see the
[docs](https://eagerx.readthedocs.io/en/master/guide/tutorials/colabs.html)
or the [eagerx\_tutorials
package](https://github.com/eager-dev/eagerx_tutorials).
Code Example
------------
Below you can find a code example of environment creation and training
using
[Stable-Baselines3](https://stable-baselines3.readthedocs.io/en/master/).
To run this code, you should install
[eagerx\_tutorials](https://github.com/eager-dev/eagerx_tutorials),
which can be done by running:
```bash
pip3 install eagerx_tutorials
```
Detailed explanation of the code can be found in [this Colab
tutorial](https://colab.research.google.com/github/eager-dev/eagerx_tutorials/blob/master/tutorials/pendulum/1_environment_creation.ipynb).
```python
import eagerx
from eagerx.backends.single_process import SingleProcess
from eagerx.wrappers import Flatten
from eagerx_tutorials.pendulum.objects import Pendulum
from eagerx_ode.engine import OdeEngine
import stable_baselines3 as sb3
import numpy as np
from typing import Dict
class PendulumEnv(eagerx.BaseEnv):
def __init__(self, name: str, rate: float, graph: eagerx.Graph, engine: eagerx.specs.EngineSpec,
backend: eagerx.specs.BackendSpec):
self.max_steps = 100
self.steps = None
super().__init__(name, rate, graph, engine, backend, force_start=True)
def step(self, action: Dict):
observation = self._step(action)
self.steps += 1
# Calculate reward and check if the episode is terminated
th = observation["angle"][0]
thdot = observation["angular_velocity"][0]
u = float(action["voltage"])
th -= 2 * np.pi * np.floor((th + np.pi) / (2 * np.pi))
cost = th ** 2 + 0.1 * thdot ** 2 + 0.01 * u ** 2
truncated = self.steps > self.max_steps
terminated = False
# Render
if self.render_mode == "human":
self.render()
return observation, -cost, terminated, truncated, {}
def reset(self, seed=None, options=None) -> Dict:
states = self.state_space.sample()
observation = self._reset(states)
self.steps = 0
# Render
if self.render_mode == "human":
self.render()
return observation, {}
if __name__ == "__main__":
rate = 30.0
pendulum = Pendulum.make("pendulum", actuators=["u"], sensors=["theta", "theta_dot"], states=["model_state"])
graph = eagerx.Graph.create()
graph.add(pendulum)
graph.connect(action="voltage", target=pendulum.actuators.u)
graph.connect(source=pendulum.sensors.theta, observation="angle")
graph.connect(source=pendulum.sensors.theta_dot, observation="angular_velocity")
engine = OdeEngine.make(rate=rate)
backend = SingleProcess.make()
env = PendulumEnv(name="PendulumEnv", rate=rate, graph=graph, engine=engine, backend=backend)
env = Flatten(env)
model = sb3.SAC("MlpPolicy", env, verbose=1)
model.learn(total_timesteps=int(150 * rate))
env.shutdown()
```
Engines
-------
EAGERx allows to create engine agnostic environments such that a single
environment can be used for simulation and reality. The following
engines are available for training and evaluation:
- [RealEngine](https://github.com/eager-dev/eagerx_reality) for
real-world experiments
- [PybulletEngine](https://github.com/eager-dev/eagerx_pybullet) for
PyBullet simulations
- [OdeEngine](https://github.com/eager-dev/eagerx_ode) for simulations
based on ordinary differential equations (ODEs)
Users can also create their own (custom) engines.
Cite EAGERx
-----------
If you are using EAGERx for your scientific publications, please cite:
``` {.sourceCode .bibtex}
@article{vanderheijden2024eagerx,
author={van der Heijden, Bas and Luijkx, Jelle and Ferranti, Laura and Kober, Jens and Babuska, Robert},
journal={IEEE Robotics \& Automation Magazine},
title={Engine Agnostic Graph Environments for Robotics (EAGERx): A Graph-Based Framework for Sim2real Robot Learning},
year={2024},
volume={},
number={},
pages={2-15},
keywords={Robots;Engines;Robot sensing systems;Delays;Robot learning;Physics;Codes},
doi={10.1109/MRA.2024.3433172}
}
```
Maintainers
-----------
EAGERx is currently maintained by Bas van der Heijden
([\@bheijden](https://github.com/bheijden)) and Jelle Luijkx
([\@jelledouwe](https://github.com/jelledouwe)).
How to contact us
-----------------
For any question, send an e-mail to *eagerx.dev@gmail.com*.
Acknowledgements
----------------
EAGERx is funded by the [OpenDR](https://opendr.eu/) Horizon 2020
project.
<p>
<img src="docs/_static/img/tu_delft.png" width="20%" alt="tu_delft"/>
<img src="docs/_static/img/opendr_logo.png" width="20%" alt="opendr"/>
</p>