To summarize, this project aims at combining the traditional MPC-based and terrain-aware foothold planner with the deep reinforcement learning(DRL) . The goal is to achieve robust control in extremely risky terrains such as stepping stone.
You can find the modifications in legged_robot_dtc.py
and legged_robot_config.py
.
In this project, we adapt a method similar to TAMOLS and Mini-Cheetah.
An estimated foothold will firstly be calculated by the formula:
where
The centrifugal term is omitted.
However, we choose the footholds solely based on quantitative score from various aspects (distance current pos, terrain variance/gradient, support area etc.), rather than solving a optimization problem.
We use the framework from isaac-gym, with PPO algorithm. With the following feature added:
-
Remove teacher-student framework
-
Add GRU and CE-net as terrain encoder. Latent dimension was increased from 64 to 512.
-
TODO: symmetric data augmentation
To integrate the foothold into DRL, the relative position to the optimized foothold was fed as observations for both actor and critic network. Moreover, a sparse reward term was also added, which will be triggered in the touch-down time.
Estimated training time is 10 hours.
pip install -e rsl_rl
pip install -e .