June, 2023
"Marc Raibert divided intelligence into Cognitive Intelligence (CI) and Athletic Intelligence (AtI). CI allows us to make abstract plans, and to understand and solve broader problems. AtI allows us to operate our bodies in such a way that we can balance, stand, walk, climb, etc. AtI also lets us do real-time perception so that we can interact with the world around us."
S. Fahmi, On Terrain-Aware Locomotion for Legged Robots, PhD. dissertation, Italian Institute of Technology, 2021.
M. Raibert and S. Kuindersma, Boston dynamics, the Challenges of Real-World Reinforcement Learning Workshop in NeurIPS, 2021.
Using Optimization and Learning
Control
State Estimation
Planning
Proprioception
Vision
Physical Properties
Geometry
STANCE: Whole-Body Control for Legged Robots
ViTAL: Vision-Based Terrain-Aware Locomotion Planning
MIMOC: Reinforcement Learning for Legged Robots
Plan the feet & base states simultaneously.
Separate the plan into feet & base plans.
Planning the feet motion ⇒ foothold selection
Planning the body motion ⇒ pose adaptation
Villarreal 2019
Barasuol 2015
V. Barasuol et al., Reactive Trotting with Foot Placement Corrections through Visual Pattern Classification, IROS, 2015.
O. Villarreal et al., Fast and Continuous Foothold Adaptation for Dynamic Locomotion through CNNs, RA-L, 2019.
M. Kalakrishnan et al., Learning, planning, and control for quadruped locomotion over challenging terrain, IJRR 2011.
P. Fankhauser et al., Robust Rough-Terrain Locomotion with a Quadrupedal Robot, IROS, 2018.
Kalakrishnan 2011
Fankhauser 2018
Propose a different paradigm.
Instead of finding body poses that are optimal w.r.t given footholds,
we find body poses that will maximize the chances of the robot reaching safe footholds.
Put the robot in a state that if it gets disturbed, it remains around a set of solutions that are still safe & reachable.
Vision-Based Foothold Adaptation (VFA)
Vision-Based Pose Adaptation (VPA)
Foothold Evaluation Criteria (FEC)
What the robot is capable of doing?
Rejects footholds that cause foot trajectory collision.
Rejects footholds that are outside the workspace during the entire gait phase.
Rejects footholds that cause leg collision with the terrain.
Rejects footholds that are near holes, spikes, edges, and corners.
First, we extract heightmaps that are centered around the projection of the leg hip location.
Then, the pose evaluation stage will evaluate the criteria for all the samples of the hip heights.
The pose evaluation stage will output samples of number of safe footholds.
The output of the function approximation is
These samples are then used to approximate a continuous function of the number of safe footholds at the function approximation stage.
Finally, the pose optimization stage finds the optimal pose that maximizes for all of the legs
Height
Roll
Pitch
front hips rising up
front hips lowering down
hind hips rising up
hind hips lowering down
O. Villarreal et al., MPC-based Controller with Terrain Insight for Dynamic Legged Locomotion, ICRA, 2020.
We presented ViTAL, a decoupled visual planning strategy
ViTAL consisted of :
VFA for foothold selection
VPA for pose adaptation
Introduced a different paradigm in pose adaptation
Finds the body poses that maximize the chances of the legs to succeed in reaching safe footholds
FEC (notion of safety) that characterizes the robot's capabilities
ViTAL allows HyQ and HyQReal to traverse a wide variety of terrains under various forward velocities.
These terrains included stairs, gaps, etc.
Forward velocities varied from 0.2m/s to 0.75 m/s.
We showed that the VPA resulted in body poses that are aware of the terrain, and of what the robot and its legs can do.
The VPA puts the robot in a pose that provides the feet with higher number of safe footholds which allows the robot to succeed unlike the baseline that fails.
S. Fahmi, V. Barasuol, D. Esteban, O. Villarreal, and C. Semini,
“ViTAL: Vision-Based Terrain-Aware Locomotion for Legged Robots,” IEEE Trans. Robot. (T-RO), Nov. 2022, doi: 10.1109/TRO.2022.3222958.
[1] N. Rudin et al., Learning to walk in minutes using massively parallel deep reinforcement learning, CoRL, 2021.
[2] G. Ji et al., Concurrent training of a control policy and a state estimator for dynamic and robust legged locomotion,” RA-L, 2022.
[3] J. Siekmann et al., Sim-to-real learning of all common bipedal gaits via periodic reward composition, ICRA, 2021
[4] X. B. Peng et al., Learning agile robotic locomotion skills by imitating animals, RSS 2020.
[5] X. B. Peng, et al., Deepmimic: Example-guided deep reinforcement learning of physics-based character skills,” ACM Trans. Graphics, 2018.
D. Kim et al., Highly dynamic quadruped locomotion via whole-body impulse control and model predictive control, ArXiV, 2019.
N. Rudin et al., Learning to walk in minutes using massively parallel deep reinforcement learning, CoRL, 2021.
A.J. Miller, S. Fahmi, M. Chignoli, and S. Kim,
"Reinforcement Learning for Legged Robots: Motion Imitation from Model-Based Optimal Control," ArXiV preprint, May 2023.