December, 2022
Extended the WBC framework to handle the full robot dynamics and challenging terrains, and to generalize beyond rigid contacts.
Investigated the effects of the standard state estimators on non-rigid contacts.
Developed a vision-based planning approach that plans footholds and body poses based on the robot's "learned" skills.
Developing the WBC of the MIT Humanoid.
Building up the RL infrastructure for Mini-Cheetah.
Developing RL controllers for Mini-Cheetah and the MIT Humanoid.
Investigating data-driven approaches in state estimation.
December, 2022
"Marc Raibert divided intelligence into Cognitive Intelligence (CI) and Athletic Intelligence (AtI). CI allows us to make abstract plans, and to understand and solve broader problems. AtI allows us to operate our bodies in such a way that we can balance, stand, walk, climb, etc. AtI also lets us do real-time perception so that we can interact with the world around us."
S. Fahmi, On Terrain-Aware Locomotion for Legged Robots, PhD. dissertation, Italian Institute of Technology, 2021.
M. Raibert and S. Kuindersma, Boston dynamics, the Challenges of Real-World Reinforcement Learning Workshop in NeurIPS, 2021.
Using Optimization and Learning
Control
State Estimation
Planning
Proprioception
Vision
Physical Properties
Geometry
STANCE: Whole-Body Control for Legged Robots
ViTAL: Vision-Based Terrain-Aware Locomotion Planning
MIMOC: Reinforcement Learning for Legged Robots
Standard WBC (sWBC)
Reformulate the sWBC to account for the contact dynamics
C3 = compliant contact consistent
How will the new WBC know about the current terrain impedance parameters?
C3 = compliant contact consistent
Feedback the terrain impedance parameters using a Terrain Compliance Estimator (TCE)
C3 = compliant contact consistent
Simultaneously, at every control loop
† S. Fahmi, M. Focchi, A. Radulescu, G. Fink, V. Barasuol, and C. Semini,
“STANCE: Locomotion Adaptation over Soft Terrain,” IEEE Trans. Robot. (T-RO), vol. 36, no. 2, pp. 443–457, Apr. 2020, doi: 10.1109/TRO.2019.2954670.
S. Fahmi, C. Mastalli, M. Focchi, and C. Semini, “Passive Whole-Body Control for Quadruped Robots: Experimental Validation over Challenging Terrain,”
IEEE Robot. Automat. Lett. (RA-L), vol. 4, no. 3, pp. 2553–2560, Jul. 2019,
doi: 10.1109/LRA.2019.2908502.
† Finalist for the IEEE RAS Italian Chapter Young Author Best Paper Award, and for the IEEE RAS Technical Committee on Model-Based Optimization for Robotics Best Paper Award.
Plan the feet & base states simultaneously.
Separate the plan into feet & base plans.
Planning the feet motion ⇒ foothold selection
Planning the body motion ⇒ pose adaptation
Villarreal 2019
Barasuol 2015
V. Barasuol et al., Reactive Trotting with Foot Placement Corrections through Visual Pattern Classification, IROS, 2015.
O. Villarreal et al., Fast and Continuous Foothold Adaptation for Dynamic Locomotion through CNNs, RA-L, 2019.
M. Kalakrishnan et al., Learning, planning, and control for quadruped locomotion over challenging terrain, IJRR 2011.
P. Fankhauser et al., Robust Rough-Terrain Locomotion with a Quadrupedal Robot, IROS, 2018.
Kalakrishnan 2011
Fankhauser 2018
Propose a different paradigm.
Instead of finding body poses that are optimal w.r.t given footholds,
we find body poses that will maximize the chances of the robot reaching safe footholds.
Put the robot in a state that if it gets disturbed, it remains around a set of solutions that are still safe & reachable.
Vision-Based Foothold Adaptation (VFA)
Vision-Based Pose Adaptation (VPA)
Foothold Evaluation Criteria (FEC)
What the robot is capable of doing?
Rejects footholds that cause foot trajectory collision.
Rejects footholds that are outside the workspace during the entire gait phase.
Rejects footholds that cause leg collision with the terrain.
Rejects footholds that are near holes, spikes, edges, and corners.
front hips rising up
front hips lowering down
hind hips rising up
hind hips lowering down
S. Fahmi, V. Barasuol, D. Esteban, O. Villarreal, and C. Semini,
“ViTAL: Vision-Based Terrain-Aware Locomotion for Legged Robots,” IEEE Trans. Robot. (T-RO), Nov. 2022, doi: 10.1109/TRO.2022.3222958.
[1] N. Rudin et al., Learning to walk in minutes using massively parallel deep reinforcement learning, CoRL, 2021.
[2] G. Ji et al., Concurrent training of a control policy and a state estimator for dynamic and robust legged locomotion,” RA-L, 2022.
[3] J. Siekmann et al., Sim-to-real learning of all common bipedal gaits via periodic reward composition, ICRA, 2021
[4] X. B. Peng et al., Learning agile robotic locomotion skills by imitating animals, RSS 2020.
[5] X. B. Peng, et al., Deepmimic: Example-guided deep reinforcement learning of physics-based character skills,” ACM Trans. Graphics, 2018.
D. Kim et al., Highly dynamic quadruped locomotion via whole-body impulse control and model predictive control, ArXiV, 2019.
N. Rudin et al., Learning to walk in minutes using massively parallel deep reinforcement learning, CoRL, 2021.
A.J. Miller*, S. Fahmi*, M. Chignoli, and S. Kim, "Reinforcement Learning for Legged Robots: Motion Imitation from Model-Based Optimal Control," under review, in IEEE Int. Conf. Robot. Automat. (ICRA), 2023.
*equal contribution
WBC vs. MPC
Hierarchical WBC (priority-based WBC)
Models in Optimization-Based Control (SRBD, Floating-Based Models, Full Model, etc.)
QP Solvers
Contact Models
WBC for Humanoids, does it differ from WBC for quadrupeds?
RL-Based Control vs. Model-Based Control
State Estimation for Legged Robots
Vision-Based Planning
Learning for Legged Robots