Document Type

Article

Publication Date

6-23-2025

Department

Department of Electrical and Computer Engineering

Abstract

Model Predictive Control (MPC) and Reinforcement Learning (RL) are two prominent strategies for controlling legged robots. RL learns control policies through system interaction, adapting to various scenarios, whereas MPC relies on a predefined mathematical model to solve optimization problems in real-time. Despite their widespread use, there is a lack of direct comparative analysis under standardized conditions. This work addresses this gap by benchmarking MPC and RL controllers on a Unitree Go1 quadruped robot within the MuJoCo simulation environment, focusing on a standardized task, straight walking at a constant velocity. Performance is evaluated based on disturbance rejection, energy efficiency, and terrain adaptability. The results show that RL excels in handling disturbances and maintaining energy efficiency but struggles with generalization to new terrains due to its dependence on learned policies tailored to specific environments. In contrast, MPC shows enhanced recovery capabilities from larger perturbations by leveraging its optimization-based approach, allowing for a balanced distribution of control efforts across the robot’s joints. The results present the advantages and limitations of both RL and MPC, offering insights into selecting an appropriate control strategy for legged robotic applications.

Publisher's Statement

© 2025 The Authors. Publisher’s version of record:https://doi.org/10.1109/ACCESS.2025.3582523

Publication Title

IEEE Access

Creative Commons License

Creative Commons Attribution 4.0 International License
This work is licensed under a Creative Commons Attribution 4.0 International License.

Version

Publisher's PDF

Share

COinS