Movatterモバイル変換


[0]ホーム

URL:


Skip to main content
Cornell University
We gratefully acknowledge support from the Simons Foundation,member institutions, and all contributors.Donate
arxiv logo>cs> arXiv:2306.09852
arXiv logo
Cornell University Logo

Computer Science > Robotics

arXiv:2306.09852 (cs)
[Submitted on 16 Jun 2023 (v1), last revised 5 Feb 2025 (this version, v7)]

Title:Actor-Critic Model Predictive Control: Differentiable Optimization meets Reinforcement Learning

View PDFHTML (experimental)
Abstract:An open research question in robotics is how to combine the benefits of model-free reinforcement learning (RL) - known for its strong task performance and flexibility in optimizing general reward formulations - with the robustness and online replanning capabilities of model predictive control (MPC). This paper provides an answer by introducing a new framework called Actor-Critic Model Predictive Control. The key idea is to embed a differentiable MPC within an actor-critic RL framework. This integration allows for short-term predictive optimization of control actions through MPC, while leveraging RL for end-to-end learning and exploration over longer horizons. Through various ablation studies, we expose the benefits of the proposed approach: it achieves better out-of-distribution behaviour, better robustness to changes in the dynamics and improved sample efficiency. Additionally, we conduct an empirical analysis that reveals a relationship between the critic's learned value function and the cost function of the differentiable MPC, providing a deeper understanding of the interplay between the critic's value and the MPC cost functions. Finally, we validate our method in a drone racing task on different tracks, in both simulation and the real world. Our method achieves the same superhuman performance as state-of-the-art model-free RL, showcasing speeds of up to 21 m/s. We show that the proposed architecture can achieve real-time control performance, learn complex behaviors via trial and error, and retain the predictive properties of the MPC to better handle out-of-distribution behavior.
Comments:18 pages, 12 figures, extension
Subjects:Robotics (cs.RO)
ACM classes:I.2.9; I.2.6; G.1.6; I.2.8; C.3
Cite as:arXiv:2306.09852 [cs.RO]
 (orarXiv:2306.09852v7 [cs.RO] for this version)
 https://doi.org/10.48550/arXiv.2306.09852
arXiv-issued DOI via DataCite

Submission history

From: Angel Romero [view email]
[v1] Fri, 16 Jun 2023 14:06:16 UTC (26,688 KB)
[v2] Mon, 18 Sep 2023 14:45:05 UTC (28,988 KB)
[v3] Thu, 21 Sep 2023 12:27:05 UTC (28,988 KB)
[v4] Wed, 28 Feb 2024 14:53:49 UTC (3,575 KB)
[v5] Fri, 12 Apr 2024 13:24:20 UTC (3,575 KB)
[v6] Wed, 27 Nov 2024 08:01:21 UTC (10,450 KB)
[v7] Wed, 5 Feb 2025 10:45:52 UTC (7,769 KB)
Full-text links:

Access Paper:

Current browse context:
cs.RO
Change to browse by:
export BibTeX citation

Bookmark

BibSonomy logoReddit logo

Bibliographic and Citation Tools

Bibliographic Explorer(What is the Explorer?)
Connected Papers(What is Connected Papers?)
scite Smart Citations(What are Smart Citations?)

Code, Data and Media Associated with this Article

CatalyzeX Code Finder for Papers(What is CatalyzeX?)
Hugging Face(What is Huggingface?)
Papers with Code(What is Papers with Code?)

Demos

Hugging Face Spaces(What is Spaces?)

Recommenders and Search Tools

Influence Flower(What are Influence Flowers?)
CORE Recommender(What is CORE?)

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community?Learn more about arXivLabs.

Which authors of this paper are endorsers? |Disable MathJax (What is MathJax?)

[8]ページ先頭

©2009-2025 Movatter.jp