Learning passive policies with virtual energy tanks in robotics

Riccardo Zanella*, Gianluca Palli, Stefano Stramigioli, Federico Califano

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

2 Citations (Scopus)
73 Downloads (Pure)

Abstract

Within a robotic context, the techniques of passivity-based control and reinforcement learning are merged with the goal of eliminating some of their reciprocal weaknesses, as well as inducing novel promising features in the resulting framework. The contribution is framed in a scenario where passivity-based control is implemented by means of virtual energy tanks, a control technique developed to achieve closed-loop passivity for any arbitrary control input. Albeit the latter result is heavily used, it is discussed why its practical application at its current stage remains rather limited, which makes contact with the highly debated claim that passivity-based techniques are associated with a loss of performance. The use of reinforcement learning allows to learn a control policy that can be passivized using the energy tank architecture, combining the versatility of learning approaches and the system theoretic properties which can be inferred due to the energy tanks. Simulations show the validity of the approach, as well as novel interesting research directions in energy-aware robotics.
Original languageEnglish
Pages (from-to)541-550
Number of pages10
JournalIET Control Theory and Applications
Volume18
Issue number5
Early online date25 Jan 2024
DOIs
Publication statusPublished - Mar 2024

Keywords

  • Energy consumption
  • Learning (artificial intelligence)
  • Robots
  • Stability

Fingerprint

Dive into the research topics of 'Learning passive policies with virtual energy tanks in robotics'. Together they form a unique fingerprint.

Cite this