Accelerated Reinforcement Learning via Dynamic Mode Decomposition

Vrushabh S. Donge, Bosen Lian, Frank L. Lewis, Ali Davoudi

Research output: Contribution to journalArticlepeer-review

5 Scopus citations

Abstract

This work applies the decomposition principle to discrete-time reinforcement learning (RL) to solve the optimal control problems for a network of subsystems. The control design is defined as a linear quadratic regulator graphical problem, where the performance function couples the subsystems' dynamics. We first present a model-free discrete-time RL algorithm based on online behaviors without using system dynamics. This could become a prohibitively long learning process for larger networks. To remedy this issue, we develop an efficient model-free RL algorithm based on dynamic mode decomposition. This decomposition method reduces the size of the measured data while the dynamic information of the original network is still retained. This algorithm is then implemented online. The proposed methodology is validated using examples of a consensus network and a power system network.

Original languageEnglish
Article number3259060
Pages (from-to)2022-2034
Number of pages13
JournalIEEE Transactions on Control of Network Systems
Volume10
Issue number4
DOIs
StatePublished - Dec 1 2023
Externally publishedYes

Keywords

  • Dynamic mode decomposition
  • large-scale systems
  • optimal control
  • reinforcement learning (RL)

Fingerprint

Dive into the research topics of 'Accelerated Reinforcement Learning via Dynamic Mode Decomposition'. Together they form a unique fingerprint.

Cite this