Abstract
This exposition discusses continuous-time reinforcement learning (CT-RL) for the control of affine nonlinear systems. We review four seminal methods that are the centerpieces of the most recent results on CT-RL control. We survey the theoretical results of the four methods, highlighting their fundamental importance and successes by including discussions on problem formulation, key assumptions, algorithm procedures, and theoretical guarantees. Subsequently, we evaluate the performance of the control designs to provide analyses and insights on the feasibility of these design methods for applications from a control designer’s point of view. Through systematic evaluations, we point out when theory diverges from practical controller synthesis. We, furthermore, introduce a new quantitative analytical framework to diagnose the observed discrepancies. Based on the analyses and the insights gained through quantitative evaluations, we point out potential future research directions to unleash the potential of CT-RL control algorithms in addressing the identified challenges.
Original language | English (US) |
---|---|
Pages (from-to) | 1-21 |
Number of pages | 21 |
Journal | IEEE Transactions on Neural Networks and Learning Systems |
DOIs | |
State | Accepted/In press - 2023 |
Keywords
- Adaptive/approximate dynamic programming (ADP)
- Convergence
- Heuristic algorithms
- Mathematical models
- Optimal control
- Power system stability
- Recurrent neural networks
- Tuning
- continuous-time (CT)
- optimal control
- policy iteration (PI)
- reinforcement learning (RL)
- value iteration (VI)
ASJC Scopus subject areas
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence