Abstract
In order to approximate nonlinear dynamic systems utilizing time-series data, recurrent neural networks (RNNs) and long short-term memory (LSTM) networks have frequently been used. The training error of neural networks may often be made suitably modest; however, the accuracy can be further improved by incorporating prior knowledge in the construction of machine learning-based models. Specifically, physics-based RNN modeling has yielded more reliable RNN models than traditional RNNs. Yet, a framework for constructing and assessing the generalization ability of such RNN models as well as LSTM models to be utilized in model predictive control (MPC) systems is lacking. In this work, we develop a methodological framework to quantify the generalization error bounds for partially-connected RNNs and LSTM models. The partially-connected RNN model is then utilized to predict the state evolution in a MPC scheme. We illustrate through open-loop and closed-loop simulations of a nonlinear chemical process of two reactors-in-series that the proposed approach provides a flexible framework for leveraging both prior knowledge and data, thereby improving the performance significantly when compared to a fully-connected modeling approach under Lyapunov-based MPC.
Original language | English |
---|---|
Pages (from-to) | 664-679 |
Number of pages | 16 |
Journal | Chemical Engineering Research and Design |
Volume | 189 |
DOIs | |
State | Published - Jan 2023 |
Funding
M. S. Alhajeri would like to express his genuine appreciation for the funding from KFAS . The first and the second authors acknowledge Kuwait University support via the KU-scholarship program. Financial support from the National Science Foundation is also gratefully acknowledged. We would like to thank Dr. Minshuo Chen for pointing out an important remark used in the LSTM generalization error bound proof. The authors would also like to thank Professor Zhe Wu for his support in the development of the theory for generalization error bounds used in this work. M. S. Alhajeri would like to express his genuine appreciation for the funding from KFAS. The first and the second authors acknowledge Kuwait University support via the KU-scholarship program. Financial support from the National Science Foundation is also gratefully acknowledged. We would like to thank Dr. Minshuo Chen for pointing out an important remark used in the LSTM generalization error bound proof. The authors would also like to thank Professor Zhe Wu for his support in the development of the theory for generalization error bounds used in this work.
Keywords
- Generalization error
- Long short-term memory
- Machine learning
- Model predictive control
- Nonlinear systems
- Partially-connected RNN
- Recurrent neural networks