TY - JOUR
T1 - On generalization error of neural network models and its application to predictive control of nonlinear processes
AU - Alhajeri, Mohammed S.
AU - Alnajdi, Aisha
AU - Abdullah, Fahim
AU - Christofides, Panagiotis D.
N1 - Publisher Copyright:
© 2022 Institution of Chemical Engineers
PY - 2023/1
Y1 - 2023/1
N2 - In order to approximate nonlinear dynamic systems utilizing time-series data, recurrent neural networks (RNNs) and long short-term memory (LSTM) networks have frequently been used. The training error of neural networks may often be made suitably modest; however, the accuracy can be further improved by incorporating prior knowledge in the construction of machine learning-based models. Specifically, physics-based RNN modeling has yielded more reliable RNN models than traditional RNNs. Yet, a framework for constructing and assessing the generalization ability of such RNN models as well as LSTM models to be utilized in model predictive control (MPC) systems is lacking. In this work, we develop a methodological framework to quantify the generalization error bounds for partially-connected RNNs and LSTM models. The partially-connected RNN model is then utilized to predict the state evolution in a MPC scheme. We illustrate through open-loop and closed-loop simulations of a nonlinear chemical process of two reactors-in-series that the proposed approach provides a flexible framework for leveraging both prior knowledge and data, thereby improving the performance significantly when compared to a fully-connected modeling approach under Lyapunov-based MPC.
AB - In order to approximate nonlinear dynamic systems utilizing time-series data, recurrent neural networks (RNNs) and long short-term memory (LSTM) networks have frequently been used. The training error of neural networks may often be made suitably modest; however, the accuracy can be further improved by incorporating prior knowledge in the construction of machine learning-based models. Specifically, physics-based RNN modeling has yielded more reliable RNN models than traditional RNNs. Yet, a framework for constructing and assessing the generalization ability of such RNN models as well as LSTM models to be utilized in model predictive control (MPC) systems is lacking. In this work, we develop a methodological framework to quantify the generalization error bounds for partially-connected RNNs and LSTM models. The partially-connected RNN model is then utilized to predict the state evolution in a MPC scheme. We illustrate through open-loop and closed-loop simulations of a nonlinear chemical process of two reactors-in-series that the proposed approach provides a flexible framework for leveraging both prior knowledge and data, thereby improving the performance significantly when compared to a fully-connected modeling approach under Lyapunov-based MPC.
KW - Generalization error
KW - Long short-term memory
KW - Machine learning
KW - Model predictive control
KW - Nonlinear systems
KW - Partially-connected RNN
KW - Recurrent neural networks
UR - http://www.scopus.com/inward/record.url?scp=85143775405&partnerID=8YFLogxK
U2 - 10.1016/j.cherd.2022.12.001
DO - 10.1016/j.cherd.2022.12.001
M3 - Article
AN - SCOPUS:85143775405
SN - 0263-8762
VL - 189
SP - 664
EP - 679
JO - Chemical Engineering Research and Design
JF - Chemical Engineering Research and Design
ER -