TY - JOUR
T1 - Applying and dissecting LSTM neural networks and regularized learning for dynamic inferential modeling
AU - LI, Jicheng
AU - QIN, S. Joe
N1 - Publisher Copyright:
© 2023 Elsevier Ltd
PY - 2023/7
Y1 - 2023/7
N2 - Deep learning models such as the long short-term memory (LSTM) network have been applied for dynamic inferential modeling. However, many studies apply LSTM as a black-box approach without examining the necessity and usefulness of the internal LSTM gates for inferential modeling. In this paper, we use LSTM as a state space realization and compare it with linear state space modeling and statistical learning methods, including N4SID, partial least squares, the Lasso, and support vector regression. Two case studies on an industrial 660MW boiler and a debutanizer column process indicate that LSTM underperforms all other methods. LSTM is shown to be capable of outperforming linear methods for a simulated reactor process with severely excited nonlinearity in the data. By dissecting the sub-components of a simple LSTM model, the effectiveness of the LSTM gates and nonlinear activation functions is scrutinized.
AB - Deep learning models such as the long short-term memory (LSTM) network have been applied for dynamic inferential modeling. However, many studies apply LSTM as a black-box approach without examining the necessity and usefulness of the internal LSTM gates for inferential modeling. In this paper, we use LSTM as a state space realization and compare it with linear state space modeling and statistical learning methods, including N4SID, partial least squares, the Lasso, and support vector regression. Two case studies on an industrial 660MW boiler and a debutanizer column process indicate that LSTM underperforms all other methods. LSTM is shown to be capable of outperforming linear methods for a simulated reactor process with severely excited nonlinearity in the data. By dissecting the sub-components of a simple LSTM model, the effectiveness of the LSTM gates and nonlinear activation functions is scrutinized.
KW - Dynamic inferential modeling
KW - LSTM
KW - Partial least squares
KW - Regularized learning
KW - Subspace identification
UR - http://www.scopus.com/inward/record.url?scp=85154581845&partnerID=8YFLogxK
U2 - 10.1016/j.compchemeng.2023.108264
DO - 10.1016/j.compchemeng.2023.108264
M3 - Journal Article (refereed)
SN - 0098-1354
VL - 175
JO - Computers and Chemical Engineering
JF - Computers and Chemical Engineering
M1 - 108264
ER -