TY - JOUR
T1 - Predictive deep reinforcement learning with multi-agent systems for adaptive time series forecasting
AU - SHAIK, Thanveer
AU - TAO, Xiaohui
AU - LI, Lin
AU - XIE, Haoran
AU - ACHARYA, U.R.
AU - GURURAJAN, Raj
AU - ZHOU, Xujuan
N1 - Publisher Copyright:
© 2025
PY - 2025/9/27
Y1 - 2025/9/27
N2 - Reinforcement learning has been increasingly applied in monitoring applications because of its ability to learn from previous experiences and make adaptive decisions. However, existing machine learning-based health monitoring applications are mostly supervised learning algorithms, trained on labels, and they cannot make adaptive decisions in an uncertain, complex environment. This study proposes a novel and generic system, predictive deep reinforcement learning (PDRL), with multiple RL agents in a time series forecasting environment. The proposed generic framework accommodates virtual Deep Q Network (DQN) agents to monitor predicted future states of a complex environment with a well-defined reward policy so that the agent learns existing knowledge while maximizing their rewards. In the evaluation process of the proposed framework, three DRL agents were deployed to monitor a subject’s future heart rate, respiration, and temperature predicted using a BiLSTM model. With each iteration, the three agents were able to learn the associated patterns, and their cumulative rewards gradually increased. It outperformed the baseline models for all three monitoring agents. The proposed PDRL framework achieves state-of-the-art performance in time series forecasting by effectively integrating reinforcement learning agents with deep learning-based prediction. The proposed DRL agents and deep learning model in the PDRL framework are customized to enable transfer learning in other forecasting applications like traffic and weather, and monitor their states. The PDRL framework is able to learn the future states of the traffic and weather forecasting, and the cumulative rewards are gradually increasing over each episode.
AB - Reinforcement learning has been increasingly applied in monitoring applications because of its ability to learn from previous experiences and make adaptive decisions. However, existing machine learning-based health monitoring applications are mostly supervised learning algorithms, trained on labels, and they cannot make adaptive decisions in an uncertain, complex environment. This study proposes a novel and generic system, predictive deep reinforcement learning (PDRL), with multiple RL agents in a time series forecasting environment. The proposed generic framework accommodates virtual Deep Q Network (DQN) agents to monitor predicted future states of a complex environment with a well-defined reward policy so that the agent learns existing knowledge while maximizing their rewards. In the evaluation process of the proposed framework, three DRL agents were deployed to monitor a subject’s future heart rate, respiration, and temperature predicted using a BiLSTM model. With each iteration, the three agents were able to learn the associated patterns, and their cumulative rewards gradually increased. It outperformed the baseline models for all three monitoring agents. The proposed PDRL framework achieves state-of-the-art performance in time series forecasting by effectively integrating reinforcement learning agents with deep learning-based prediction. The proposed DRL agents and deep learning model in the PDRL framework are customized to enable transfer learning in other forecasting applications like traffic and weather, and monitor their states. The PDRL framework is able to learn the future states of the traffic and weather forecasting, and the cumulative rewards are gradually increasing over each episode.
KW - Behavior pattern
KW - Decision making
KW - Monitoring
KW - Reinforcement learning
KW - Time series forecasting
UR - https://www.scopus.com/pages/publications/105010182638
U2 - 10.1016/j.knosys.2025.113941
DO - 10.1016/j.knosys.2025.113941
M3 - Journal Article (refereed)
SN - 0950-7051
VL - 326
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 113941
ER -