Govindhasamy, James J. and McLoone, Sean F. and Irwin, George W.
Second-Order Training of Adaptive Critics
for Online Process Control.
IEEE Trans. Systems Man. and Cybernetics, Part B, 35 (2).
This paper deals with reinforcement lear ning for process
modeling and control using a model-free, action- dependent adaptive
critic (ADAC). A new modified recursive Levenberg Marquardt (RLM)
training algorithm, called temporal difference RLM, is developed to
improve the ADAC performance. Novel application results for a simulated
continuously-stirred-tank-reactor process are included to show the superiority
of the new algorithm to conventional temporal-difference stochastic
||Copyright ÃÂ© 2005 IEEE.Â Â Reprinted fromÂ (relevant publication info).
This material is posted here with permission of the IEEE. Such permission
of the IEEE does not in any way imply IEEE endorsement of any of NUI
Maynooth ePrints and eTheses Archive's products or services.Â Internal or
personal use of this material is permitted.Â However, permission
for creating new collective works for resale or redistribution must be
obtained from the IEEE by writing to firstname.lastname@example.org.
copyright laws protecting it.
||Action- dependent Adaptive Critic, Levenberg Marquardt, Online process control
||Science & Engineering > Electronic Engineering
||23 Sep 2008
|Journal or Publication Title:
||IEEE Trans. Systems Man. and Cybernetics, Part B
||Institute of Electrical and Electronics Engineers
Repository Staff Only(login required)
||Item control page