-
2
-
-
0003529238
-
-
Ph.D. dissertation, Committee Appl. Math., Harvard Univ., Cambridge, MA
-
P. J. Werbos, "Beyond regression: New tools for prediction and analysis in the behavioral sciences," Ph.D. dissertation, Committee Appl. Math., Harvard Univ., Cambridge, MA, 1974.
-
(1974)
Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences
-
-
Werbos, P.J.1
-
3
-
-
0024888479
-
Neural networks for control and system identification
-
Tampa, FL Dec
-
P. J. Werbos, "Neural networks for control and system identification," in Proc. IEEE Conf. Decis. Control, vol. 1. Tampa, FL, Dec. 1989, pp. 260-265.
-
(1989)
Proc. IEEE Conf. Decis. Control
, vol.1
, pp. 260-265
-
-
Werbos, P.J.1
-
5
-
-
0002031779
-
Approximate dynamic programming for real-time control and neural modeling
-
New York: Van Nostrand
-
P. J. Werbos, "Approximate dynamic programming for real-time control and neural modeling," in Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches, vol. 15. New York: Van Nostrand, 1992, pp. 493-525.
-
(1992)
Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches
, vol.15
, pp. 493-525
-
-
Werbos, P.J.1
-
6
-
-
85046476577
-
-
Boca Raton FL: CRC Press
-
L. Busoniu, R. Babuska, B. De Schutter, and D. Ernst, Reinforcement Learning and Dynamic Programming Using Function Approximators. Boca Raton, FL: CRC Press, 2010.
-
(2010)
Reinforcement Learning and Dynamic Programming Using Function Approximators
-
-
Busoniu, L.1
Babuska, R.2
De Schutter, B.3
Ernst, D.4
-
7
-
-
34547098844
-
Kernel-based least squares policy iteration for reinforcement learning
-
DOI 10.1109/TNN.2007.899161, Neural Networks for Feedback Control Systems
-
X. Xu, D. Hu, and X. Lu, "Kernel-based least squares policy iteration for reinforcement learning," IEEE Trans. Neural Netw., vol. 18, no. 4, pp. 973-992, Jul. 2007. (Pubitemid 47098876)
-
(2007)
IEEE Transactions on Neural Networks
, vol.18
, Issue.4
, pp. 973-992
-
-
Xu, X.1
Hu, D.2
Lu, X.3
-
9
-
-
67650458822
-
Basis function adaptation methods for cost approximation in MDP
-
Learn., Mar.-Apr
-
H. Yu and D. Bertsekas, "Basis function adaptation methods for cost approximation in MDP," in Proc. IEEE Symp. Adapt. Dyn. Program. Reinforce. Learn., Mar.-Apr. 2009, pp. 74-81.
-
(2009)
Proc. IEEE Symp. Adapt. Dyn. Program. Reinforce
, pp. 74-81
-
-
Yu, H.1
Bertsekas, D.2
-
10
-
-
70349116541
-
Reinforcement learning and adaptive dynamic programming for feedback control
-
Sep
-
F. L. Lewis and D. Vrabie, "Reinforcement learning and adaptive dynamic programming for feedback control," IEEE Circuits Syst. Mag., vol. 9, no. 3, pp. 32-50, Sep. 2009.
-
(2009)
IEEE Circuits Syst. Mag.
, vol.9
, Issue.3
, pp. 32-50
-
-
Lewis, F.L.1
Vrabie, D.2
-
11
-
-
0028584964
-
Adaptive linear quadratic control using policy iteration
-
Jul
-
S. J. Bradtke, B. E. Ydstie, and A. G. Barto, "Adaptive linear quadratic control using policy iteration," in Proc. Amer. Control Conf., vol. 3. Jul. 1994, pp. 3475-3479.
-
(1994)
Proc. Amer. Control Conf.
, vol.3
, pp. 3475-3479
-
-
Bradtke, S.J.1
Ydstie, B.E.2
Barto, A.G.3
-
12
-
-
58349110975
-
Adaptive optimal control for continuous-time linear systems based on policy iteration
-
Feb
-
D. Vrabie, O. Pastravanu, M. Abu-Khalaf, and F. L. Lewis, "Adaptive optimal control for continuous-time linear systems based on policy iteration," Automatica, vol. 45, no. 2, pp. 477-484, Feb. 2009.
-
(2009)
Automatica
, vol.45
, Issue.2
, pp. 477-484
-
-
Vrabie, D.1
Pastravanu, O.2
Abu-Khalaf, M.3
Lewis, F.L.4
-
13
-
-
78650246160
-
Approximate dynamic programming for output feedback control
-
Beijing, China, Jul
-
Y. Jiang and Z. P. Jiang, "Approximate dynamic programming for output feedback control," in Proc. 29th Chin. Control Conf., Beijing, China, Jul. 2010, pp. 5815-5820.
-
(2010)
Proc. 29th Chin. Control Conf.
, pp. 5815-5820
-
-
Jiang, Y.1
Jiang, Z.P.2
-
14
-
-
79551685808
-
Reinforcement learning for partially observable dynamic processes: Adaptive dynamic programming using measured output data
-
Feb
-
F. L. Lewis and K. G. Vamvoudakis, "Reinforcement learning for partially observable dynamic processes: Adaptive dynamic programming using measured output data," IEEE Trans. Syst., Man, Cybern., Part B: Cybern., vol. 41, no. 1, pp. 14-25, Feb. 2011.
-
(2011)
IEEE Trans. Syst., Man, Cybern., Part B: Cybern.
, vol.41
, Issue.1
, pp. 14-25
-
-
Lewis, F.L.1
Vamvoudakis, K.G.2
-
15
-
-
33846781129
-
Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control
-
DOI 10.1016/j.automatica.2006.09.019, PII S0005109806004249
-
A. Al-Tamimi, F. L. Lewis, and M. Abu-Khalaf, "Model-free Q-learning designs for linear discrete-time zero-sum games with application to Hinfinity control," Automatica, vol. 43, no. 3, pp. 473-481, Mar. 2007. (Pubitemid 46209050)
-
(2007)
Automatica
, vol.43
, Issue.3
, pp. 473-481
-
-
Al-Tamimi, A.1
Lewis, F.L.2
Abu-Khalaf, M.3
-
16
-
-
49049089962
-
Discrete-time nonlinear HJB solution using approximate dynamic programming: Convergence proof
-
Aug
-
A. Al-Tamimi, F. L. Lewis, and M. Abu-Khalaf, "Discrete-time nonlinear HJB solution using approximate dynamic programming: Convergence proof," IEEE Trans. Syst., Man, Cybern., Part B: Cybern., vol. 38, no. 4, pp. 943-949, Aug. 2008.
-
(2008)
IEEE Trans. Syst., Man, Cybern., Part B: Cybern.
, vol.38
, Issue.4
, pp. 943-949
-
-
Al-Tamimi, A.1
Lewis, F.L.2
Abu-Khalaf, M.3
-
17
-
-
0014612678
-
Optimal stationary control of linear systems with controldependent noise
-
Dec
-
D. Kleinman, "Optimal stationary control of linear systems with controldependent noise," IEEE Trans. Autom. Control, vol. 14, no. 6, pp. 673-677, Dec. 1969.
-
(1969)
IEEE Trans. Autom. Control
, vol.14
, Issue.6
, pp. 673-677
-
-
Kleinman, D.1
-
18
-
-
66449130966
-
Adaptive dynamic programming: An introduction
-
May
-
F. Wang, H. Zhang, and D. Liu, "Adaptive dynamic programming: An introduction," IEEE Comp. Intell. Mag., vol. 4, no. 2, pp. 39-47, May 2009.
-
(2009)
IEEE Comp. Intell. Mag.
, vol.4
, Issue.2
, pp. 39-47
-
-
Wang, F.1
Zhang, H.2
Liu, D.3
|