-
3
-
-
0036832953
-
Variable-resolution discretization in optimal control
-
R. Munos and A. Moore, "Variable-resolution discretization in optimal control," Machine Learning, Vol. 49, no. 2-3, pp. 291-323, 2002.
-
(2002)
Machine Learning
, vol.49
, Issue.2-3
, pp. 291-323
-
-
Munos, R.1
Moore, A.2
-
5
-
-
21844465127
-
Tree-based batch mode reinforcement learning
-
D. Ernst, P. Geurts, and L. Wehenkel, "Tree-based batch mode reinforcement learning," Journal of Machine Learning Research, Vol. 6, pp. 503-556, 2005. (Pubitemid 40958851)
-
(2005)
Journal of Machine Learning Research
, vol.6
-
-
Ernst, D.1
Geurts, P.2
Wehenkel, L.3
-
6
-
-
35748957806
-
Proto-value functions: A Laplacian framework for learning representation and control in Markov decision processes
-
S. Mahadevan and M. Maggioni, "Proto-value functions: A Laplacian framework for learning representation and control in Markov decision processes," Journal of Machine Learning Research, Vol. 8, pp. 2169-2231, 2007. (Pubitemid 350046199)
-
(2007)
Journal of Machine Learning Research
, vol.8
, pp. 2169-2231
-
-
Mahadevan, S.1
Maggioni, M.2
-
7
-
-
84898939480
-
Policy gradient methods for reinforcement learning with function approximation
-
S. A. Solla, T. K. Leen, and K.-R. Müller, Eds. MIT Press
-
R. S. Sutton, D. A. McAllester, S. P. Singh, and Y. Mansour, "Policy gradient methods for reinforcement learning with function approximation," in Advances in Neural Information Processing Systems 12, S. A. Solla, T. K. Leen, and K.-R. Müller, Eds. MIT Press, 2000, pp. 1057-1063.
-
(2000)
Advances in Neural Information Processing Systems
, vol.12
, pp. 1057-1063
-
-
Sutton, R.S.1
Mc Allester, D.A.2
Singh, S.P.3
Mansour, Y.4
-
9
-
-
33646399442
-
Policy gradient in continuous time
-
R. Munos, "Policy gradient in continuous time," Journal of Machine Learning Research, Vol. 7, pp. 771-791, 2006.
-
(2006)
Journal of Machine Learning Research
, vol.7
, pp. 771-791
-
-
Munos, R.1
-
10
-
-
1942516890
-
The cross-entropy method for fast policy search
-
US, 21-24 August
-
S. Mannor, R. Y. Rubinstein, and Y. Gat, "The cross-entropy method for fast policy search," in Proceedings 20th International Conference on Machine Learning (ICML-03), Washington, US, 21-24 August 2003, pp. 512-519.
-
(2003)
Proceedings 20th International Conference on Machine Learning (ICML-03), Washington
, pp. 512-519
-
-
Mannor, S.1
Rubinstein, R.Y.2
Gat, Y.3
-
12
-
-
33646714634
-
Evolutionary function approximation for reinforcement learning
-
S. Whiteson and P. Stone, "Evolutionary function approximation for reinforcement learning," Journal of Machine Learning Research, Vol. 7, pp. 877-917, 2006. (Pubitemid 43736560)
-
(2006)
Journal of Machine Learning Research
, vol.7
, pp. 877-917
-
-
Whiteson, S.1
Stone, P.2
-
13
-
-
3543089271
-
The cross entropy method
-
M. Jordan, J. Kleinberg, B. Scholkopf, F. Kelly, and I. Witten, Eds. Springer
-
R. Y. Rubinstein and D. P. Kroese, The Cross Entropy Method. A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation, and Machine Learning, ser. Information Science and Statistics, M. Jordan, J. Kleinberg, B. Scholkopf, F. Kelly, and I. Witten, Eds. Springer, 2004.
-
(2004)
A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation, and Machine Learning, ser. Information Science and Statistics
-
-
Rubinstein, R.Y.1
Kroese, D.P.2
-
14
-
-
4043069840
-
On actor-critic algorithms
-
V. R. Konda and J. N. Tsitsiklis, "On actor-critic algorithms," SIAM Journal on Control and Optimization, Vol. 42, no. 4, pp. 1143-1166, 2003.
-
(2003)
SIAM Journal on Control and Optimization
, vol.42
, Issue.4
, pp. 1143-1166
-
-
Konda, V.R.1
Tsitsiklis, J.N.2
-
15
-
-
34250354563
-
Convergence properties of the cross-entropy method for discrete optimization
-
DOI 10.1016/j.orl.2006.11.005, PII S0167637706001313
-
A. Costa, O. D. Jones, and D. Kroese, "Convergence properties of the cross-entropy method for discrete optimization," Operations Research Letters, Vol. 35, pp. 573-580, 2007. (Pubitemid 47198343)
-
(2007)
Operations Research Letters
, vol.35
, Issue.5
, pp. 573-580
-
-
Costa, A.1
Jones, O.D.2
Kroese, D.3
-
16
-
-
49949101369
-
Continuousstate reinforcement learning with fuzzy approximation
-
K. Tuyls, A. Nowé, Z. Guessoum, and D. Kudenko, Eds. Springer
-
L. Buşoniu, D. Ernst, B. De Schutter, and R. Babuška, "Continuousstate reinforcement learning with fuzzy approximation," in Adaptive Agents and Multi-Agent Systems III, ser. Lecture Notes in Computer Science, K. Tuyls, A. Nowé, Z. Guessoum, and D. Kudenko, Eds. Springer, 2008, Vol. 4865, pp. 27-43.
-
(2008)
Adaptive Agents and Multi-Agent Systems III, ser. Lecture Notes in Computer Science
, vol.4865
, pp. 27-43
-
-
Buşoniu, L.1
Ernst, D.2
De Schutter, B.3
Babuška, R.4
-
17
-
-
1642401055
-
Learning to drive a bicycle using reinforcement learning and shaping
-
Madison, US, 24-27 July
-
J. Randløv and P. Alstrøm, "Learning to drive a bicycle using reinforcement learning and shaping," in Proceedings 15th International Conference on Machine Learning (ICML-98), Madison, US, 24-27 July 1998, pp. 463-471.
-
(1998)
Proceedings 15th International Conference on Machine Learning (ICML-98)
, pp. 463-471
-
-
Randløv, J.1
Alstrøm, P.2
|