-
1
-
-
84864030941
-
An application of reinforcement learning to aerobatic helicopter flight
-
Abbeel, P., Coates, A., Quigley, M., & Ng, A. Y. (2007). An application of reinforcement learning to aerobatic helicopter flight. In Proc. neural information processing systems.
-
(2007)
Proc. Neural Information Processing Systems
-
-
Abbeel, P.1
Coates, A.2
Quigley, M.3
Ng, A.Y.4
-
3
-
-
0008586604
-
A method for using belief networks as influence diagrams
-
Minneapolis, Minnesota
-
Cooper, G. F. (1988). A method for using belief networks as influence diagrams. In Proc. 4th workshop on uncertainty in artificial intelligence (pp. 55-63), Minneapolis, Minnesota.
-
(1988)
Proc. 4th Workshop on Uncertainty in Artificial Intelligence
, pp. 55-63
-
-
Cooper, G.F.1
-
4
-
-
0346982426
-
Using expectation-maximization for reinforcement learning
-
0876.68090 10.1162/neco.1997.9.2.271
-
P. Dayan G. E. Hinton 1997 Using expectation-maximization for reinforcement learning Neural Computation 9 2 271 278 0876.68090 10.1162/neco.1997.9.2.271
-
(1997)
Neural Computation
, vol.9
, Issue.2
, pp. 271-278
-
-
Dayan, P.1
Hinton, G.E.2
-
6
-
-
70350090880
-
Bayesian policy learning with trans-dimensional MCMC
-
Hoffman, M., Doucet, A., de Freitas, N., & Jasra, A. (2008). Bayesian policy learning with trans-dimensional MCMC. In Proc. neural information processing systems.
-
(2008)
Proc. Neural Information Processing Systems
-
-
Hoffman, M.1
Doucet, A.2
De Freitas, N.3
Jasra, A.4
-
7
-
-
33645325949
-
Dynamic analysis of a nonholonomic two-wheeled inverted pendulum robot
-
10.1007/s10846-005-9022-4
-
Y. Kim S. H. Kim Y. K. Kwak 2005 Dynamic analysis of a nonholonomic two-wheeled inverted pendulum robot Journal of Intelligent and Robotic Systems 44 1 25 46 10.1007/s10846-005-9022-4
-
(2005)
Journal of Intelligent and Robotic Systems
, vol.44
, Issue.1
, pp. 25-46
-
-
Kim, Y.1
Kim, S.H.2
Kwak, Y.K.3
-
9
-
-
70349325516
-
A Bayesian exploration-exploitation approach for optimal online sensing and planning with a visually guided mobile robot
-
doi: 10.1007/s10514-009-9130-2
-
Martinez-Cantin, R., de Freitas, N., Castellanos, J. A., & Doucet, A. (2009). A Bayesian exploration-exploitation approach for optimal online sensing and planning with a visually guided mobile robot. Autonomous Robots. doi: 10.1007/s10514-009-9130-2.
-
(2009)
Autonomous Robots
-
-
Martinez-Cantin, R.1
De Freitas, N.2
Castellanos, J.A.3
Doucet, A.4
-
10
-
-
0002788893
-
A view of the em algorithm that justifies incremental, sparse, and other variants
-
M. I. Jordan (eds). Kluwer Academic Dordrecht
-
Neal, R. M., & Hinton, G. E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. In M. I. Jordan (Ed.), Learning in graphical models (pp. 355-368). Dordrecht: Kluwer Academic.
-
(1998)
Learning in Graphical Models
, pp. 355-368
-
-
Neal, R.M.1
Hinton, G.E.2
-
12
-
-
40649106649
-
Natural actor critic
-
10.1016/j.neucom.2007.11.026
-
J. Peters S. Schaal 2008 Natural actor critic Neurocomputing 71 7-9 1180 1190 10.1016/j.neucom.2007.11.026
-
(2008)
Neurocomputing
, vol.71
, Issue.79
, pp. 1180-1190
-
-
Peters, J.1
Schaal, S.2
-
13
-
-
44949241322
-
Reinforcement learning of motor skills with policy gradients
-
10.1016/j.neunet.2008.02.003
-
J. Peters S. Schaal 2008 Reinforcement learning of motor skills with policy gradients Neural Networks 21 4 682 697 10.1016/j.neunet.2008.02.003
-
(2008)
Neural Networks
, vol.21
, Issue.4
, pp. 682-697
-
-
Peters, J.1
Schaal, S.2
-
15
-
-
67650996818
-
Reinforcement learning for robot soccer
-
10.1007/s10514-009-9120-4 This issue, part A
-
M. Riedmiller T. Gabel R. Hafner S. Lange 2009 Reinforcement learning for robot soccer Autonomous Robots 27 1 55 73 10.1007/s10514-009-9120-4 This issue, part A
-
(2009)
Autonomous Robots
, vol.27
, Issue.1
, pp. 55-73
-
-
Riedmiller, M.1
Gabel, T.2
Hafner, R.3
Lange, S.4
-
16
-
-
70349307164
-
State-dependent exploration for policy gradient methods
-
Rückstie, T., Felder, M., & Schmidhuber, J. (2008). State-dependent exploration for policy gradient methods. In Proc. European conf. on machine learning.
-
(2008)
Proc. European Conf. on Machine Learning
-
-
Rückstie, T.1
-
19
-
-
34250728061
-
Probabilistic inference for solving discrete and continuous state Markov decision processes
-
Toussaint, M., & Storkey, A. (2006). Probabilistic inference for solving discrete and continuous state Markov decision processes. In Proc. int. conf. on machine learning.
-
(2006)
Proc. Int. Conf. on Machine Learning
-
-
Toussaint, M.1
Storkey, A.2
-
21
-
-
84950432017
-
A Monte Carlo implementation of the em algorithm and the poor man's data augmentation algorithm
-
10.2307/2290005
-
G. Wei M. Tanner 1990 A Monte Carlo implementation of the EM algorithm and the poor man's data augmentation algorithm Journal of the American Statistical Association 85 699 704 10.2307/2290005
-
(1990)
Journal of the American Statistical Association
, vol.85
, pp. 699-704
-
-
Wei, G.1
Tanner, M.2
|