메뉴 건너뛰기




Volumn 27, Issue 2, 2009, Pages 123-130

Learning Model-free robot control by a Monte Carlo em algorithm

Author keywords

EM algorithm; Model free robot control; Probabilistic inference; Reinforcement learning

Indexed keywords

CONTROLLER PARAMETER; EM ALGORITHM; GEOMETRIC DISTRIBUTION; INFINITE HORIZONS; LEARNING MODELS; MODEL FREE; MODEL-FREE ROBOT CONTROL; MONTE CARLO EM ALGORITHM; MONTREAL , CANADA; NEURAL INFORMATION PROCESSING SYSTEMS; ON-MACHINES; PROBABILISTIC INFERENCE; PROBABILISTIC MODELS; REAL ROBOT; ROBOT CONTROLS; ROBOT TRAJECTORY;

EID: 70349327392     PISSN: 09295593     EISSN: None     Source Type: Journal    
DOI: 10.1007/s10514-009-9132-0     Document Type: Article
Times cited : (56)

References (21)
  • 4
    • 0346982426 scopus 로고    scopus 로고
    • Using expectation-maximization for reinforcement learning
    • 0876.68090 10.1162/neco.1997.9.2.271
    • P. Dayan G. E. Hinton 1997 Using expectation-maximization for reinforcement learning Neural Computation 9 2 271 278 0876.68090 10.1162/neco.1997.9.2.271
    • (1997) Neural Computation , vol.9 , Issue.2 , pp. 271-278
    • Dayan, P.1    Hinton, G.E.2
  • 7
    • 33645325949 scopus 로고    scopus 로고
    • Dynamic analysis of a nonholonomic two-wheeled inverted pendulum robot
    • 10.1007/s10846-005-9022-4
    • Y. Kim S. H. Kim Y. K. Kwak 2005 Dynamic analysis of a nonholonomic two-wheeled inverted pendulum robot Journal of Intelligent and Robotic Systems 44 1 25 46 10.1007/s10846-005-9022-4
    • (2005) Journal of Intelligent and Robotic Systems , vol.44 , Issue.1 , pp. 25-46
    • Kim, Y.1    Kim, S.H.2    Kwak, Y.K.3
  • 9
    • 70349325516 scopus 로고    scopus 로고
    • A Bayesian exploration-exploitation approach for optimal online sensing and planning with a visually guided mobile robot
    • doi: 10.1007/s10514-009-9130-2
    • Martinez-Cantin, R., de Freitas, N., Castellanos, J. A., & Doucet, A. (2009). A Bayesian exploration-exploitation approach for optimal online sensing and planning with a visually guided mobile robot. Autonomous Robots. doi: 10.1007/s10514-009-9130-2.
    • (2009) Autonomous Robots
    • Martinez-Cantin, R.1    De Freitas, N.2    Castellanos, J.A.3    Doucet, A.4
  • 10
    • 0002788893 scopus 로고    scopus 로고
    • A view of the em algorithm that justifies incremental, sparse, and other variants
    • M. I. Jordan (eds). Kluwer Academic Dordrecht
    • Neal, R. M., & Hinton, G. E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. In M. I. Jordan (Ed.), Learning in graphical models (pp. 355-368). Dordrecht: Kluwer Academic.
    • (1998) Learning in Graphical Models , pp. 355-368
    • Neal, R.M.1    Hinton, G.E.2
  • 12
    • 40649106649 scopus 로고    scopus 로고
    • Natural actor critic
    • 10.1016/j.neucom.2007.11.026
    • J. Peters S. Schaal 2008 Natural actor critic Neurocomputing 71 7-9 1180 1190 10.1016/j.neucom.2007.11.026
    • (2008) Neurocomputing , vol.71 , Issue.79 , pp. 1180-1190
    • Peters, J.1    Schaal, S.2
  • 13
    • 44949241322 scopus 로고    scopus 로고
    • Reinforcement learning of motor skills with policy gradients
    • 10.1016/j.neunet.2008.02.003
    • J. Peters S. Schaal 2008 Reinforcement learning of motor skills with policy gradients Neural Networks 21 4 682 697 10.1016/j.neunet.2008.02.003
    • (2008) Neural Networks , vol.21 , Issue.4 , pp. 682-697
    • Peters, J.1    Schaal, S.2
  • 15
    • 67650996818 scopus 로고    scopus 로고
    • Reinforcement learning for robot soccer
    • 10.1007/s10514-009-9120-4 This issue, part A
    • M. Riedmiller T. Gabel R. Hafner S. Lange 2009 Reinforcement learning for robot soccer Autonomous Robots 27 1 55 73 10.1007/s10514-009-9120-4 This issue, part A
    • (2009) Autonomous Robots , vol.27 , Issue.1 , pp. 55-73
    • Riedmiller, M.1    Gabel, T.2    Hafner, R.3    Lange, S.4
  • 16
    • 70349307164 scopus 로고    scopus 로고
    • State-dependent exploration for policy gradient methods
    • Rückstie, T., Felder, M., & Schmidhuber, J. (2008). State-dependent exploration for policy gradient methods. In Proc. European conf. on machine learning.
    • (2008) Proc. European Conf. on Machine Learning
    • Rückstie, T.1
  • 19
    • 34250728061 scopus 로고    scopus 로고
    • Probabilistic inference for solving discrete and continuous state Markov decision processes
    • Toussaint, M., & Storkey, A. (2006). Probabilistic inference for solving discrete and continuous state Markov decision processes. In Proc. int. conf. on machine learning.
    • (2006) Proc. Int. Conf. on Machine Learning
    • Toussaint, M.1    Storkey, A.2
  • 20
  • 21
    • 84950432017 scopus 로고
    • A Monte Carlo implementation of the em algorithm and the poor man's data augmentation algorithm
    • 10.2307/2290005
    • G. Wei M. Tanner 1990 A Monte Carlo implementation of the EM algorithm and the poor man's data augmentation algorithm Journal of the American Statistical Association 85 699 704 10.2307/2290005
    • (1990) Journal of the American Statistical Association , vol.85 , pp. 699-704
    • Wei, G.1    Tanner, M.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.