메뉴 건너뛰기




Volumn , Issue , 2007, Pages

Reinforcement learning in continuous action spaces through sequential Monte Carlo methods

Author keywords

[No Author keywords available]

Indexed keywords

APPROXIMATION ALGORITHMS; LEARNING ALGORITHMS; REINFORCEMENT LEARNING;

EID: 85161968592     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (88)

References (14)
  • 1
    • 0036475447 scopus 로고    scopus 로고
    • A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking
    • DOI 10.1109/78.978374, PII S1053587X0200569X
    • M. Sanjeev Arulampalam, Simon Maskell, Neil Gordon, and Tim Clapp. A tutorial on particle filters for online nonlinear/non-gaussian bayesian tracking. IEEE Trans. on Signal Processing, 50(2):174-188, 2002. (Pubitemid 34291500)
    • (2002) IEEE Transactions on Signal Processing , vol.50 , Issue.2 , pp. 174-188
    • Arulampalam, M.S.1    Maskell, S.2    Gordon, N.3    Clapp, T.4
  • 6
    • 0008550260 scopus 로고    scopus 로고
    • Reinforcement learning for continuous action using stochastic gradient ascent
    • H. Kimura and S. Kobayashi. Reinforcement learning for continuous action using stochastic gradient ascent. In 5th Intl. Conf. on Intelligent Autonomous Systems, pages 288-295, 1998.
    • (1998) 5th Intl. Conf. on Intelligent Autonomous Systems , pp. 288-295
    • Kimura, H.1    Kobayashi, S.2
  • 9
    • 0036832960 scopus 로고    scopus 로고
    • Continuous-action Q-learning
    • DOI 10.1023/A:1017988514716
    • Jose Del R. Millan, Daniele Posenato, and Eric Dedieu. Continuous-action q-learning. Machine Learning, 49:247-265, 2002. (Pubitemid 34325689)
    • (2002) Machine Learning , vol.49 , Issue.2-3 , pp. 247-265
    • Millan, J.D.R.1    Posenato, D.2    Dedieu, E.3
  • 10
    • 34250635407 scopus 로고    scopus 로고
    • Policy gradient methods for robotics
    • DOI 10.1109/IROS.2006.282564, 4058714, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2006
    • Jan Peters and Stefen Schaal. Policy gradient methods for robotics. In Proceedings of the IEEE International Conference on Intelligent Robotics Systems (IROS), pages 2219-2225, 2006. (Pubitemid 46928224)
    • (2006) IEEE International Conference on Intelligent Robots and Systems , pp. 2219-2225
    • Peters, J.1    Schaal, S.2
  • 11
    • 0031231885 scopus 로고    scopus 로고
    • Experiments with reinforcement learning in problems with continuous state and action spaces
    • J. C. Santamaria, R. S: Sutton, and A. Ram. Experiments with reinforcement learning in problems with continuous state and action spaces. Adaptive Behavior, 6:163-217, 1998. (Pubitemid 127175211)
    • (1997) Adaptive Behavior , vol.6 , Issue.2 , pp. 163-217
    • Santamaria, J.C.1    Ram, A.2    Sutton, R.S.3
  • 12
    • 26944466214 scopus 로고    scopus 로고
    • Function approximation via tile coding: Automating parameter choice
    • LNAI. Springer Verlag
    • Alexander A. Sherstov and Peter Stone. Function approximation via tile coding: Automating parameter choice. In SARA 2005, LNAI, pages 194-205. Springer Verlag, 2005.
    • (2005) SARA 2005 , pp. 194-205
    • Sherstov, A.A.1    Stone, P.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.