메뉴 건너뛰기




Volumn , Issue , 2007, Pages 662-669

Batch reinforcement learning in a complex domain

Author keywords

Evolution and adaptation; Learning; Perception and action

Indexed keywords

AGENT-BASED; ALGORITHMIC VARIANTS; ASYMPTOTIC PERFORMANCE; BATCH METHODS; BATCH REINFORCEMENT LEARNING; COMPLEX DOMAINS; DIMENSIONAL CONTROLS; EMPIRICAL PERFORMANCE; EVOLUTION AND ADAPTATION; EXPERIENCE DATUM; KEEPAWAY; LEARNING; MULTI AGENTS; ON-LINE ALGORITHMS; PERCEPTION AND ACTION; POLE BALANCING; ROBOCUP SOCCERS; SAMPLE COMPLEXITY; TEMPORAL DIFFERENCE REINFORCEMENT LEARNING;

EID: 60349130974     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: 10.1145/1329125.1329241     Document Type: Conference Paper
Times cited : (44)

References (16)
  • 2
    • 21844465127 scopus 로고    scopus 로고
    • Tree-based batch mode reinforcement learning
    • D. Ernst, P. Geurts, and L. Wehenkel. Tree-based batch mode reinforcement learning. J. Mach. Learn. Res., 6:503-556, 2005.
    • (2005) J. Mach. Learn. Res , vol.6 , pp. 503-556
    • Ernst, D.1    Geurts, P.2    Wehenkel, L.3
  • 5
    • 0000123778 scopus 로고
    • Self-improving reactive agents based on reinforcement learning, planning and teaching
    • L.-J. Lin. Self-improving reactive agents based on reinforcement learning, planning and teaching. Machine Learning, 8:293-321, 1992.
    • (1992) Machine Learning , vol.8 , pp. 293-321
    • Lin, L.-J.1
  • 7
    • 84898980684 scopus 로고    scopus 로고
    • Autonomous helicopter flight via reinforcement learning
    • S. Thrun, L. Saul, and B. Schölkopf, editors, MIT Press, Cambridge, MA
    • A. Y. Ng, H. J. Kim, M. I. Jordan, and S. Sastry. Autonomous helicopter flight via reinforcement learning. In S. Thrun, L. Saul, and B. Schölkopf, editors, Advances in Neural Information Processing Systems 16. MIT Press, Cambridge, MA, 2004.
    • (2004) Advances in Neural Information Processing Systems 16
    • Ng, A.Y.1    Kim, H.J.2    Jordan, M.I.3    Sastry, S.4
  • 8
    • 33646398129 scopus 로고    scopus 로고
    • Neural fitted q iteration - first experiences with a data efficient neural reinforcement learning method
    • J. Gama, R. Camacho, P. Brazdil, A. Jorge, and L. Torgo, editors, ECML, of, Springer
    • M. Riedmiller. Neural fitted q iteration - first experiences with a data efficient neural reinforcement learning method. In J. Gama, R. Camacho, P. Brazdil, A. Jorge, and L. Torgo, editors, ECML, volume 3720 of Lecture Notes in Computer Science, pages 317-328. Springer, 2005.
    • (2005) Lecture Notes in Computer Science , vol.3720 , pp. 317-328
    • Riedmiller, M.1
  • 9
    • 0003636089 scopus 로고
    • On-line Q-learning using connectionist systems
    • Cambridge University Engineering Department
    • G. A. Rummery and M. Niranjan. On-line Q-learning using connectionist systems. Technical Report CUED/F-INFENG/TR 166, Cambridge University Engineering Department, 1994.
    • (1994) Technical Report CUED/F-INFENG/TR , vol.166
    • Rummery, G.A.1    Niranjan, M.2
  • 11
    • 27544506565 scopus 로고    scopus 로고
    • Reinforcement learning for RoboCup-soccer keepaway
    • P. Stone, R. S. Sutton, and G. Kuhlmann. Reinforcement learning for RoboCup-soccer keepaway. Adaptive. Behavior, 13(3):165-188, 2005.
    • (2005) Adaptive. Behavior , vol.13 , Issue.3 , pp. 165-188
    • Stone, P.1    Sutton, R.S.2    Kuhlmann, G.3
  • 13
    • 0033170372 scopus 로고    scopus 로고
    • Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning
    • R. S. Sutton, D. Precup, and S. P. Singh. Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 112(1-2):181-211, 1999.
    • (1999) Artificial Intelligence , vol.112 , Issue.1-2 , pp. 181-211
    • Sutton, R.S.1    Precup, D.2    Singh, S.P.3
  • 15
    • 27544473171 scopus 로고    scopus 로고
    • Behavior transfer for value-function-based reinforcement learning
    • F. Dignum, V. Dignum, S. Koenig, S. Kraus, M. P. Singh, and M. Wooldridge, editors, New York, NY, July, ACM Press
    • M. E. Taylor and P. Stone. Behavior transfer for value-function-based reinforcement learning. In F. Dignum, V. Dignum, S. Koenig, S. Kraus, M. P. Singh, and M. Wooldridge, editors, The Fourth International Joint Conference on Autonomous Agents and Multiagent Systems, pages 53-59, New York, NY, July 2005. ACM Press.
    • (2005) The Fourth International Joint Conference on Autonomous Agents and Multiagent Systems , pp. 53-59
    • Taylor, M.E.1    Stone, P.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.