메뉴 건너뛰기




Volumn 7416 LNCS, Issue , 2012, Pages 1-12

WrightEagle and UT Austin Villa: RoboCup 2011 simulation league champions

Author keywords

[No Author keywords available]

Indexed keywords

3D SIMULATIONS; INTERNATIONAL COMPETITIONS; ROBOCUP; ROBOCUP SIMULATION LEAGUE; SCIENCE AND TECHNOLOGY; SIMULATION LEAGUE; UNIVERSITY OF TEXAS;

EID: 84865710452     PISSN: 03029743     EISSN: 16113349     Source Type: Book Series    
DOI: 10.1007/978-3-642-32060-6_1     Document Type: Article
Times cited : (10)

References (14)
  • 2
    • 84865728553 scopus 로고    scopus 로고
    • Hierarchical reinforcement learning with the MAXQ value function decomposition
    • Dietterich, T.G.: Hierarchical reinforcement learning with the MAXQ value function decomposition. Journal of Machine Learning Research 13(1), 63 (1999)
    • (1999) Journal of Machine Learning Research , vol.13 , Issue.1 , pp. 63
    • Dietterich, T.G.1
  • 7
    • 0032073263 scopus 로고    scopus 로고
    • Planning and acting in partially observable stochastic domains
    • PII S000437029800023X
    • Kaelbling, L.P., Littman, M.L., Cassandra, A.R.: Planning and acting in partially observable stochastic domains. Artificial Intelligence 101(1-2), 99-134 (1998) (Pubitemid 128387390)
    • (1998) Artificial Intelligence , vol.101 , Issue.1-2 , pp. 99-134
    • Kaelbling, L.P.1    Littman, M.L.2    Cassandra, A.R.3
  • 10
    • 84865777939 scopus 로고    scopus 로고
    • Action-driven markov decision process and the application in RoboCup
    • Shi, K., Chen, X.: Action-driven markov decision process and the application in RoboCup. Journal of Chinese Computer Systems 32, 511-515 (2011)
    • (2011) Journal of Chinese Computer Systems , vol.32 , pp. 511-515
    • Shi, K.1    Chen, X.2
  • 11
    • 0033170372 scopus 로고    scopus 로고
    • Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning
    • DOI 10.1016/S0004-3702(99)00052-1
    • Sutton, R.S., Precup, D., Singh, S.: Between mdps and semi-mdps: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence 112(1-2), 181-211 (1999) (Pubitemid 32079890)
    • (1999) Artificial Intelligence , vol.112 , Issue.1 , pp. 181-211
    • Sutton, R.S.1    Precup, D.2    Singh, S.3
  • 13
    • 50249116786 scopus 로고    scopus 로고
    • Solving large-scale and sparse-reward dec-pomdps with correlation-MDPs
    • Visser, U., Ribeiro, F., Ohashi, T., Dellaert, F. (eds.) 9. Springer, Heidelberg
    • Wu, F., Chen, X.: Solving Large-Scale and Sparse-Reward DEC-POMDPs with Correlation-MDPs. In: Visser, U., Ribeiro, F., Ohashi, T., Dellaert, F. (eds.) RoboCup 2007. LNCS (LNAI), vol. 5001, pp. 208-219. Springer, Heidelberg (2008)
    • (2008) RoboCup 2007. LNCS (LNAI) , vol.5001 , pp. 208-21
    • Wu, F.1    Chen, X.2
  • 14
    • 78650942474 scopus 로고    scopus 로고
    • Online planning for multi-agent systems with bounded communication
    • Wu, F., Zilberstein, S., Chen, X.: Online planning for multi-agent systems with bounded communication. Artificial Intelligence 175(2), 487-511 (2011)
    • (2011) Artificial Intelligence , vol.175 , Issue.2 , pp. 487-511
    • Wu, F.1    Zilberstein, S.2    Chen, X.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.