메뉴 건너뛰기




Volumn 2, Issue , 2012, Pages 1495-1502

Apprenticeship learning for model parameters of partially observable environments

Author keywords

[No Author keywords available]

Indexed keywords

ACTION SELECTION; APPRENTICESHIP LEARNING; DIALOGUE SYSTEMS; ENVIRONMENT MODELS; EXPLICIT MODELING; MODEL PARAMETERS; OPTIMAL ACTIONS; PARTIALLY OBSERVABLE ENVIRONMENTS;

EID: 84867126700     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (13)

References (24)
  • 2
    • 77955809093 scopus 로고    scopus 로고
    • Autonomous helicopter aerobatics through apprenticeship learning
    • Abbeel, P., Coates, A., and Ng, A. Y. Autonomous helicopter aerobatics through apprenticeship learning. International Journal of Robotics Research, 29 (13):1608-1639, 2010.
    • (2010) International Journal of Robotics Research , vol.29 , Issue.13 , pp. 1608-1639
    • Abbeel, P.1    Coates, A.2    Ng, A.Y.3
  • 7
    • 79955875655 scopus 로고    scopus 로고
    • Inverse reinforcement learning in partially observable environments
    • Choi, J. and Kim, K.-E. Inverse reinforcement learning in partially observable environments. Journal of Machine Learning Research, 12:691-730, 2011.
    • (2011) Journal of Machine Learning Research , vol.12 , pp. 691-730
    • Choi, J.1    Kim, K.-E.2
  • 10
    • 85162071686 scopus 로고    scopus 로고
    • What makes some POMDP problems easy to approximate?
    • Platt, J., Koller, D., Singer, Y., and Roweis, S. (eds.) MIT Press, Cambridge, MA
    • Hsu, D., Lee, W. S., and Rong, N. What makes some POMDP problems easy to approximate? In Platt, J., Koller, D., Singer, Y., and Roweis, S. (eds.), Advances in Neural Information Processing Systems 20, pp. 689-696. MIT Press, Cambridge, MA, 2008.
    • (2008) Advances in Neural Information Processing Systems 20 , pp. 689-696
    • Hsu, D.1    Lee, W.S.2    Rong, N.3
  • 12
    • 0032073263 scopus 로고    scopus 로고
    • Planning and acting in partially observable stochastic domains
    • Kaelbling, L. P., Littman, M. L., and Cassandra, A. R. Planning and acting in partially observable stochastic domains. Artificial Intelligence, 101:99-134, 1998.
    • (1998) Artificial Intelligence , vol.101 , pp. 99-134
    • Kaelbling, L.P.1    Littman, M.L.2    Cassandra, A.R.3
  • 14
    • 70349645087 scopus 로고    scopus 로고
    • SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces
    • Kurniawati, H., Hsu, D., and Lee, W. S. SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces. In Proc. Robotics: Science and Systems, 2008.
    • (2008) Proc. Robotics: Science and Systems
    • Kurniawati, H.1    Hsu, D.2    Lee, W.S.3
  • 17
    • 72449199041 scopus 로고    scopus 로고
    • Training parsers by inverse reinforcement learning
    • Neu, G. and Szepesvári, C. Training parsers by inverse reinforcement learning. Machine Learning, 77(2-3): 303-337, 2009.
    • (2009) Machine Learning , vol.77 , Issue.2-3 , pp. 303-337
    • Neu, G.1    Szepesvári, C.2
  • 18
    • 85011436515 scopus 로고    scopus 로고
    • Direct search algorithms for optimization calculations
    • Powell, M. J. D. Direct search algorithms for optimization calculations. Acta Numerica, 7:287-336, 1998.
    • (1998) Acta Numerica , vol.7 , pp. 287-336
    • Powell, M.J.D.1
  • 21
    • 0015658957 scopus 로고
    • The optimal control of partially observable Markov processes over a finite horizon
    • Smallwood, R. and Sondik, E. The optimal control of partially observable Markov processes over a finite horizon,. Operations Research, 21:1071-1088, 1973.
    • (1973) Operations Research , vol.21 , pp. 1071-1088
    • Smallwood, R.1    Sondik, E.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.