메뉴 건너뛰기




Volumn , Issue , 2008, Pages 1433-1438

Maximum Entropy Inverse Reinforcement Learning

Author keywords

[No Author keywords available]

Indexed keywords

ENTROPY; INVERSE PROBLEMS;

EID: 85148975703     PISSN: None     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (1062)

References (11)
  • 1
    • 14344251217 scopus 로고    scopus 로고
    • Apprenticeship learning via inverse reinforcement learning
    • Abbeel, P., and Ng, A. Y. 2004. Apprenticeship learning via inverse reinforcement learning. In Proc. ICML, 1–8.
    • (2004) Proc. ICML , pp. 1-8
    • Abbeel, P.1    Ng, A. Y.2
  • 2
    • 33746050365 scopus 로고    scopus 로고
    • Maximum entropy distribution estimation with generalized regularization
    • Dudík, M., and Schapire, R. E. 2006. Maximum entropy distribution estimation with generalized regularization. In Proc. COLT, 123–138.
    • (2006) Proc. COLT , pp. 123-138
    • Dudík, M.1    Schapire, R. E.2
  • 3
    • 11944266539 scopus 로고
    • Information theory and statistical mechanics
    • Jaynes, E. T. 1957. Information theory and statistical mechanics. Physical Review 106:620–630.
    • (1957) Physical Review , vol.106 , pp. 620-630
    • Jaynes, E. T.1
  • 4
    • 33750318706 scopus 로고    scopus 로고
    • Predestination: Inferring destinations from partial trajectories
    • Krumm, J., and Horvitz, E. 2006. Predestination: Inferring destinations from partial trajectories. In Proc. Ubicomp, 243–260.
    • (2006) Proc. Ubicomp , pp. 243-260
    • Krumm, J.1    Horvitz, E.2
  • 5
    • 0142192295 scopus 로고    scopus 로고
    • Conditional random fields: Probabilistic models for segmenting and labeling sequence data
    • Lafferty, J.; McCallum, A.; and Pereira, F. 2001. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proc. ICML, 282–289.
    • (2001) Proc. ICML , pp. 282-289
    • Lafferty, J.1    McCallum, A.2    Pereira, F.3
  • 6
    • 33750727337 scopus 로고    scopus 로고
    • Trip router with individualized preferences (trip): Incorporating personalization into route planning
    • Letchner, J.; Krumm, J.; and Horvitz, E. 2006. Trip router with individualized preferences (trip): Incorporating personalization into route planning. In Proc. IAAI, 1795–1800.
    • (2006) Proc. IAAI , pp. 1795-1800
    • Letchner, J.1    Krumm, J.2    Horvitz, E.3
  • 7
    • 34147182909 scopus 로고    scopus 로고
    • Learning and inferring transportation routines
    • Liao, L.; Patterson, D. J.; Fox, D.; and Kautz, H. 2007. Learning and inferring transportation routines. Artificial Intelligence 171(5-6):311–331.
    • (2007) Artificial Intelligence , vol.171 , Issue.5-6 , pp. 311-331
    • Liao, L.1    Patterson, D. J.2    Fox, D.3    Kautz, H.4
  • 8
    • 80053212134 scopus 로고    scopus 로고
    • Apprenticeship learning using inverse reinforcement learning and gradient methods
    • Neu, G., and Szepesvri, C. 2007. Apprenticeship learning using inverse reinforcement learning and gradient methods. In Proc. UAI, 295–302.
    • (2007) Proc. UAI , pp. 295-302
    • Neu, G.1    Szepesvri, C.2
  • 9
    • 0042547347 scopus 로고    scopus 로고
    • Algorithms for inverse reinforcement learning
    • Ng, A. Y., and Russell, S. 2000. Algorithms for inverse reinforcement learning. In Proc. ICML, 663–670.
    • (2000) Proc. ICML , pp. 663-670
    • Ng, A. Y.1    Russell, S.2
  • 10
    • 77956052826 scopus 로고    scopus 로고
    • Bayesian inverse reinforcement learning
    • Ramachandran, D., and Amir, E. 2007. Bayesian inverse reinforcement learning. In Proc. IJCAI, 2586–2591.
    • (2007) Proc. IJCAI , pp. 2586-2591
    • Ramachandran, D.1    Amir, E.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.