메뉴 건너뛰기




Volumn 16, Issue 3, 2005, Pages 611-617

Reinforcement learning for mobile robot: From reaction to deliberation

Author keywords

Deliberative control; Mobile robot; Reactive control; Reinforcement learning

Indexed keywords

DELIBERATIVE CONTROL; HIERARCHICAL METHODS; MODULAR METHODS; REACTIVE CONTROL; REINFORCEMENT LEARNING;

EID: 27744491246     PISSN: 16711793     EISSN: None     Source Type: Journal    
DOI: None     Document Type: Article
Times cited : (11)

References (21)
  • 5
    • 0029277469 scopus 로고
    • A sensor-based navigation for a mobile robot using fuzzy logic and reinforcement learning
    • Beom H R, Cho H S. A sensor-based navigation for a mobile robot using fuzzy logic and reinforcement learning. IEEE Trans. on Systems, Man and Cybernetics. 1995, 25(3): 464-477.
    • (1995) IEEE Trans. on Systems, Man and Cybernetics , vol.25 , Issue.3 , pp. 464-477
    • Beom, H.R.1    Cho, H.S.2
  • 7
    • 33847202724 scopus 로고
    • Learning to predict by the methods of temporal difference
    • Sutton R. Learning to predict by the methods of temporal difference. Machine Learning, 1988, 3(1): 9-44.
    • (1988) Machine Learning , vol.3 , Issue.1 , pp. 9-44
    • Sutton, R.1
  • 10
    • 0031231885 scopus 로고    scopus 로고
    • Experiments with reinforcement learning in problems with continuous state and action spaces
    • Santamaria J, Sutton R, Ram A. Experiments with reinforcement learning in problems with continuous state and action spaces. Behavior, 1997, 6(2).
    • (1997) Behavior , vol.6 , Issue.2
    • Santamaria, J.1    Sutton, R.2    Ram, A.3
  • 12
    • 0003979861 scopus 로고    scopus 로고
    • Incorporating prior knowledge and previously learned information into reinforcement learning agents (Technical report)
    • Carnegie Mellon University, Institute for Complex Engineered Systems
    • Dixon K R, Malak R J, Khosla P K. Incorporating prior knowledge and previously learned information into reinforcement learning agents (Technical report). Carnegie Mellon University, Institute for Complex Engineered Systems, January, 2000.
    • (2000)
    • Dixon, K.R.1    Malak, R.J.2    Khosla, P.K.3
  • 14
    • 0008898273 scopus 로고    scopus 로고
    • Action selection methods using reinforcement learning
    • Trinity Hall, Cambridge
    • Humphrys M. Action selection methods using reinforcement learning. Trinity Hall, Cambridge. June 1997.
    • (1997)
    • Humphrys, M.1
  • 17
    • 0033170372 scopus 로고    scopus 로고
    • Between mdps and semi-mdps: A framework for temporal abstraction in reinforcement learning
    • Sutton R, Precup D, Singh S. Between mdps and semi-mdps: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 1999, 112: 181-211.
    • (1999) Artificial Intelligence , vol.112 , pp. 181-211
    • Sutton, R.1    Precup, D.2    Singh, S.3
  • 19
    • 0002278788 scopus 로고    scopus 로고
    • Hierarchical reinforcement learning with the Max q value function decomposition
    • Dietterich T G. Hierarchical reinforcement learning with the Max q value function decomposition. Journal of Artificial Intelligence Research, 2000, 13: 227-303.
    • (2000) Journal of Artificial Intelligence Research , vol.13 , pp. 227-303
    • Dietterich, T.G.1
  • 20
    • 0013498457 scopus 로고    scopus 로고
    • Hierarchical learning and planning in partially observable Markov decision processes
    • Department of Computer Science and Engineering, Michigan State University, USA
    • Theocharous G. Hierarchical learning and planning in partially observable Markov decision processes. Department of Computer Science and Engineering, Michigan State University, USA. 2002.
    • (2002)
    • Theocharous, G.1
  • 21
    • 0031998630 scopus 로고    scopus 로고
    • Learning metric-topological maps for indoor mobile robot navigation
    • Thrun S. Learning metric-topological maps for indoor mobile robot navigation. Artificial Intelligence, 1998, 99(1): 21-71.
    • (1998) Artificial Intelligence , vol.99 , Issue.1 , pp. 21-71
    • Thrun, S.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.