메뉴 건너뛰기




Volumn 56, Issue 9, 2008, Pages 717-727

Reinforcement learning for problems with symmetrical restricted states

Author keywords

Elevator group control system; High dimensionality; Reinforcement learning; Symmetry

Indexed keywords

EDUCATION; OPTICAL FILTERS; REINFORCEMENT LEARNING; STANDARDS;

EID: 49649084490     PISSN: 09218890     EISSN: None     Source Type: Journal    
DOI: 10.1016/j.robot.2008.01.004     Document Type: Article
Times cited : (11)

References (24)
  • 1
    • 85153940465 scopus 로고
    • Generalization in reinforcement learning: Safely approximating the value function
    • MIT Press, Cambridge
    • Boyan J.A., and Moore A.W. Generalization in reinforcement learning: Safely approximating the value function. Advances in Neural Information Processing Systems vol. 7 (1995), MIT Press, Cambridge 369-376
    • (1995) Advances in Neural Information Processing Systems , vol.7 , pp. 369-376
    • Boyan, J.A.1    Moore, A.W.2
  • 2
    • 85156221438 scopus 로고    scopus 로고
    • Generalization in reinforcement learning: Successful examples using sparse coding
    • MIT Press, Cambridge
    • Sutton R.S. Generalization in reinforcement learning: Successful examples using sparse coding. Advances in Neural Information Processing Systems vol. 8 (1996), MIT Press, Cambridge 1038-1044
    • (1996) Advances in Neural Information Processing Systems , vol.8 , pp. 1038-1044
    • Sutton, R.S.1
  • 3
    • 49649111875 scopus 로고    scopus 로고
    • P. Stone, R.S. Sutton, Scalling reinforcement learning toward robocup soccer, in: Proc. of the 18th Int. Conf. on Machine learning, 2001
    • P. Stone, R.S. Sutton, Scalling reinforcement learning toward robocup soccer, in: Proc. of the 18th Int. Conf. on Machine learning, 2001
  • 4
    • 0037288370 scopus 로고    scopus 로고
    • Recent advances in hierarchical reinforcement learning
    • Barto A.G., and Mahadevan S. Recent advances in hierarchical reinforcement learning. Discrete Event Dynamic Systems 13 4 (2003) 41-77
    • (2003) Discrete Event Dynamic Systems , vol.13 , Issue.4 , pp. 41-77
    • Barto, A.G.1    Mahadevan, S.2
  • 6
    • 0034248853 scopus 로고    scopus 로고
    • Stochastic dynamic programming with factored representations
    • Boutilier C., Dearden R., and Goldszmidt M. Stochastic dynamic programming with factored representations. Artificial Intelligence 121 (2000) 49-107
    • (2000) Artificial Intelligence , vol.121 , pp. 49-107
    • Boutilier, C.1    Dearden, R.2    Goldszmidt, M.3
  • 7
    • 37249010686 scopus 로고    scopus 로고
    • B. Hengst, Generating hierarchical structure in reinforcement learning from state variables, in: PRICAI 2000 Topics in Artificial Intelligence, 2000, pp. 533-543
    • B. Hengst, Generating hierarchical structure in reinforcement learning from state variables, in: PRICAI 2000 Topics in Artificial Intelligence, 2000, pp. 533-543
  • 8
    • 0001041553 scopus 로고
    • Rapid task learning for real robot
    • Kluwer Academic Publishers
    • Connel J., and Mahadevan S. Rapid task learning for real robot. Robot Learning (1993), Kluwer Academic Publishers
    • (1993) Robot Learning
    • Connel, J.1    Mahadevan, S.2
  • 9
    • 49649104694 scopus 로고    scopus 로고
    • M.A.S. Kamal, J. Murata, K. Hirasawa, Task-oriented reinforcement learning for continuous tasks in dynamic environment, in: Proc. of the Conf. of Soceity of Instrument and Control Eng., 2002, pp. 932-935
    • M.A.S. Kamal, J. Murata, K. Hirasawa, Task-oriented reinforcement learning for continuous tasks in dynamic environment, in: Proc. of the Conf. of Soceity of Instrument and Control Eng., 2002, pp. 932-935
  • 10
    • 49649090817 scopus 로고    scopus 로고
    • M.A.S. Kamal, J. Murata, K. Hirasawa, Task-oriented multiagent reinforcement learning control for a real time high-dimensional problem, in: Proc. Int. Symposium on Artificial Life and Robotics, vol. 2, 2003, pp. 353-356
    • M.A.S. Kamal, J. Murata, K. Hirasawa, Task-oriented multiagent reinforcement learning control for a real time high-dimensional problem, in: Proc. Int. Symposium on Artificial Life and Robotics, vol. 2, 2003, pp. 353-356
  • 11
    • 49649105211 scopus 로고    scopus 로고
    • M. Zinkevich, T. Balch, Symmetry in Markov decision process and their implication for single agent and multi-agent learning, in: Proc. of the 18th Int. Conf. on Machine learning, 2001, pp. 632-640
    • M. Zinkevich, T. Balch, Symmetry in Markov decision process and their implication for single agent and multi-agent learning, in: Proc. of the 18th Int. Conf. on Machine learning, 2001, pp. 632-640
  • 12
    • 84880771557 scopus 로고    scopus 로고
    • B. Ravindran, A.G. Barto, SMDP homomorphisms: An algebraic approach to abstraction in semi Markov decision processes, in: Proc. of the 18th Int. Joint Conf. on Artificial Intel., IJCAI 2003, pp. 1011-1016
    • B. Ravindran, A.G. Barto, SMDP homomorphisms: An algebraic approach to abstraction in semi Markov decision processes, in: Proc. of the 18th Int. Joint Conf. on Artificial Intel., IJCAI 2003, pp. 1011-1016
  • 13
    • 84880677563 scopus 로고    scopus 로고
    • M. Kearns, D. Koller, Efficient reinforcement learning in factored MDPs, in: Proc. of the 16th Int. Joint Conf. on Artificial Intel., IJCAI, 1999, pp. 740-747
    • M. Kearns, D. Koller, Efficient reinforcement learning in factored MDPs, in: Proc. of the 16th Int. Joint Conf. on Artificial Intel., IJCAI, 1999, pp. 740-747
  • 14
    • 49649096798 scopus 로고    scopus 로고
    • Brian Sallans, Reinforcement learning for factored Markov decision process, Ph.D. Dissertation, Dept. of Computer Science, Torento University, 2002
    • Brian Sallans, Reinforcement learning for factored Markov decision process, Ph.D. Dissertation, Dept. of Computer Science, Torento University, 2002
  • 15
    • 49649110565 scopus 로고    scopus 로고
    • B. Ravindran, An algebraic approach to abstraction in reinforcement learning, Doctoral Dissertation, University of Massachusetts, 2004
    • B. Ravindran, An algebraic approach to abstraction in reinforcement learning, Doctoral Dissertation, University of Massachusetts, 2004
  • 16
    • 15744368771 scopus 로고    scopus 로고
    • M.A.S. Kamal, J. Murata, Reinforcement learning for high-dimensional problems with symmetrical actions, in: Proc. of Annual Conf. of IEEE Systems Man and Cybernetics, IEEE SMC, The Hague, Netherlands, October 2004, pp. 6192-6197
    • M.A.S. Kamal, J. Murata, Reinforcement learning for high-dimensional problems with symmetrical actions, in: Proc. of Annual Conf. of IEEE Systems Man and Cybernetics, IEEE SMC, The Hague, Netherlands, October 2004, pp. 6192-6197
  • 20
    • 0028497630 scopus 로고
    • Asynchronous stochastic approximation and Q-learning
    • Tsitsiklis J.N. Asynchronous stochastic approximation and Q-learning. Machine Learning 16 (1994) 185-202
    • (1994) Machine Learning , vol.16 , pp. 185-202
    • Tsitsiklis, J.N.1
  • 21
    • 49649127605 scopus 로고    scopus 로고
    • M. Tan, Multi-agent reinforcement learning: Independent vs. cooperative agents, in: Proc. Int. Conf. on Machine Learning, 1993, pp. 330-337
    • M. Tan, Multi-agent reinforcement learning: Independent vs. cooperative agents, in: Proc. Int. Conf. on Machine Learning, 1993, pp. 330-337
  • 22
    • 0032208335 scopus 로고    scopus 로고
    • Elevator group control using multiple reinforcement learning agents
    • Crites R.H., and Barto A.G. Elevator group control using multiple reinforcement learning agents. Machine Learning 33 2-3 (1998) 235-262
    • (1998) Machine Learning , vol.33 , Issue.2-3 , pp. 235-262
    • Crites, R.H.1    Barto, A.G.2
  • 24
    • 49649092578 scopus 로고    scopus 로고
    • Elevator group control using multiagent task-oriented reinforcement learning
    • Kamal M.A.S., Murata J., and Hirasawa K. Elevator group control using multiagent task-oriented reinforcement learning. Transactions of IEE Japan, Part C 125 7 (2005) 1140-1146
    • (2005) Transactions of IEE Japan, Part C , vol.125 , Issue.7 , pp. 1140-1146
    • Kamal, M.A.S.1    Murata, J.2    Hirasawa, K.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.