메뉴 건너뛰기




Volumn 7700 LECTURE NO, Issue , 2012, Pages 735-757

10 Steps and some tricks to set up neural reinforcement controllers

Author keywords

batch reinforcement learning; fitted Q; learning control; Neural reinforcement learning

Indexed keywords

BENCHMARKING; CONTROLLERS; LEARNING SYSTEMS;

EID: 84872531075     PISSN: 03029743     EISSN: 16113349     Source Type: Book Series    
DOI: 10.1007/978-3-642-35289-8_39     Document Type: Article
Times cited : (27)

References (32)
  • 4
  • 6
    • 80052242429 scopus 로고    scopus 로고
    • Improved neural fitted q iteration applied to a novel computer gaming and learning benchmark. in: Proceedings of the
    • ADPRL 2011), Paris, France. IEEE Press (April
    • Gabel, T., Lutz, C., Riedmiller, M.: Improved Neural Fitted Q Iteration Applied to a Novel Computer Gaming and Learning Benchmark. In: Proceedings of the IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning (ADPRL 2011), Paris, France. IEEE Press (April 2011)
    • (2011) IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning
    • Gabel, T.1    Lutz, C.2    Riedmiller, M.3
  • 11
    • 79958779459 scopus 로고    scopus 로고
    • Reinforcement learning in feedback control
    • 10.1007/s10994-011-5235-x
    • Hafner, R., Riedmiller, M.: Reinforcement learning in feedback control. Machine Learning 27(1), 55-74 (2011), 10.1007/s10994-011-5235-x
    • (2011) Machine Learning , vol.27 , Issue.1 , pp. 55-74
    • Hafner, R.1    Riedmiller, M.2
  • 12
    • 79956136559 scopus 로고    scopus 로고
    • Safe exploration for reinforcement learning
    • Hans, A., Schneegass, D., Schafer, A.M., Udluft, S.: Safe exploration for reinforcement learning. In: ESANN, pp. 143-148 (2008)
    • (2008) ESANN , pp. 143-148
    • Hans, A.1    Schneegass, D.2    Schafer, A.M.3    Udluft, S.4
  • 14
    • 0001857994 scopus 로고    scopus 로고
    • Orr, G.B., Muller, K.-R. (eds.) NIPS-WS 1996. LNCS Springer, Heidelberg
    • LeCun, Y., Bottou, L., Orr, G.B., Muller, K.-R.: Efficient backProp. In: Orr, G.B., Muller, K.-R. (eds.) NIPS-WS 1996. LNCS, vol. 1524, pp. 9-50. Springer, Heidelberg (1998)
    • (1998) Efficient backProp , vol.1524 , pp. 9-50
    • LeCun, Y.1    Bottou, L.2    Orr, G.B.3    Muller, K.-R.4
  • 15
  • 17
    • 84943274699 scopus 로고
    • A direct adaptive method for faster backpropagation learning: The RPROP algorithm
    • Ruspini, H. (ed.) (ICNN), San Francisco
    • Riedmiller, M., Braun, H.: A direct adaptive method for faster backpropagation learning: The RPROP algorithm. In: Ruspini, H. (ed.) Proceedings of the IEEE International Conference on Neural Networks (ICNN), San Francisco, pp. 586-591 (1993)
    • (1993) Proceedings of the IEEE International Conference on Neural Networks , pp. 586-591
    • Riedmiller, M.1    Braun, H.2
  • 18
    • 84858054371 scopus 로고    scopus 로고
    • Distributed policy search reinforcement learning for job-shop scheduling tasks
    • Available online from (May 2011
    • Riedmiller, M., Gabel, T.: Distributed Policy Search Reinforcement Learning for Job-Shop Scheduling Tasks. TPRS International Journal of Production Research 50(1) (2012); Available online from (May 2011)
    • (2012) TPRS International Journal of Production Research , vol.50 , Issue.1
    • Riedmiller, M.1    Gabel, T.2
  • 22
    • 0009267623 scopus 로고    scopus 로고
    • Generating continuous control signals for reinforcement controllers using dynamic output elements
    • ESANN 1997, Bruges
    • Riedmiller, M.: Generating continuous control signals for reinforcement controllers using dynamic output elements. In: European Symposium on Artificial Neural Networks, ESANN 1997, Bruges (1997)
    • (1997) European Symposium on Artificial Neural Networks
    • Riedmiller, M.1
  • 23
    • 33646398129 scopus 로고    scopus 로고
    • Neural fitted Q iteration - First experiences with a data efficient neural reinforcement learning method
    • Gama, J., Camacho, R., Brazdil, P.B., Jorge, A.M., Torgo, L. (eds.) Springer Heidelberg
    • Riedmiller, M.: Neural Fitted Q Iteration - First Experiences with a Data Efficient Neural Reinforcement Learning Method. In: Gama, J., Camacho, R., Brazdil, P.B., Jorge, A.M., Torgo, L. (eds.) ECML 2005. LNCS (LNAI), vol. 3720, pp. 317-328. Springer, Heidelberg (2005)
    • (2005) ECML 2005. LNCS (LNAI) , vol.3720 , pp. 317-328
    • Riedmiller, M.1
  • 24
    • 27944453854 scopus 로고    scopus 로고
    • Neural reinforcement learning to swing-up and balance a real pole
    • 2005, Big Island, USA October
    • Riedmiller, M.: Neural reinforcement learning to swing-up and balance a real pole. In: Proc. of the Int. Conference on Systems, Man and Cybernetics, 2005, Big Island, USA (October 2005)
    • (2005) Proc. of the Int. Conference on Systems Man and Cybernetics
    • Riedmiller, M.1
  • 32
    • 38049150972 scopus 로고    scopus 로고
    • Kok, J.N., Koronacki, J., Lopez de Mantaras, R., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI) Springer, Heidelberg
    • Walsh, T.J., Nouri, A., Li, L., Littman, M.L.: Planning and Learning in Environments with Delayed Feedback. In: Kok, J.N., Koronacki, J., Lopez de Mantaras, R., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI), vol. 4701, pp. 442-453. Springer, Heidelberg (2007)
    • (2007) Planning and Learning in Environments with Delayed Feedback , vol.4701 , pp. 442-453
    • Walsh, T.J.1    Nouri, A.2    Li, L.3    Littman, M.L.4


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.