메뉴 건너뛰기




Volumn 11, Issue 6, 2011, Pages 4097-4109

Knowledge of opposite actions for reinforcement learning

Author keywords

NOQ( ) algorithm; Opposite action; Opposition weight; Opposition based learning (OBL); OQ( ) algorithm; Q( ); Reinforcement learning

Indexed keywords

ACTION SPACES; DISCRETE STATE; FASTER CONVERGENCE; MACHINE INTELLIGENCE; MODEL FREE; OPPOSITE ACTION; OPPOSITION WEIGHT; OPPOSITION-BASED LEARNING; OPTIMAL POLICIES; REAL-WORLD PROBLEM; STATE SPACE; VALUE FUNCTIONS;

EID: 79956159518     PISSN: 15684946     EISSN: None     Source Type: Journal    
DOI: 10.1016/j.asoc.2011.01.045     Document Type: Article
Times cited : (18)

References (39)
  • 1
    • 15744386698 scopus 로고    scopus 로고
    • Emotionally motivated reinforcement learning based controller
    • The Hague, The Netherlands
    • A. Ayesh Emotionally motivated reinforcement learning based controller IEEE SMC The Hague, The Netherlands 2004
    • (2004) IEEE SMC
    • Ayesh, A.1
  • 2
    • 0028731609 scopus 로고
    • Fuzzy Q-learning: A new approach for fuzzy dynamic programming, IEEE World Congress on Computational Intelligence
    • H.R. Berenji Fuzzy Q-learning: a new approach for fuzzy dynamic programming, IEEE World Congress on Computational Intelligence Fuzzy Systems 1 1994 486 491
    • (1994) Fuzzy Systems , vol.1 , pp. 486-491
    • Berenji, H.R.1
  • 4
    • 0010820738 scopus 로고    scopus 로고
    • HarperCollins Publishers 77-85 Fulham Palace Road, London, England
    • Collins Cobuild English Dictionary 2000 HarperCollins Publishers 77-85 Fulham Palace Road, London, England
    • (2000) Collins Cobuild English Dictionary
  • 5
    • 0003259931 scopus 로고    scopus 로고
    • Improving elevator performance using reinforcement learning
    • D.S. Touretzky, M.C. Mozer, M.E. Hasselmo (Eds.) MIT Press, Cambridge, MA
    • R.H. Crites, A.G. Barto, Improving elevator performance using reinforcement learning, in: D.S. Touretzky, M.C. Mozer, M.E. Hasselmo (Eds.), Advances in Neural Information Processing Systems, vol. 8, MIT Press, Cambridge, MA, 1996
    • (1996) Advances in Neural Information Processing Systems , vol.8
    • Crites, R.H.1    Barto, A.G.2
  • 6
    • 4444312102 scopus 로고    scopus 로고
    • Integrating guidance into relational reinforcement learning
    • K. Driessens, and S. Dzeroski Integrating guidance into relational reinforcement learning Machine Learning 57 2004 271 304
    • (2004) Machine Learning , vol.57 , pp. 271-304
    • Driessens, K.1    Dzeroski, S.2
  • 13
    • 0010498504 scopus 로고    scopus 로고
    • The NSF Workshop on Reinforcement Learning: Summary and Observations
    • S. Mahadevan, and L.P. Kaelbling The NSF Workshop on Reinforcement Learning: Summary and Observations AI Magazine 1996
    • (1996) AI Magazine
    • Mahadevan, S.1    Kaelbling, L.P.2
  • 17
    • 84988783053 scopus 로고    scopus 로고
    • Convergence of reinforcement learning algorithms and acceleration of learning
    • A. Potapov, and M.K. Ali Convergence of reinforcement learning algorithms and acceleration of learning Physical Review E 67 2003 026706
    • (2003) Physical Review e , vol.67 , pp. 026706
    • Potapov, A.1    Ali, M.K.2
  • 19
    • 34547240097 scopus 로고    scopus 로고
    • Opposition-based differential evolution for optimization of noisy problems
    • 1688534, 2006 IEEE Congress on Evolutionary Computation, CEC 2006
    • S. Rahnamayan, H.R. Tizhoosh, and M. Salama Opposition-based differential evolution for optimization of noisy problems IEEE Congress on Evolutionary Computation (CEC-2006) Vancouver 2006 1865 1872 (Pubitemid 47130723)
    • (2006) 2006 IEEE Congress on Evolutionary Computation, CEC 2006 , pp. 1865-1872
    • Rahnamayan, S.1    Tizhoosh, H.R.2    Salama, M.M.A.3
  • 20
    • 34249011503 scopus 로고    scopus 로고
    • A novel population initialization method for accelerating evolutionary algorithms
    • DOI 10.1016/j.camwa.2006.07.013, PII S0898122107001344
    • S. Rahnamayan, H.R. Tizhoosh, and M.M.A. Salama A novel population initialization method for accelerating evolutionary algorithms Elsevier Journal on Computers and Mathematics with Applications 53 10 2007 1605 1614 (Pubitemid 46783389)
    • (2007) Computers and Mathematics with Applications , vol.53 , Issue.10 , pp. 1605-1614
    • Rahnamayan, S.1    Tizhoosh, H.R.2    Salama, M.M.A.3
  • 24
    • 51649087431 scopus 로고    scopus 로고
    • Using background knowledge to speed reinforcement learning in physical agents
    • Montreal, Quebec, Canada
    • D. Shapiro, P. Langley, and R. Shachter Using background knowledge to speed reinforcement learning in physical agents AGENTS'01 Montreal, Quebec, Canada 2001
    • (2001) AGENTS'01
    • Shapiro, D.1    Langley, P.2    Shachter, R.3
  • 27
    • 56349096864 scopus 로고    scopus 로고
    • Tradeoff between exploration and exploitation of OQ(λ) with non-Markovian update in dynamic environments
    • M. Shokri, H.R. Tizhoosh, and M.S. Kamel Tradeoff between exploration and exploitation of OQ(λ) with non-Markovian update in dynamic environments IEEE International Joint Conference on Neural Networks 2008 2916 2922
    • (2008) IEEE International Joint Conference on Neural Networks , pp. 2916-2922
    • Shokri, M.1    Tizhoosh, H.R.2    Kamel, M.S.3
  • 28
    • 79956133635 scopus 로고    scopus 로고
    • The concept of opposition and its use in Q-learning and Q(λ) techniques
    • H.R. Tizhoosh, M. Ventresca (Eds.) Springer Physika-Verlag
    • M. Shokri, H.R. Tizhoosh, M.S. Kamel, The concept of opposition and its use in Q-learning and Q(λ) techniques, in: H.R. Tizhoosh, M. Ventresca (Eds.), Oppositional Concepts in Computational Intelligence, Springer Physika-Verlag, 2008
    • (2008) Oppositional Concepts in Computational Intelligence
    • Shokri, M.1    Tizhoosh, H.R.2    Kamel, M.S.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.