메뉴 건너뛰기




Volumn 40, Issue , 2011, Pages 1-24

Non-deterministic policies in markovian decision processes

Author keywords

[No Author keywords available]

Indexed keywords

ACTION SELECTION; CONVENTIONAL METHODS; DECISION MAKING PROCESS; DECISION-MAKING PROBLEM; DISCRETE DOMAINS; HUMAN SUBJECTS; MARKOVIAN DECISION PROCESS; MARKOVIAN ENVIRONMENT; MARKOVIAN PROCESS; MEDICAL DOMAINS; NEAR-OPTIMAL SOLUTIONS; REAL-WORLD PROBLEM; RUNNING TIME; STOCHASTIC ENVIRONMENT; WEB NAVIGATION;

EID: 79956364385     PISSN: None     EISSN: 10769757     Source Type: Journal    
DOI: 10.1613/jair.3175     Document Type: Article
Times cited : (25)

References (22)
  • 6
    • 0034160101 scopus 로고    scopus 로고
    • Planning treatment of ischemic heart disease with partially observable Markov decision processes
    • DOI 10.1016/S0933-3657(99)00042-1, PII S0933365799000421
    • Hauskrecht, M., & Fraser, H. (2000). Planning treatment of ischemic heart disease with partially observable Markov decision processes. Artificial Intelligence in Medicine, 18 (3), 221-244. (Pubitemid 30089530)
    • (2000) Artificial Intelligence in Medicine , vol.18 , Issue.3 , pp. 221-244
    • Hauskrecht, M.1    Fraser, H.2
  • 8
    • 51249181779 scopus 로고
    • A new polynomial-time algorithm for linear programming
    • Karmarkar, N. (1984). A new polynomial-time algorithm for linear programming. Combi- natorica, 4 (4), 373-395.
    • (1984) Combi- natorica , vol.4 , Issue.4 , pp. 373-395
    • Karmarkar, N.1
  • 9
    • 0036832954 scopus 로고    scopus 로고
    • Near-optimal reinforcement learning in polynomial time
    • Kearns, M., & Singh, S. (2002). Near-optimal reinforcement learning in polynomial time. Machine Learning, 49.
    • (2002) Machine Learning , vol.49
    • Kearns, M.1    Singh, S.2
  • 10
    • 0034530018 scopus 로고    scopus 로고
    • Deciding when to intervene: A Markov decision process approach
    • DOI 10.1016/S1386-5056(00)00099-X, PII S138650560000099X
    • Magni, P., Quaglini, S., Marchetti, M., & Barosi, G. (2000). Deciding when to intervene: A Markov decision process approach. International Journal of Medical Informatics, 60 (3), 237-253. (Pubitemid 32007952)
    • (2000) International Journal of Medical Informatics , vol.60 , Issue.3 , pp. 237-253
    • Magni, P.1    Quaglini, S.2    Marchetti, M.3    Barosi, G.4
  • 12
    • 33847336943 scopus 로고    scopus 로고
    • Bias and variance approximation in value function estimates
    • DOI 10.1287/mnsc.1060.0614
    • Mannor, S., Simester, D., Sun, P., & Tsitsiklis, J. N. (2007). Bias and variance approxima- tion in value function estimates. Management Science, 53 (2), 308-322. (Pubitemid 46326182)
    • (2007) Management Science , vol.53 , Issue.2 , pp. 308-322
    • Mannor, S.1    Simester, D.2    Sun, P.3    Tsitsiklis, J.N.4
  • 13
    • 79956340213 scopus 로고    scopus 로고
    • Amazon mechanical turk. In
    • MTurk (2010). Amazon mechanical turk. In http://www.mturk.com/.
    • (2010)
    • Turk, M.1
  • 14
    • 19144362679 scopus 로고    scopus 로고
    • An experimental design for the development of adaptive treatment strategies
    • DOI 10.1002/sim.2022
    • Murphy, S. A. (2005). An experimental design for the development of adaptive treatment strategies. Statistics in Medicine, 24 (10), 1455-1481. (Pubitemid 40716347)
    • (2005) Statistics in Medicine , vol.24 , Issue.10 , pp. 1455-1481
    • Murphy, S.A.1
  • 15
    • 34047273906 scopus 로고    scopus 로고
    • Constructing evidence-based treatment strategies using methods from computer science
    • DOI 10.1016/j.drugalcdep.2007.01.005, PII S0376871607000270
    • Pineau, J., Bellemare, M. G., Rush, A. J., Ghizaru, A., & Murphy, S. A. (2007). Construct- ing evidence-based treatment strategies using methods from computer science. Drug and Alcohol Dependence, 88 (Supplement 2), S52 - S60. (Pubitemid 46546455)
    • (2007) Drug and Alcohol Dependence , vol.88 , Issue.SUPPL. 2
    • Pineau, J.1    Bellemare, M.G.2    Rush, A.J.3    Ghizaru, A.4    Murphy, S.A.5
  • 17
    • 84944041010 scopus 로고    scopus 로고
    • Variance-Penalized Reinforcement Learning for Risk-Averse Asset Allocation
    • Sato, M., & Kobayashi, S. (2000). Variance-penalized reinforcement learning for risk-averse asset allocation. In Proceedings of the Second International Conference on Intelligent Data Engineering and Automated Learning, Data Mining, Financial Engineering, and Intelligent Agents, pp. 244-249. Springer-Verlag. (Pubitemid 33211324)
    • (2000) LECTURE NOTES IN COMPUTER SCIENCE , Issue.1983 , pp. 244-249
    • Sato, M.1    Kobayashi, S.2
  • 19
    • 79956364284 scopus 로고    scopus 로고
    • Schools-Wikipedia 2008/9 wikipedia selection for schools. In
    • Schools-Wikipedia (2009). 2008/9 wikipedia selection for schools. In http://schools- wikipedia.org/.
    • (2009)
  • 21
    • 26844472827 scopus 로고    scopus 로고
    • Agent based decision support system using re- inforcement learning under emergency circumstances
    • Thapa, D., Jung, I., & Wang, G. (2005). Agent based decision support system using re- inforcement learning under emergency circumstances. Lecture Notes in Computer Science, 3610, 888.
    • (2005) Lecture Notes in Computer Science , vol.3610 , pp. 888
    • Thapa, D.1    Jung, I.2    Wang, G.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.