메뉴 건너뛰기




Volumn 55, Issue 2, 2010, Pages 492-497

A distributed actor-critic algorithm and applications to mobile sensor network coordination problems

Author keywords

Actor critic methods; Consensus; Markov decision processes (MDP); Multi agent coordination; Sensor networks

Indexed keywords

ACTOR-CRITIC ALGORITHM; ACTOR-CRITIC METHODS; COMMUNICATION CONSTRAINTS; COORDINATION PROBLEMS; GENERAL CLASS; KNOWN ENVIRONMENTS; MARKOV DECISION PROBLEM; MARKOV DECISION PROCESSES; MARKOV DECISION PROCESSES (MDP); MOBILE SENSOR NETWORKS; MULTI-AGENT COORDINATIONS; MULTIPLE AGENTS; SINGLE-AGENT;

EID: 76949102585     PISSN: 00189286     EISSN: None     Source Type: Journal    
DOI: 10.1109/TAC.2009.2037462     Document Type: Article
Times cited : (43)

References (10)
  • 1
  • 3
    • 0038548185 scopus 로고    scopus 로고
    • Coordination of groups of mobile autonomous agents using nearest neighbor rules
    • Jun.
    • A. Jadbabaie, J. Lin, and A. S. Morse, "Coordination of groups of mobile autonomous agents using nearest neighbor rules," IEEE Trans. Autom. Control, vol.48, no.6, pp. 988-1001, Jun. 2003.
    • (2003) IEEE Trans. Autom. Control , vol.48 , Issue.6 , pp. 988-1001
    • Jadbabaie, A.1    Lin, J.2    Morse, A.S.3
  • 5
    • 33847192512 scopus 로고    scopus 로고
    • Distributed cooperative coverage control of sensor networks
    • W. Li and C. G. Cassandras, "Distributed cooperative coverage control of sensor networks," in Proc. 44th IEEE Conf. Decision Control, 2005, pp. 2542-2547.
    • (2005) Proc. 44th IEEE Conf. Decision Control , pp. 2542-2547
    • Li, W.1    Cassandras, C.G.2
  • 6
  • 7
    • 62749157532 scopus 로고    scopus 로고
    • Solving sensor network coverage problems by distributed asynchronous actor-critic methods
    • New Orleans, LA, Dec.
    • P. Pennesi and I. C. Paschalidis, "Solving sensor network coverage problems by distributed asynchronous actor-critic methods," in Proc. 46th IEEE Conf. Decision Control, New Orleans, LA, Dec. 2007, pp. 5300-5305.
    • (2007) Proc. 46th IEEE Conf. Decision Control , pp. 5300-5305
    • Pennesi, P.1    Paschalidis, I.C.2
  • 8
    • 0035249254 scopus 로고    scopus 로고
    • Simulation-based optimization of Markov reward processes
    • Feb.
    • P. Marbach and J. Tsitsiklis, "Simulation-based optimization of Markov reward processes," IEEE Trans. Autom. Control, vol.46, no.2, pp. 191-209, Feb. 2001.
    • (2001) IEEE Trans. Autom. Control , vol.46 , Issue.2 , pp. 191-209
    • Marbach, P.1    Tsitsiklis, J.2
  • 9
    • 0031143730 scopus 로고    scopus 로고
    • An analysis of temporal-difference learning with function approximation
    • May
    • J. Tsitsiklis and B. V. Roy, "An analysis of temporal-difference learning with function approximation," IEEE Trans. Autom. Control, vol.42, no.5, pp. 674-690, May 1997.
    • (1997) IEEE Trans. Autom. Control , vol.42 , Issue.5 , pp. 674-690
    • Tsitsiklis, J.1    Roy, B.V.2
  • 10
    • 0022783899 scopus 로고
    • Distributed asynchronous deterministic and stochastic gradient optimization algorithms
    • Sep.
    • J. Tsitsiklis, D. Bertsekas, and M. Athans, "Distributed asynchronous deterministic and stochastic gradient optimization algorithms," IEEE Trans. Autom. Control, vol.AC-31, no.9, pp. 803-812, Sep. 1986.
    • (1986) IEEE Trans. Autom. Control , vol.AC-31 , Issue.9 , pp. 803-812
    • Tsitsiklis, J.1    Bertsekas, D.2    Athans, M.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.