메뉴 건너뛰기




Volumn 12, Issue 4, 2011, Pages 1248-1260

Cooperative adaptive cruise control: A reinforcement learning approach

Author keywords

Autonomous vehicle control; cooperative adaptive cruise control (CACC); neural networks; policy gradient algorithms; reinforcement learning (RL)

Indexed keywords

ADAPTIVE CRUISE CONTROL; AUTONOMOUS VEHICLE CONTROL; AUTONOMOUS VEHICLES; COMPUTING TECHNOLOGY; CONSTANT SPEED; CONTROL POLICY; CONTROL TASK; COOPERATIVE ADAPTIVE CRUISE CONTROL; CURRENT SPEED; DESIGN APPROACHES; DRIVER-ASSISTANCE SYSTEMS; FOLLOWING VEHICLE; FUNCTION APPROXIMATION TECHNIQUES; GRADIENT-DESCENT; INTER VEHICLE COMMUNICATIONS; MACHINE LEARNING TECHNIQUES; REINFORCEMENT LEARNING APPROACH; SAFE DISTANCE; VEHICLE-TO-VEHICLE COMMUNICATION;

EID: 82455188002     PISSN: 15249050     EISSN: None     Source Type: Journal    
DOI: 10.1109/TITS.2011.2157145     Document Type: Article
Times cited : (308)

References (58)
  • 1
    • 50849142855 scopus 로고    scopus 로고
    • Advanced driver assistance systems from autonomous to cooperative approach
    • Sep.
    • J. Piao and M. McDonald, "Advanced driver assistance systems from autonomous to cooperative approach,"Transp. Rev., vol. 28, no. 5, pp. 659-684, Sep. 2008.
    • (2008) Transp. Rev. , vol.28 , Issue.5 , pp. 659-684
    • Piao, J.1    McDonald, M.2
  • 3
    • 44949241322 scopus 로고    scopus 로고
    • Reinforcement learning of motor skills with policy gradients
    • May
    • J. Peters and S. Schaal, "Reinforcement learning of motor skills with policy gradients,"Neural Netw., vol. 21, no. 4, pp. 682-697, May 2008.
    • (2008) Neural Netw. , vol.21 , Issue.4 , pp. 682-697
    • Peters, J.1    Schaal, S.2
  • 4
    • 0035461024 scopus 로고    scopus 로고
    • Human-centered design of an ACC with braking and forward-crash-warning system
    • P. Fancher, Z. Bareket, and R. Ervin, "Human-centered design of an ACC with braking and forward-crash-warning system,"Vehicle Syst. Dyn., vol. 36, no. 2, pp. 203-223, 2001.
    • (2001) Vehicle Syst. Dyn. , vol.36 , Issue.2 , pp. 203-223
    • Fancher, P.1    Bareket, Z.2    Ervin, R.3
  • 5
    • 82455193318 scopus 로고
    • Description of three PROMETHEUS demonstrators having potential safety effects
    • D. Augello, "Description of three PROMETHEUS demonstrators having potential safety effects,"in Proc. 13th Int. Tech. Conf. Exp. Safety Vehicles, 1993, pp. 4-7.
    • (1993) Proc. 13th Int. Tech. Conf. Exp. Safety Vehicles , pp. 4-7
    • Augello, D.1
  • 6
    • 84862409208 scopus 로고    scopus 로고
    • Fuel consumption reduction experienced by two promote chauffeur trucks in electronic tow bar operation
    • C. Bonnet and H. Fritz, "Fuel consumption reduction experienced by two promote chauffeur trucks in electronic tow bar operation,"in Proc. 7th World Congr. Intell. Transp. Syst., 2000.
    • (2000) Proc. 7th World Congr. Intell. Transp. Syst.
    • Bonnet, C.1    Fritz, H.2
  • 9
    • 74349087100 scopus 로고    scopus 로고
    • An introduction to Demo 2000: The cooperative driving scenario
    • Jul.
    • S. Tsugawa, "An introduction to Demo 2000: The cooperative driving scenario,"IEEE Intell. Syst., vol. 15, no. 4, pp. 78-79, Jul. 2000.
    • (2000) IEEE Intell. Syst. , vol.15 , Issue.4 , pp. 78-79
    • Tsugawa, S.1
  • 10
    • 39349087026 scopus 로고    scopus 로고
    • Enhancing road weather information through vehicle infrastructure integration
    • K. Petty and W. Mahoney, "Enhancing road weather information through vehicle infrastructure integration,"Transp. Res. Rec., no. 2015, pp. 132- 140, 2007.
    • (2007) Transp. Res. Rec. , Issue.2015 , pp. 132-140
    • Petty, K.1    Mahoney, W.2
  • 11
    • 4544259801 scopus 로고    scopus 로고
    • Radar and vision data fusion for hybrid adaptive cruise control on highways
    • Apr.
    • U. Hofmann, A. Rieder, and E. Dickmanns, "Radar and vision data fusion for hybrid adaptive cruise control on highways,"Mach. Vis. Appl., vol. 14, no. 1, pp. 42-49, Apr. 2003.
    • (2003) Mach. Vis. Appl. , vol.14 , Issue.1 , pp. 42-49
    • Hofmann, U.1    Rieder, A.2    Dickmanns, E.3
  • 16
    • 77956517783 scopus 로고    scopus 로고
    • A comprehensive review of the development of adaptive cruise control systems
    • L. Xiao and F. Gao, "A comprehensive review of the development of adaptive cruise control systems,"Vehicle Syst. Dyn., vol. 48, no. 10, pp. 1167-1192, 2010.
    • (2010) Vehicle Syst. Dyn. , vol.48 , Issue.10 , pp. 1167-1192
    • Xiao, L.1    Gao, F.2
  • 20
    • 33845537933 scopus 로고    scopus 로고
    • The impact of cooperative adaptive cruise control on traffic-flow characteristics
    • Dec.
    • B. Van Arem, C. Van Driel, and R. Visser, "The impact of cooperative adaptive cruise control on traffic-flow characteristics,"IEEE Trans. Intell. Transp. Syst., vol. 7, no. 4, pp. 429-436, Dec. 2006.
    • (2006) IEEE Trans. Intell. Transp. Syst. , vol.7 , Issue.4 , pp. 429-436
    • Van Arem, B.1    Van Driel, C.2    Visser, R.3
  • 21
    • 0008570374 scopus 로고
    • Neural network vision for robot driving
    • M. Arbib, Ed. Cambridge, MA: MIT Press
    • D. Pomerleau, "Neural network vision for robot driving,"in The Handbook of Brain Theory and Neural Networks, M. Arbib, Ed. Cambridge, MA: MIT Press, 1995.
    • (1995) The Handbook of Brain Theory and Neural Networks
    • Pomerleau, D.1
  • 22
    • 0029509215 scopus 로고
    • Road following with continuous learning
    • G. Yu and I. Sethi, "Road following with continuous learning,"in Proc. Intell. Vehicles Symp., 1995, pp. 412-417.
    • (1995) Proc. Intell. Vehicles Symp. , pp. 412-417
    • Yu, G.1    Sethi, I.2
  • 23
    • 70349116541 scopus 로고    scopus 로고
    • Reinforcement learning and adaptive dynamic programming for feedback control
    • Third Quarter
    • F. Lewis and D. Vrabie, "Reinforcement learning and adaptive dynamic programming for feedback control,"IEEE Circuits Syst. Mag., vol. 9, no. 3, pp. 32-50, Third Quarter, 2009.
    • (2009) IEEE Circuits Syst. Mag. , vol.9 , Issue.3 , pp. 32-50
    • Lewis, F.1    Vrabie, D.2
  • 24
    • 0033685792 scopus 로고    scopus 로고
    • A new reinforcement learning vehicle control architecture for vision-based road following
    • May
    • S. Oh, J. Lee, and D. Choi, "A new reinforcement learning vehicle control architecture for vision-based road following,"IEEE Trans. Veh. Technol., vol. 49, no. 3, pp. 997-1005, May 2000.
    • (2000) IEEE Trans. Veh. Technol. , vol.49 , Issue.3 , pp. 997-1005
    • Oh, S.1    Lee, J.2    Choi, D.3
  • 27
    • 82455195523 scopus 로고    scopus 로고
    • Reinforcement learning of dynamic collaborative driving-Part I: Longitudinal adaptive control
    • L. Ng, C. Clark, and J. Huissoon, "Reinforcement learning of dynamic collaborative driving-Part I: Longitudinal adaptive control,"Int. J. Vehicle Inf. Commun. Syst., vol. 1, no. 3, pp. 208-228, 2008.
    • (2008) Int. J. Vehicle Inf. Commun. Syst. , vol.1 , Issue.3 , pp. 208-228
    • Ng, L.1    Clark, C.2    Huissoon, J.3
  • 39
    • 0001251942 scopus 로고    scopus 로고
    • Reinforcement learning in POMDPS with function approximation
    • H. Kimura, M. Yamamura, and S. Kobayashi, "Reinforcement learning in POMDPS with function approximation,"in Proc. 14th ICML, 1997, pp. 152-160.
    • (1997) Proc. 14th ICML , pp. 152-160
    • Kimura, H.1    Yamamura, M.2    Kobayashi, S.3
  • 40
    • 0034859944 scopus 로고    scopus 로고
    • Autonomous helicopter control using reinforcement learning policy search methods
    • Seoul, Korea,May
    • J. A. Bagnell and J. G. Schneider, "Autonomous helicopter control using reinforcement learning policy search methods,"in Proc. IEEE ICRA, Seoul, Korea, May 2001, pp. 1615-1620.
    • (2001) Proc. IEEE ICRA , pp. 1615-1620
    • Bagnell, J.A.1    Schneider, J.G.2
  • 41
    • 82455200419 scopus 로고    scopus 로고
    • M.S. thesis, Victoria Univ. Wellington, Wellington, New Zealand
    • T. Field, "Policy-gradient learning for motor control,"M.S. thesis, Victoria Univ. Wellington, Wellington, New Zealand, 2005.
    • (2005) Policy-Gradient Learning for Motor Control
    • Field, T.1
  • 42
    • 3042534761 scopus 로고    scopus 로고
    • Policy-gradient reinforcement learning for fast quadrupedal locomotion
    • New Orleans, LA,May
    • N. Kohl and P. Stone, "Policy-gradient reinforcement learning for fast quadrupedal locomotion,"in Proc. IEEE ICRA, New Orleans, LA, May 2004, pp. 2619-2624.
    • (2004) Proc. IEEE ICRA , pp. 2619-2624
    • Kohl, N.1    Stone, P.2
  • 43
    • 67149110572 scopus 로고    scopus 로고
    • Direct gradient-based reinforcement learning for robot behavior learning
    • Sep.
    • A. El-Fakdi, M. Carreras, and P. Ridao, "Direct gradient-based reinforcement learning for robot behavior learning,"in Proc. 2nd ICINCO, Sep. 2005, pp. 225-231.
    • (2005) Proc. 2nd ICINCO , pp. 225-231
    • El-Fakdi, A.1    Carreras, M.2    Ridao, P.3
  • 44
    • 0013535965 scopus 로고    scopus 로고
    • Infinite-horizon policy-gradient estimation
    • Jul.
    • J. Baxter and P. Bartlett, "Infinite-horizon policy-gradient estimation,"J. Artif. Intell. Res., vol. 15, no. 4, pp. 319-350, Jul. 2001.
    • (2001) J. Artif. Intell. Res. , vol.15 , Issue.4 , pp. 319-350
    • Baxter, J.1    Bartlett, P.2
  • 45
    • 0013495368 scopus 로고    scopus 로고
    • Experiments with infinitehorizon policy-gradient estimation
    • Jul.
    • J. Baxter, P. Bartlett, and L. Weaver, "Experiments with infinitehorizon policy-gradient estimation,"J. Artif. Intell. Res., vol. 15, no. 1, pp. 351-381, Jul. 2001.
    • (2001) J. Artif. Intell. Res. , vol.15 , Issue.1 , pp. 351-381
    • Baxter, J.1    Bartlett, P.2    Weaver, L.3
  • 46
    • 34250635407 scopus 로고    scopus 로고
    • Policy-gradient methods for robotics
    • Los Alamitos, CA,Oct.
    • J. Peters and S. Schaal, "Policy-gradient methods for robotics,"in Proc. IEEE/RSJ Int. Conf. IROS, Los Alamitos, CA, Oct. 2006, pp. 2219-2225.
    • (2006) Proc. IEEE/RSJ Int. Conf. IROS , pp. 2219-2225
    • Peters, J.1    Schaal, S.2
  • 47
    • 0000337576 scopus 로고
    • Simple statistical gradient-following algorithms for connectionist reinforcement learning
    • May
    • R. J. Williams, "Simple statistical gradient-following algorithms for connectionist reinforcement learning,"Mach. Learn., vol. 8, no. 3/4, pp. 229-256, May 1992.
    • (1992) Mach. Learn. , vol.8 , Issue.3-4 , pp. 229-256
    • Williams, R.J.1
  • 49
    • 85152624316 scopus 로고
    • Reinforcement learning by stochastic hill climbing on discounted reward
    • H. Kimura,M. Yamamura, and S. Kobayashi, "Reinforcement learning by stochastic hill climbing on discounted reward,"in Proc. 12th ICML, 1995, pp. 295-303.
    • (1995) Proc. 12th ICML , pp. 295-303
    • Kimura, H.1    Yamamura, M.2    Kobayashi, S.3
  • 50
    • 0034439308 scopus 로고    scopus 로고
    • Stochastic optimization of controlled partially observable Markov decision processes
    • P. Bartlett and J. Baxter, "Stochastic optimization of controlled partially observable Markov decision processes,"in Proc. 39th IEEE Conf. Decision Control, 2000, vol. 1, pp. 124-129.
    • (2000) Proc. 39th IEEE Conf. Decision Control , vol.1 , pp. 124-129
    • Bartlett, P.1    Baxter, J.2
  • 55
    • 0141596576 scopus 로고    scopus 로고
    • Policy invariance under reward transformations: Theory and application to reward shaping
    • A. Y. Ng, D. Harada, and S. J. Russell, "Policy invariance under reward transformations: Theory and application to reward shaping,"in Proc. 16th ICML, 1999, pp. 278-287.
    • (1999) Proc. 16th ICML , pp. 278-287
    • Ng, A.Y.1    Harada, D.2    Russell, S.J.3
  • 56
    • 0024861871 scopus 로고
    • Approximation by superpositions of a sigmoidal function
    • Dec.
    • G. V. Cybenko, "Approximation by superpositions of a sigmoidal function,"Math. Control, Signals Syst., vol. 2, no. 4, pp. 303-314, Dec. 1989.
    • (1989) Math. Control, Signals Syst. , vol.2 , Issue.4 , pp. 303-314
    • Cybenko, G.V.1
  • 57
    • 77956340777 scopus 로고    scopus 로고
    • Situation assessment for automatic lane-change maneuvers
    • Sep.
    • R. Schubert, K. Schulze, and G.Wanielik, "Situation assessment for automatic lane-change maneuvers,"IEEE Trans. Intell. Transp. Syst., vol. 11, no. 3, pp. 607-616, Sep. 2010.
    • (2010) IEEE Trans. Intell. Transp. Syst. , vol.11 , Issue.3 , pp. 607-616
    • Schubert, R.1    Schulze, K.2    Wanielik, G.3
  • 58
    • 33746031424 scopus 로고    scopus 로고
    • Partial local friendQ multiagent learning: Application to team automobile coordination problem
    • Jun
    • J. Laumônier and B. Chaib-draa, "Partial local friendQ multiagent learning: Application to team automobile coordination problem,"in Proc. Can. AI, Lecture Notes in Artificial Intelligence, Jun. 2006, pp. 361-372.
    • (2006) Proc. Can. AI Lecture Notes in Artificial Intelligence , pp. 361-372
    • Laumônier, J.1    Chaib-draa, B.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.