XVII International Conference on Systems, Automatic Control and Measurements, SAUM 2024 (pp. 92-95)
АУТОР(И) / AUTHOR(S): Jianxun Cui , Huidong Gao , Staniša Perić , Marko Milojković , Miroslav Milovanović
Download Full Pdf
DOI: 10.46793/SAUM24.092C
САЖЕТАК / ABSTRACT:
This paper reviews the core component of autonomous driving technology—the main methods of decision-making and planning. Special emphasis is placed on decision-making and planning as the key bridge connecting perception and control, and the development of behavior decision-making systems is thoroughly discussed. The article first introduces rule-based decision-making methods, including the three structures of finite state machines and their advantages and disadvantages. It then shifts to learning-based behavior decision-making methods, discussing in detail two strategies: imitation learning and reinforcement learning, including their advantages and limitations. Finally, the paper explores large model-based behavior decision-making methods, which leverage the general knowledge understanding and reasoning capabilities of large language models to provide decision support in the form of natural language.
КЉУЧНЕ РЕЧИ / KEYWORDS:
autonomous driving, reinforcement learning, language large model, behavior decision
ЛИТЕРАТУРА / REFERENCES:
- Kuefler, J. Morton, T. Wheeler, and M. Kochenderfer. Imitating driver behavior with generative adversarial networks. 2017 IEEE Intelligent Vehicles Symposium (IV), 2017: 204-211.
- González, J. Pérez, V. Milanés, and F. Nashashibi. A review of motion planning techniques for automated vehicles. IEEE Transactions on Intelligent Transportation Systems, 2015, 17(4): 1135- 1145.
- Leonard, J. How, S. Teller, S. Campbell, G. Fiore, and J. Williams. A perception-driven autonomous urban vehicle. Journal of Field Robotics, 2008, 25(10): 727-774.
- Montemerlo, J. Becker, S. Bhat, D. Dolgov, S. Ettinger, and S. Thrun. Junior: The stanford entry in the urban challenge. Journal of field Robotics, 2008, 25(9): 569-597.
- Xu, Y. Gao, F. Yu, and T. Darrell. End-to-end learning of driving models from large-scale video datasets. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017: 2174-2182.
- Pan, C. A. Cheng, K. Saigol, K. Lee, X. Yan, E. A. Theodorou, and B. Boots. Imitation learning for agile autonomous driving. The International Journal of Robotics Research, 2020, 39(2-3): 286- 302.
- Ross, and D. Bagnell. Efficient reductions for imitation learning. Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings, 2010: 661-668.
- Levine, and V. Koltun. Continuous inverse optimal control with locally optimal examples. arXiv preprint arXiv:1206.4617, 2012.
- D. Ziebart, A. L. Maas, J. A. Bagnell, and A. K. Dey. Maximum entropy inverse reinforcement learning. AAAI, 2008, 8: 1433-1438.
- Levine, Z. Popovic, and V. Koltun. Nonlinear inverse reinforcement learning with gaussian processes. Advances in neural information processing systems, 2011, 24: 19-27.
- Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, and Y. Bengio. Generative adversarial nets. Advances in neural information processing systems, 2014, 27.
- Ho, and S. Ermon. Generative adversarial imitation learning. Advances in neural information processing systems, 2016, 29: 4565-4573.
- Bhattacharyya, B. Wulfe, D. Phillips, A. Kuefler, J. Morton, R. Senanayake, and M. Kochenderfer. Modeling human driving behavior through generative adversarial imitation learning. arXiv preprint arXiv:2006.06412, 2020.
- J. Shin, and J. Kim. Randomized adversarial imitation learning for autonomous driving. arXiv preprint arXiv:1905.05637, 2019.
- Li, J. Song, and S. Ermon. Infogail: Interpretable imitation learning from visual demonstrations. Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017: 3815-3825.
- S. Sutton, and A. G. Barto. Reinforcement learning: An introduction. MIT press, 2018.
- Mirchevska, M. Blum, L. Louis, J. Boedecker, and M. Werling. Reinforcement learning for autonomous maneuvering in highway scenarios. Workshop for Driving Assistance Systems and Autonomous Driving, 2017: 32-41.
- Mirchevska, C. Pek, M. Werling, M. Althoff, and J. Boedecker. High-level decision making for safe and reasonable autonomous lane changing using reinforcement learning. 2018 21st International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2018: 2156-2162.
- Mo, X. Pei, and Z. Chen. Decision-Making for Oncoming Traffic Overtaking Scenario using Double DQN. 2019 3rd Conference on Vehicle Control and Intelligence (CVCI). IEEE, 2019: 1-4.
- Zhang, R. Zhang, T. Wu, R. Weng, M. Han, and Y. Zhao. Safe reinforcement learning with stability guarantee for motion planning of autonomous vehicles. IEEE Transactions on Neural Networks and Learning Systems, 2021, 32(12): 5435-5444.
- Zhu, X. Wang, and Y. Wang. Human-like autonomous car-following model with deep reinforcement learning. Transportation research part C: emerging technologies, 2018, 97: 348-368.
- Zhu, Y. Wang, Z. Pu, J. Hu, X. Wang, and R. Ke. Safe, efficient, and comfortable velocity control based on reinforcement learning for autonomous driving. Transportation Research Part C: Emerging Technologies, 2020, 117: 102662.
- Sha, Y. Mu, Y. Jiang, L. Chen, C. Xu, P. Luo, S. E. Li, M. Tomizuka, W. Zhan, and M. Ding, “Languagempc: Large language models as decision makers for autonomous driving,” arXiv preprint arXiv:2310.03026, 2023.
- Cui, Y. Ma, X. Cao, W. Ye, and Z. Wang, “Drive as you speak: Enabling human-like interaction with large language models in autonomous vehicles,” arXiv preprint arXiv:2309.10228, 2023.
- Chen, O. Sinavski, J. Hunermann, A. Karnsund, A. J. Willmott, ¨ D. Birch, D. Maund, and J. Shotton, “Driving with llms: Fusing object-level vector modality for explainable autonomous driving,” arXiv preprint arXiv:2310.01957, 2023.
- Wen, D. Fu, X. Li, X. Cai, T. Ma, P. Cai, M. Dou, B. Shi, L. He, and Y. Qiao, “Dilu: A knowledge-driven approach to autonomous driving with large language models,” arXiv preprint arXiv:2309.16292, 2023.
- Fu, X. Li, L. Wen, M. Dou, P. Cai, B. Shi, and Y. Qiao, “Drive like a human: Rethinking autonomous driving with large language models,” arXiv preprint arXiv:2307.07162, 2023.
- Gao, J. Han, R. Zhang, Z. Lin, S. Geng, A. Zhou, W. Zhang, P. Lu, C. He, X. Yue, H. Li, and Y. Qiao, “Llama-adapter v2: Parameterefficient visual instruction model,” 2023.