FHPE-Net: Pedestrian Intention Prediction Using Fusion with Head Pose Estimation Based on RNN

Authors

  • Zhiyong Yang zyy@cqvie.edu.cn
  • Zihang Guo
  • Ruixiang Zhang
  • Jieru Guo
  • Yu Zhou

DOI:

https://doi.org/10.5755/j01.itc.53.3.34807

Keywords:

pedestrian action, autonomous vehicles, transport safety, fusion strategy

Abstract

Accurate real-time prediction of pedestrian crossing intent during the autonomous driving process is crucial for ensuring the safety of both pedestrians and passengers, as well as improving riding comfort. However, existing methods for pedestrian crossing intent detection mostly rely on extracting complete pose information of pedestrians, leading to reduced accuracy when pedestrians are occluded. To address this issue, this paper proposes FHPE-Net: a lightweight, multi-branch prediction model that utilizes only the head pose features of pedestrians. In pedestrian crossing scenarios, pedestrian behavior is highly influenced by surrounding vehicles and the environment. FHPE-Net encodes pedestrian head poses and global context semantic image sequences to comprehensively capture spatiotemporal interaction features between pedestrians, vehicles, and the environment, thereby enhancing the accuracy of pedestrian crossing intent prediction. To improve the robustness of the FHPE-Net method, this study further processes bounding box positions and vehicle velocity features, making it more stable and reliable in complex traffic scenarios. Finally, a novel U-BiGRUs module is introduced for feature fusion, and an optimal fusion strategy is employed to achieve the best predictive performance in terms of F1 score and accuracy (ACC). Extensive ablation experiments are conducted on the PIE dataset, and performance analysis demonstrates that FHPE-Net achieves an accuracy of 90%, outperforming baseline methods such as PCPA and Multi-RNN, while using only pedestrian head pose features. This research holds significant guidance in enhancing traffic safety and optimizing urban traffic management. Furthermore, it provides essential technological support for advancing the commercialization of autonomous driving.

Author Biographies

Zhiyong Yang, zyy@cqvie.edu.cn

Zhiyong Yang (zyy@cqvie.edu.cn) earned his Ph.D. degree in control engineering from the University of Chongqing. Currently, he is a professor at the College of Big Data and Internet of Things, Chongqing Vocational Institute of Engineering, and the College of Computer and Information Science, Chongqing Normal University, Chongqing, 402246, China. He is a senior member of the China Computer Federation, and he is the Vice Chairman of CCF YOCSEF. His current research interests include image recognition, artificial intelligence, and autonomous driving

Zihang Guo

Zihang Guo (guody123@foxmail.com) is currently pursuing a master’s degree at Chongqing Normal University, Chongqing, 401331, China. His research interests include autonomous driving and computer vision

Ruixiang Zhang

Ruixiang Zhang (z81077@outlook.com) is currently pursuing a master’s degree at Chongqing Normal University, Chong-qing, 401331, China. His research inter-ests include autonomous driving and computer vision.

Jieru Guo

Jieru Guo (cdguru@foxmail.com) is currently pursuing a master’s degree at Chongqing Normal University, Chong-qing, 401331, China. His research inter-ests include autonomous driving and im-age recognition.

Yu Zhou

Yu Zhou (zy1982@cqvie.edu.cn) earned her Master's degree in accounting from the University of Chongqing. Currently, she is an associate professor at the College of Finance and Tourism, Chongqing Voca-tional Institute of Engineering, Chong-qing, 402246, China. Her current research interests include big data analysis and data mining

Downloads

Published

2024-09-25

Issue

Section

Articles