|
|
|
| Efficient Multi-task Visual Perception Model in Autonomous Driving Scenarios |
| LIU Bohang, ZHAO Qiang, TANG Zhenglin, TANG Yinglong, LI Yeqi
|
| College of Mechanical and Electrical Engineering, Northeast Forestry University, Harbin 150040, China |
|
|
|
|
Abstract To efficiently utilize the hardware computing power of autonomous vehicles, a multi-task perception model OLAD is constructed based on YOLOv5,which can simultaneously achieve traffic object detection, lane lines recognition, and drivable area segmentation. By introducing an improved SPPFCSPC module and redesigning the feature fusion network based on Slim Neck, OLAD enhances feature extraction capabilities, inference speed, and detection accuracy, the loss function is improved by incorporating MPDIoU to boost the accuracy of traffic objects detection. In terms of model performance validation, a comprehensive performance evaluation is conducted by supplementing the self-made domestic road dataset in the BDD100K validation set. The results show that the detection accuracy and speed of OLAD are better than the YOLOP of SOTA; In addition, public road images from different time periods in Suzhou are randomly selected to test the performance of the model on domestic roads. The results show that the perception results of the OLAD model in this paper are more accurate and suitable for domestic roads.
|
|
Received: 01 February 2024
Published: 13 February 2026
|
|
|
|
|
|
|
|
|