TY - GEN
T1 - Q-YOLOP
T2 - 2023 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2023
AU - Chang, Chi Chih
AU - Lin, Wei Cheng
AU - Wang, Pei Shuo
AU - Yu, Sheng Feng
AU - Lu, Yu Chen
AU - Lin, Kuan Cheng
AU - Wu, Kai Chiang
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - In this work, we present an efficient and quantization-aware panoptic driving perception model (Q-YOLOP) for object detection, drivable area segmentation, and lane line segmentation, in the context of autonomous driving. Our model employs the Efficient Layer Aggregation Network (ELAN) as its backbone and task-specific heads for each task. We employ a four-stage training process that includes pretraining on the BDD100K dataset, finetuning on both the BDD100K and iVS datasets, and quantization-aware training (QAT) on BDD100K. During the training process, we use powerful data augmentation techniques, such as random perspective and mosaic, and train the model on a combination of the BDD100K and iVS datasets. Both strategies enhance the model's generalization capabilities. The proposed model achieves state-of-the-art performance with an [email protected] of 0.622 for object detection and an mIoU of 0.612 for segmentation, while maintaining low computational and memory requirements.
AB - In this work, we present an efficient and quantization-aware panoptic driving perception model (Q-YOLOP) for object detection, drivable area segmentation, and lane line segmentation, in the context of autonomous driving. Our model employs the Efficient Layer Aggregation Network (ELAN) as its backbone and task-specific heads for each task. We employ a four-stage training process that includes pretraining on the BDD100K dataset, finetuning on both the BDD100K and iVS datasets, and quantization-aware training (QAT) on BDD100K. During the training process, we use powerful data augmentation techniques, such as random perspective and mosaic, and train the model on a combination of the BDD100K and iVS datasets. Both strategies enhance the model's generalization capabilities. The proposed model achieves state-of-the-art performance with an [email protected] of 0.622 for object detection and an mIoU of 0.612 for segmentation, while maintaining low computational and memory requirements.
KW - Object detection
KW - autonomous driving
KW - quantization-aware training
KW - semantic segmentation
UR - http://www.scopus.com/inward/record.url?scp=85172393910&partnerID=8YFLogxK
U2 - 10.1109/ICMEW59549.2023.00015
DO - 10.1109/ICMEW59549.2023.00015
M3 - Conference contribution
AN - SCOPUS:85172393910
T3 - Proceedings - 2023 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2023
SP - 52
EP - 56
BT - Proceedings - 2023 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2023
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 10 July 2023 through 14 July 2023
ER -