ESTIMATING THE TRAFFIC FLOW SPEED ON TRAFFIC VIDEOS USING YOLOV8 AND BYTETRACK | Phương | TNU Journal of Science and Technology

ESTIMATING THE TRAFFIC FLOW SPEED ON TRAFFIC VIDEOS USING YOLOV8 AND BYTETRACK

About this article

Received: 17/01/24                Revised: 28/03/24                Published: 29/03/24

Authors

1. Vu Le Quynh Phuong Email to author, Kien Giang Teachers Training College
2. Pham Nguyen Khang, Can Tho University
3. Tran Nguyen Minh Thu, Can Tho University

Abstract


Traffic flow is a critical aspect of economic, social, and environmental development. To assess traffic flow, estimating the speed of traffic is crucial. In this research, we propose a model for estimating traffic speed based on data collected from traffic surveillance cameras. The main objective is to count and track vehicles to estimate traffic flow by combining the Yolov8 and ByteTrack models, then calculating the average speed of vehicles. To train and evaluate the model's performance, data collected from the Vinh Thanh Van Police - Rach Gia City, including 10,092 images and over 96,024 labeled objects in various conditions, were used. The study experimented and compared the performance of our model with models combining Yolov8 and DeepSort. The results indicate that the proposed model has the lowest execution time and the capability to estimate traffic flow close to reality, with an accuracy of 91.39%. The dataset used in this research can be explored and utilized as a benchmark for similar problems.

Keywords


Traffic flow speed; YOLOv8; ByteTrack; Object recognition; Object tracking

References


[1] A. Fedorov, K. Nikolskaia, S. Ivanov, V. Shepelev, and A. Minbaleev, "Traffic flow estimation with data from a video surveillance camera," Journal of Big Data, vol. 6, pp. 1-15, 2019.

[2] M. M. Rahman and N. Nower, “Attention based Deep Hybrid Networks for Traffic Flow Prediction using Google Maps Data,” in Proceedings of the 2023 8th International Conference on Machine Learning Technologies, 2023, pp. 74-81.

[3] L. Q. P. Vu, N. T. Bui, K. H. Nguyen, N. M. T. Tran, and N. K. Pham, “Estimating the traffic density from traffic cameras,” in Future Data and Security Engineering. Big Data, Security and Privacy, Smart City and Industry 4.0 Applications, Springer, 2021, pp. 248-263.

[4] T. Abirami, C. Nivas, R. Naveen, and G. Nithishkumar, "Deep Learning based Traffic Analysis of Motor Cycles in Urban City," in 2022 6th International Conference on Intelligent Computing and Control Systems (ICICCS), 2022, pp. 1304-1309.

[5] F. Zhang, C. Li, and F. Yang, "Vehicle Detection in Urban Traffic Surveillance Images Based on Convolutional Neural Networks with Feature Concatenation," Sensors, vol. 19, pp. 549-570, 2019.

[6] T. Abirami, C. Nivas, R. Naveen, and T. G. Nithishkumar, “Deep Learning based Traffic Analysis of Motor Cycles in Urban City,” in 2022 6th International Conference on Intelligent Computing and Control Systems (ICICCS), IEEE, 2022, pp. 1304-1309.

[7] A. Vats and D. C. Anastasiu, “Enhancing Retail Checkout Through Video Inpainting, YOLOv8 Detection, and DeepSort Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5529-5536.

[8] Y. Zhang, P. Sun, Y. Jiang, D. Yu, F. Weng, Z. Yuan, and X. Wang, “Bytetrack: Multi-object tracking by associating every detection box,” in European Conference on Computer Vision, Cham: Springer Nature Switzerland, 2022, pp. 1-21.

[9] N. D. T. Yung, W. K. Wong, F. H. Juwono, and Z. A. Sim, “Safety helmet detection using deep learning: Implementation and comparative study using YOLOv5, YOLOv6, and YOLOv7,” in 2022 International Conference on Green Energy, Computing and Sustainable Technology (GECOST) IEEE, 2022, pp. 164-170.

[10] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You only look once: unified, real-time object detection," in 2016 IEEE conference on computer vision and pattern recognition (CVPR), 2016, pp. 779-788.

[11] A. Bochkovskiy, C. Y. Wang, and H. Y. M. Liao, “Yolov4: Optimal speed and accuracy of object detection,” arXiv preprint arXiv:2004.10934, 2020.

[12] C. Y. Wang, A. Bochkovskiy, and H. Y. M. Liao, “YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7464-7475.

[13] Y. Du, Z. Zhao, Y. Song, Y. Zhao, F. Su, T. Gong, and H. Meng, “Strongsort: Make deepsort great again,” IEEE Transactions on Multimedi, vol. 25, pp. 8725-8737, 2023.




DOI: https://doi.org/10.34238/tnu-jst.9604

Refbacks

  • There are currently no refbacks.
TNU Journal of Science and Technology
Rooms 408, 409 - Administration Building - Thai Nguyen University
Tan Thinh Ward - Thai Nguyen City
Phone: (+84) 208 3840 288 - E-mail: jst@tnu.edu.vn
Based on Open Journal Systems
©2018 All Rights Reserved