دورية أكاديمية

Faster CNN-based vehicle detection and counting strategy for fixed camera scenes

التفاصيل البيبلوغرافية
العنوان: Faster CNN-based vehicle detection and counting strategy for fixed camera scenes
المؤلفون: Gomaa, Ahmed, Minematsu, Tsubasa, Abdelwahab, Moataz M., Abo-Zahhad, Mohammed, Taniguchi, Rin-ichiro
المساهمون: The National Research Institute of Astronomy and Geophysics
المصدر: Multimedia Tools and Applications ; volume 81, issue 18, page 25443-25471 ; ISSN 1380-7501 1573-7721
بيانات النشر: Springer Science and Business Media LLC
سنة النشر: 2022
مصطلحات موضوعية: Computer Networks and Communications, Hardware and Architecture, Media Technology, Software
الوصف: Automatic detection and counting of vehicles in a video is a challenging task and has become a key application area of traffic monitoring and management. In this paper, an efficient real-time approach for the detection and counting of moving vehicles is presented based on YOLOv2 and features point motion analysis. The work is based on synchronous vehicle features detection and tracking to achieve accurate counting results. The proposed strategy works in two phases; the first one is vehicle detection and the second is the counting of moving vehicles. Different convolutional neural networks including pixel by pixel classification networks and regression networks are investigated to improve the detection and counting decisions. For initial object detection, we have utilized state-of-the-art faster deep learning object detection algorithm YOLOv2 before refining them using K-means clustering and KLT tracker. Then an efficient approach is introduced using temporal information of the detection and tracking feature points between the framesets to assign each vehicle label with their corresponding trajectories and truly counted it. Experimental results on twelve challenging videos have shown that the proposed scheme generally outperforms state-of-the-art strategies. Moreover, the proposed approach using YOLOv2 increases the average time performance for the twelve tested sequences by 93.4% and 98.9% from 1.24 frames per second achieved using Faster Region-based Convolutional Neural Network (F R-CNN ) and 0.19 frames per second achieved using the background subtraction based CNN approach (BS-CNN ), respectively to 18.7 frames per second.
نوع الوثيقة: article in journal/newspaper
اللغة: English
DOI: 10.1007/s11042-022-12370-9
DOI: 10.1007/s11042-022-12370-9.pdf
DOI: 10.1007/s11042-022-12370-9/fulltext.html
الإتاحة: https://doi.org/10.1007/s11042-022-12370-9Test
حقوق: https://creativecommons.org/licenses/by/4.0Test ; https://creativecommons.org/licenses/by/4.0Test
رقم الانضمام: edsbas.54F12A73
قاعدة البيانات: BASE