Dept. of CSE, East West University (EWU), Dhaka, Bangladesh.
Dept. of ECE, East West University (EWU), Dhaka, Bangladesh.
PLoS One. 2024 Sep 9;19(9):e0308460. doi: 10.1371/journal.pone.0308460. eCollection 2024.
The Corona Virus Disease (COVID-19) has a huge impact on all of humanity, and people's disregard for COVID-19 regulations has sped up the disease's spread. Our study uses a state-of-the-art object detection model like YOLOv4 (You Only Look Once, version 4), a very effective tool, on real-time 25fps, 1920 X 1080 video data streamed live by a camera-mounted Unmanned Aerial Vehicle (UAV) quad-copter to observe proper maintenance of social distance in an area of 35m range in this study. The model has demonstrated remarkable efficacy in identifying and quantifying instances of social distancing, with an accuracy of 82% and little latency. It has been able to work efficiently with real-time streaming at 25-30 ms. Our model is based on CSPDarkNet-53, which was trained on the MS COCO dataset for image classification. It includes additional layers to capture feature maps from different phases. Additionally, the model's neck is made up of PANet, which is used to aggregate the parameters from various CSPDarkNet-53 layers. The CSPDarkNet-53's 53 convolutional layers are followed by 53 more layers in the model head, for a total of 106 completely convolutional layers in the design. This architecture is further integrated with YOLOv3, resulting in the YOLOv4 model, which will be used by our detection model. Furthermore, to differentiate humans The aforementioned method was used to evaluate drone footage and count social distance violations in real time. Our findings show that our model was reliable and successful at detecting social distance violations in real-time with an average accuracy of 82%.
新冠病毒病(COVID-19)对全人类造成了巨大影响,人们对 COVID-19 规定的无视加速了疾病的传播。我们的研究使用了最先进的目标检测模型,如 YOLOv4(只看一次,版本 4),这是一种非常有效的工具,能够实时处理由安装在无人机(UAV)四轴飞行器上的摄像机实时流式传输的 25fps、1920 X 1080 视频数据,以观察在 35m 范围内的区域内是否正确保持社交距离。该模型在识别和量化社交距离方面表现出了显著的效果,准确率为 82%,延迟很小。它能够以 25-30ms 的实时效率工作。我们的模型基于 CSPDarkNet-53,该模型在 MS COCO 数据集上进行图像分类训练。它包括附加的层,用于捕获来自不同阶段的特征图。此外,模型的颈部由 PANet 组成,用于聚合来自不同 CSPDarkNet-53 层的参数。CSPDarkNet-53 的 53 个卷积层之后,模型头部还有 53 个卷积层,总共 106 个完全卷积层。该架构进一步与 YOLOv3 集成,形成了我们的检测模型将使用的 YOLOv4 模型。此外,为了区分人类,我们使用上述方法来实时评估无人机拍摄的视频并计算社交距离违规行为。我们的研究结果表明,我们的模型能够可靠且成功地实时检测社交距离违规行为,平均准确率为 82%。