College of Electronic Science, National University of Defense Technology, Changsha 410073, China.
Sensors (Basel). 2019 Nov 20;19(23):5060. doi: 10.3390/s19235060.
Smoke detection technology based on computer vision is a popular research direction in fire detection. This technology is widely used in outdoor fire detection fields (e.g., forest fire detection). Smoke detection is often based on features such as color, shape, texture, and motion to distinguish between smoke and non-smoke objects. However, the salience and robustness of these features are insufficiently strong, resulting in low smoke detection performance under complex environment. Deep learning technology has improved smoke detection performance to a certain degree, but extracting smoke detail features is difficult when the number of network layers is small. With no effective use of smoke motion characteristics, indicators such as false alarm rate are high in video smoke detection. To enhance the detection performance of smoke objects in videos, this paper proposes a concept of change-cumulative image by converting the YUV color space of multi-frame video images into a change-cumulative image, which can represent the motion and color-change characteristics of smoke. Then, a fusion deep network is designed, which increases the depth of the VGG16 network by arranging two convolutional layers after each of its convolutional layer. The VGG16 and Resnet50 (Deep residual network) network models are also arranged using the fusion deep network to improve feature expression ability while increasing the depth of the whole network. Doing so can help extract additional discriminating characteristics of smoke. Experimental results show that by using the change-cumulative image as the input image of the deep network model, smoke detection performance is superior to the classic RGB input image; the smoke detection performance of the fusion deep network model is better than that of the single VGG16 and Resnet50 network models; the smoke detection accuracy, false positive rate, and false alarm rate of this method are better than those of the current popular methods of video smoke detection.
基于计算机视觉的烟雾检测技术是火灾检测中的一个热门研究方向。这项技术广泛应用于户外火灾检测领域(如森林火灾检测)。烟雾检测通常基于颜色、形状、纹理和运动等特征来区分烟雾和非烟雾物体。然而,这些特征的显著性和鲁棒性不够强,导致在复杂环境下烟雾检测性能较低。深度学习技术在一定程度上提高了烟雾检测性能,但在网络层数较小时,提取烟雾细节特征较为困难。由于没有有效利用烟雾运动特征,视频烟雾检测的误报率等指标较高。为了提高视频中烟雾目标的检测性能,本文提出了一种通过将多帧视频图像的 YUV 颜色空间转换为变化累积图像的概念,该图像可以表示烟雾的运动和颜色变化特征。然后,设计了一种融合深度网络,在其每个卷积层后增加两个卷积层,增加 VGG16 网络的深度。还使用融合深度网络排列 VGG16 和 Resnet50(深度残差网络)网络模型,以提高特征表达能力,同时增加整个网络的深度。这样可以帮助提取烟雾的额外鉴别特征。实验结果表明,通过使用变化累积图像作为深度网络模型的输入图像,烟雾检测性能优于经典的 RGB 输入图像;融合深度网络模型的烟雾检测性能优于单一的 VGG16 和 Resnet50 网络模型;该方法的烟雾检测准确率、误报率和误报率均优于当前流行的视频烟雾检测方法。