Open Access Paper
2 February 2023 Video-based smoke detection by using motion, color, and texture features
Author Affiliations +
Proceedings Volume 12462, Third International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022); 124620N (2023) https://doi.org/10.1117/12.2660953
Event: International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022), 2022, Xi'an, China
Abstract
Video smoke detection benefits life safety and environment protection, its early warning is of great importance. In response of many disadvantages of traditional smoke detectors, a method of video smoke detection based on dynamic, color and texture features are proposed. Firstly, the motion area is extracted through improved Vibe algorithm. Then, the suspected area corresponding to smoke is identified in CIELAB color space and segmented using a color filtering method. Finally, uniform local binary mode and gray level cooccurrence matrix are extracted from the image within suspected area and used to form the input vector of machine learning classifier for recognizing smoke. The classifier is tested with 400 images, and the results show that the detection system based on random forest algorithm has better performance, the selected smoke features have high recognition accuracy.

1.

INTRODUCTION

The fire accident usually causes economical and ecological damage as well as endangering people’s lives. Automatic visual object detection has been fully used for alarming fire disaster as soon as possible. Generally, most of the objects will generate smoke before it catches fire and this motivates that the smoke detection is employed to provide alarm of fire accident. However, the traditional smoke detection system is usually based on smoke sensors and difficult to play a complete role in outdoor environment, and it can’t provide more essential information about the process of fire accident. With the development of image processing technology, smoke detection based on vision technology has attracted researchers’ attention.

Several algorithms have been reported in the literature regarding video smoke detection. Emmy Prema et al. [1] converted the RGB color space to YUV space to determine the suspected smoke area. Ding et al. [2] proposed a method according to spectral features used in large and complex scenes. However, the accuracy of smoke detection is not very well for lack of features. Wu et al. [3] developed a new fire smoke model based on Gabor wavelet, and used the smoke change energy model and direction angle distribution model to analyze the dynamic properties of texture changes. Wang et al. [4] proposed a smoke detection method based on the swing and diffusion judgment model by analyzing the relevant characteristics of the gray level co-occurrence matrix (GLCM). Li et al. [5] used the MSER tracking algorithm to determine the smoke candidate regions in the video, and use static and dynamic features to identify smoke. James [6] used the optical flow method to extract smoke candidate regions and neural networks for classification and recognition. Liu et al. [7] divided the smoke into overlapping blocks, extracted the wavelet texture features and HSV color features of each image, obtained the latent semantic analysis. Some classical algorithms have been used for smoke detection, for example, Yuan used local binary pattern [8], Islam et al. used Gaussian Mixture Model [9].

Liu et al. [10] separated suspected smoke area in YUV color space, used the Local Binary Patterns (LBP) and discrete wavelet transform to extract smoke features, the AdaBoost classifier is used for image classification. But when the volume fraction is small, the system cannot quickly identify smoke. Wang [11] proposed an image enhancement method based on fuzzy logic to alleviate the interference of low and uneven lighting in coal mine, and used support vector machine (SVM) to classify smoke by features of perimeter and area ratio, area randomness and drift characteristics. And the convolutional neural network CNN has been widely used in video smoke detection [12-14].

In this paper, multi-feature fusion method is used for smoke detection. Based on the motion area extracted by the improved Vibe algorithm, the suspected smoke area is segmented in the CIELAB color space by color filtering rules, and then the texture features of smoke are mainly discussed. The uniform local binary pattern codes (ULBP), and GLCM are used. Machine learning classifiers SVM and random forest are used for model training to distinguish smoke and non-smoke. The optimal algorithm is selected from the analysis results, and the test is carried out through the simulated fire smoke monitoring video to verify the accuracy of the detection method.

In the following sections, the method including motion objects detection, color detection and texture detection in detail is described in Section II. Section III gives experimental results while conclusions are drawn in Section IV.

2.

FEATURES EXTRACTION

2.1

Motion objects detection

Aiming at the relationship between the upward movement characteristics of smoke and fixed surveillance cameras, the moving target detection method under static background is used to extract the moving area. The Vibe algorithm introduces the randomization method into the background model for the first time, which is fast, has a small calculation and a certain robustness to noise. The detection area is more complete, and it can quickly achieve background extraction and moving target detection [15].

However, the original Vibe algorithm has the following problems when extracting the motion area: the algorithm is updated too fast, and sometimes the initial smoke movement is slow, which will absorb the smoke area as the background, resulting in no smoke detected; if the algorithm is updated too slowly, it will leave shadow; misclassification of pixels may be caused due to parameter setting issues. Aiming at the problems existing in the original Vibe algorithm, the improved algorithm is proposed.

Firstly, the sampling range is extended from 8 neighborhoods to 12 neighborhoods. Enlarging the sampling range can avoid a pixel from being sampled multiple times and improve the quality of the initial model.

Figure 1.

12 neighborhoods model

00023_PSISDG12462_124620N_page_2_1.jpg

The original Vibe algorithm is performed by calculating the current pixel and the sample set the Euclidean distance D between each pixel for foreground and background segmentation. As is shown in Figure2, set v(x) (current point)as the center, R as radius, v1, v2, v3vn is grayscale values in sample set. Count each distance of sample set with v(x), if the count of distance which is less than threshold T, then take v(x)as foreground, otherwise, take it as background. In original Vibe algorithm, R = 20, T = 2, n = 20.

Figure 2.

Pixel classification in Euclidean distance

00023_PSISDG12462_124620N_page_2_2.jpg

On the basis of the original Vibe algorithm, the OTSU is introduced to calculate the optimal segmentation threshold of the current frame, and the foreground points are double-discriminated [16]. OTSU is defined as:

00023_PSISDG12462_124620N_page_3_1.jpg

Where the subscript 0, 1 and without subscript indicate foreground, background and entire image, respectively. ω indicate the proportion of the number of pixels, μ are the average grayscale value. Set the segmentation threshold be t, and obtained when the σ2 reaches the maximum.

In the detection results of the traditional vibe algorithm, N background pixels are randomly selected, the average gray value VN of these background points is calculated, and the secondary discrimination is performed:

00023_PSISDG12462_124620N_page_3_2.jpg

Secondly, a dynamic threshold is proposed: taking the average distance meanD(x, y) from each pixel to the background sample set first. The dynamic threshold is calculated as follows:

00023_PSISDG12462_124620N_page_3_3.jpg

In this paper, α=0.5, α2 =0.25, β =3, r is original radius parameter 20. The larger the meanD(x, y) is, the more complex the background is, so a larger R is needed, otherwise a smaller R is required. The parameter β is adjusted to an appropriate value according to the background change, and take β meanD(x, y) as the threshold judgment standard.

After morphological processing operations such as dilation, erosion, opening operation and closing operation, the detection result of smoke video motion area is shown in the figure3.

Figure 3.

The result of motion detection

00023_PSISDG12462_124620N_page_3_4.jpg

The motion area extracted by the motion detection method includes not only the smoke diffusion area, but also may moving vehicles, walking person, etc. Therefore, further segment is needed to obtain a more accurate smoke motion area.

2.2

Color Extraction

The color attributes have an important role in smoke detection. Color is a powerful descriptive factor in human vision, it carries a large amount of visual information. Although the smoke area has been determined by the motion extraction, there have non-smoke motion areas in these areas that are different in color with smoke. It would cost much time if feature extraction is performed directly on this result. Therefore, color segmentation is performed on the motion region. CIELAB [17] is a color system of CIE that is used to confirm the numerical information of color. It can express colors based on human vision system. The CIELAB values cannot be obtained directly from RGB, but can be obtained roughly from sRGB [18]:

00023_PSISDG12462_124620N_page_4_1.jpg
00023_PSISDG12462_124620N_page_4_2.jpg

Among this, Xn =0.950456, Yn =1, Zn =1.088754.

00023_PSISDG12462_124620N_page_4_3.jpg

f(X), f(Y), f(Z) can be expressed as the filter:

00023_PSISDG12462_124620N_page_4_4.jpg

By extracting and analyzing the color components of the smoke image in CIELAB color spaces, a color statistical model based on the Lab components is proposed.

00023_PSISDG12462_124620N_page_4_5.jpg

C stands for chroma which indicates the vividness of the color.

After using the CIELAB color filtering to segment the suspected smoke area, although most of the interference between the color and the smoke moving area can be eliminated (see Fig 4 and 5), there may still be very few moving areas similar to the smoke color. Therefore, in-depth analysis based on more essential characteristics of the smoke image should be carried out.

Figure 4.

Image before and after color filtering

00023_PSISDG12462_124620N_page_4_6.jpg

Figure 5.

Suspected smoke area before and after filtering

00023_PSISDG12462_124620N_page_4_7.jpg

2.3

Texture detection

Texture feature is a global feature, which reflects the organization of slow transformation and periodic change on the surface of the object. Better results can be achieved by texture feature in recognizing smoke image. LBP and GLCM are the most widely used method in texture feature extraction.

The earliest LBP descriptor was proposed by Ojala et al. [19] for the texture image classification. Taking the gray value of the center pixel as a threshold for binary quantization, then the neighborhood gray value is encoded as 1 if it is greater than or equal to the gray value of the center point, otherwise, it will be encoded as 0. Then, the LBP value of center pixel is obtained by transforming the 8-bit binary number into decimal number. In the meanwhile, there will be 28=256 patterns.

The uniform LBP [20] was proposed to achieve effective dimensionality reduction, which is defined as when the cyclic binary number corresponding to a certain LBP has at most two jumps from 0 to 1 or from 1 to 0. The binary corresponding to that LBP is called the uniform class, which is proven to be a powerful texture. The rest of the patterns are grouped into another class, called the mixed pattern class. Therefore, the generated binary codes are classified into 58 uniform patterns and one mixed pattern, for reducing computational complexity.

GLCM [21] can reflect the comprehensive information of image gray scale, such as direction, adjacent interval and amplitude of change, but it cannot directly provide the features of different textures. Therefore, statistical parameters that can quantitatively describe texture features are extracted by counting the values of normalized GLCM.

00023_PSISDG12462_124620N_page_5_1.jpg
00023_PSISDG12462_124620N_page_5_2.jpg
00023_PSISDG12462_124620N_page_5_3.jpg
00023_PSISDG12462_124620N_page_5_4.jpg

Where (i, j) is the grayscale value of any point(x, y) in image and another point(x + Δx, y + Δy), their distance is d and the angle is θ, let the level of gray value be g. For the whole image, count the number of occurrences of every (i, j) and arrange them into a square matrix. Normalize the occurrences to get the probability, and P(i, j,d,θ) is obtained. Among this, 00023_PSISDG12462_124620N_page_5_5.jpg, 00023_PSISDG12462_124620N_page_5_6.jpg, 00023_PSISDG12462_124620N_page_5_7.jpg, 00023_PSISDG12462_124620N_page_5_8.jpg.

In this paper, energy characteristics, entropy, correlation and contrast, 4 directions and 16 features vectors are selected to further identify smoke.

3.

EXPERIMENTS

3.1

Experiment configuration and evaluation index

The first experiment for comparing between different texture features and different classifiers was first conducted based on Windows 11 system with Intel Core i5 CPU; the software is Matlab2018. In the experiment, the smoke dataset from the National Laboratory of Fire Science, University of Science and Technology of China is used [22]. It includes 800 still images of 100×100 pixels, dividing into two groups, 400 images containing smoke and 400 images not containing smoke. 59 uniform LBP patterns and 16 GLCM paraments, totally 75 feature vectors were extracted and feed into a classifier for training to obtain a classification model. SVM and random forest (RF) were compared.

In order to measure the performance of the algorithm quantitatively, the confusion matrix is chosen as the index to evaluate the retrieval performance of the algorithm, TN denote the number of samples which is predicted to be negative from smokeless, FP denote the number of samples which is predicted to be positive from smokeless, FN denote the number of samples which is predicted to be negative from smoke and TP denote the number of samples which is predicted to be positive from smoke. Accuracy (Acc), precision (Prec), recall rate (Rec) and the value of in (13)-(16) is used as measures for testing algorithms.

00023_PSISDG12462_124620N_page_5_9.jpg
00023_PSISDG12462_124620N_page_6_1.jpg
00023_PSISDG12462_124620N_page_6_2.jpg
00023_PSISDG12462_124620N_page_6_3.jpg

3.2

Experiment results

Three different features, LBP, GLCM and the combination of LBP and GLCM, and two classifiers, SVM and random forest (RF), were compared in the experiment. 400 random images, half with smoke and half without smoke, were selected as the training images and test images respectively. The comparison result is presented in Table I and Figure 6 for 3 texture features and two classifiers. In Table I and Figure 6, LG denotes the combination of LBP and GLCM features, RF denotes random forest.

Figure 6.

Comparison of evaluation index of classification

00023_PSISDG12462_124620N_page_6_4.jpg

Table 1.

Evaluation of different method

IndexLBPGLCMLG
SVMRFSVMRFSVMRF
Acc83.090.591.087.081.096.0
Prec81.786.579.877.579.295.1
Rec85.096.091.093.084.097.0
F183.391.085.084.581.596.0

It can be seen from Table I that the Acc, Prec, Rec, and value of SVM is better than those of RF when only classifying GLCM features. However, when it comes to LBP and the combination of LBP and GLCM, the result of RF is better than SVM. By analyzing the characteristics of the SVM, it can be seen that when there are too many features, it is difficult to find an optimal hyperplane to divide two types of samples, the reliability of segmentation is not enough, and the generalization ability is insufficient. Therefore, SVM has better effect on small sample training, but if the number of image features is large, the classification effect is poor. By contrast, RF has a high degree of self-adaptation and good generalization ability.

In the second experiment, the RF classifier that has been trained is used to test smoke video. In the experiment, the proposed Vibe algorithm and color filter were implemented to segment suspected smoke area, and then combination of LBP and GLCM features were extracted and feed into trained RF classifier. The results were compared that from with Liu’s method [23] as is shown in the Table II. It used background modeling and color information to mix the static and dynamic features, then SVM was used to discriminate texture features. On the whole, the performances of the proposed method outperform Liu’s method. Liu’s algorithm is easy to be influenced by surroundings, and some object with subtle texture, blurred color and irregular movement are likely to be detected as smoke.

Table 2.

Smoke video detection results

Smoke videosNumber of framesProposed methodLiu’s method
Smoke With gray walls51797.8100
Smoke With walking people288698.475.4
Smoke With green leaves120099.693

4.

CONCLUSION

This paper proposes an effective smoke detection method based on surveillance video images. Combined with the proposed motion feature and color filtering, the complete smoke area is segmented first to improve detection. The eigenvectors extracted by the uniform LBP and the GLCM can effectively express the essential characteristics of smoke; the random forest classifier is selected for distinguishing smoke and non-smoke. The test results show that the method has higher detection efficiency, compared with traditional smoke detectors.

ACKNOWLEDGMENT

This work was supported by the National Natural Science Foundation of China (No.61775170).

REFERENCES

[1] 

C. Emmy P and S. S. Vinsley and S. Suresh, “Multi Feature Analysis of Smoke in YUV Color Space for Early Forest Fire Detection,” Fire Technology, 52 1319 –1342 (2016). https://doi.org/10.1007/s10694-016-0580-8 Google Scholar

[2] 

DING Xiong, LU Yan, “Early smoke detection of forest fires based on SVM image segmentation,” Journal of Forest Science, 65 150 –159 (2019). https://doi.org/10.17221/82/2018-JFS Google Scholar

[3] 

WU Zhangxian, YANG Guotian, LIU Xiangjie, “A new method for fire smoke identification based on Gabor wavelet,” Journal of Instrumentation, 31 1 –7 (2010). Google Scholar

[4] 

Shidong W, “Early smoke detection in video using swaying and diffusion feature,” Journal of Intelligent & Fuzzy Systems, 26 267 –275 (2014). https://doi.org/10.3233/IFS-120735 Google Scholar

[5] 

S. Li, B. Wang, L. Gong, “A novel smoke detection algorithm based on MSER tracking,” in The 27th Chinese Control and Decision Conference (2015 CCDC), 5676 –5681 (2015). Google Scholar

[6] 

James M., “Flame and smoke estimation for fire detection in video based on optical flow and neural networks,” International Journal of Research in Engineering and Technology, 03 324 –328 (2014). https://doi.org/10.15623/ijret Google Scholar

[7] 

LIU Ying, GU Xiaodong, LI Daxiang, “Fire and Smoke Detection Algorithm Based on LSA and SVM,” Journal of Xi’an University of Posts and Telecommunications, 19 6 –10 (2014). Google Scholar

[8] 

Yuan, F., Shi, J., Xia, X., Yang, Y., Fang, Y., Wang, R., “Sub oriented histograms of local binary patterns for smoke detection and texture classification,” KSII Trans. Internet Informat. Syst. (TIIS), 10 (4), 1807 –1823 (2016). Google Scholar

[9] 

Islam, M.R., Amiruzzaman, M., Nasim, S., Shin, J, “Smoke object segmentation and the dynamic growth feature model for video-based smoke detection systems,” Symmetry, 12 1075 –1093 (2020). https://doi.org/10.3390/sym12071075 Google Scholar

[10] 

LIU Kai, LIU Xiang, CHANG Liping, “Video smoke detection based on YUV color space and multiple feature fusion,” Chinese Journal of Sensors and Actuators, 32 237 –244 (2019). Google Scholar

[11] 

WANG Yuanbin, “Research on fire detection and recognition in coal mine based on image features,” Xi ′ an University of Science and Technology, Xi′ an (2013). Google Scholar

[12] 

Frizzi S, Kaabi R, Bouchouicha M, “Convolutional neural network for video fire and smoke detection,” in Conference of the IEEE Industrial Electronics Society, 877 –882 (2016). Google Scholar

[13] 

Gagliardi A, Gioia F D, Saponara S, “A real-time video smoke detection algorithm based on Kalman filter and CNN,” Journal of Real-Time Image Processing, 2085 –2095 2021). https://doi.org/10.1007/s11554-021-01094-y Google Scholar

[14] 

Hohberg S P, “Wildfire Smoke Detection using Convolutional Neural Networks,” Freie Universität Berlin, (2015). Google Scholar

[15] 

Barnich O, Droogenbroeck M V, “ViBe: A Universal Background Subtraction Algorithm for Video Sequences,” in Image Processing of the IEEE Transactions, 1709 –1724 (2011). Google Scholar

[16] 

Otsu N., “A threshold selection method from Gray-Level Histograms,” Trransactions on System Man and Cybernetic of the IEEE, 9 62 –66 1979). https://doi.org/10.1109/TSMC.1979.4310076 Google Scholar

[17] 

CIE, Colorimrtry. CIE, 15 2004 CIE Central Bureau, Vienna (2004). Google Scholar

[18] 

Stokes M, Anderson M, Chandrasekar S, Motta R, “A Standard Default Color Space for the Internet — sRGB,” (1996). Google Scholar

[19] 

T. Ojala, M. Pietikainen and D. Harwood, “A Comparative Study of Texture Measures with Classification Based on Feature Distributions,” Pattern Recognition, 29 51 –59 (1996). https://doi.org/10.1016/0031-3203(95)00067-4 Google Scholar

[20] 

T. Ojala, M. Pietikainen and T. Maenpaa, “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” IEEE Trans. Pattern Anal. Mach. Intell, 24 971 –987 (2002). https://doi.org/10.1109/TPAMI.2002.1017623 Google Scholar

[21] 

HARALICK R M, “Statistical and structural approaches to texture,” in Proceedings of the IEEE, 786 (2005). Google Scholar

[23] 

LIU Minjie, “Design and implementation of video-based fire smoke detection,” Nanjing University of Posts and Telecommunications, (2019). Google Scholar
© (2023) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Zheng Zhao, Guihua Cui, Donghui Li, and Qifan Chen "Video-based smoke detection by using motion, color, and texture features", Proc. SPIE 12462, Third International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022), 124620N (2 February 2023); https://doi.org/10.1117/12.2660953
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Binary data

Detection and tracking algorithms

Video

Feature extraction

Motion detection

RGB color model

Image segmentation

RELATED CONTENT

Human detection and tracking
Proceedings of SPIE (October 26 2013)
Salient region detection for object tracking
Proceedings of SPIE (May 08 2012)
Human visual system-based smoking event detection
Proceedings of SPIE (May 08 2012)

Back to Top