3.2 Classification results
19552 samples of beech, cotton, polyurethane, n-heptane smokes, and oil fume are obtained on the experimental platform for the classification which are divided into training and test sets in a ratio of 7:3. These algorithms run in PyCharm (PyCharm Community Edition 2020.1) using PC of Intel(R) Core(TM) i7-10510U CPU @ 1.80 GHz 2.30 GHz. The EST-CNN model is trained and tested using multiple sets of parameters to find the optimal parameters with low amount of network parameters and high classification accuracy.
CNN for feature learning and extraction typically have more than 3 convolutional layers used to increase and decrease the dimensionality of the data. To realize the model running on low-computing power embedded chips, the number of parameters of the model is required to be as few as possible. Therefore, the EST-CNN model applies only 3 convolutional layers in the feature extraction network, and the input data size of the first layer is 4 (number of channels) × 4 (sampling time points). The previous input and final output channels of the feature extraction network are connected with the 2D-TS matrix and full connected classification layer, respectively, which are set to constant values of 8 and 64. In this way, the main parameters that can be adjusted in the model are the number of input and output channels in the middle layer, stride, and padding. For the channels, the number of output channels in each layer of the network is required to be the same as the number of input channels in the next layer. Stride represents the number of data that the convolution kernel slides over in each slide. Padding is the complementary zeros around the boundaries of the input matrix. The parameters of the partial sets of the feature extraction network are shown in Table. 3, and the corresponding training and testing results are shown in Figs. 5-9. In these figures, Class 1, 2, 3, 4, and 5 represent the beech smoke, cotton smoke, polyurethane smoke, n-heptane smoke, and oil fume, respectively.
Table 3.
The parameters of the feature extraction network.
Table 3.
The parameters of the feature extraction network.
Set |
Layer number |
Layer type |
Input channel |
Output channel |
Convolutional kernel size |
Stride |
Padding |
Parameters |
1 |
1 |
Conv |
8 |
16 |
3×3 |
1 |
1 |
66 kB |
2 |
Conv |
16 |
32 |
3×3 |
1 |
0 |
3 |
Conv |
32 |
64 |
2×2 |
1 |
0 |
2 |
1 |
Conv |
8 |
32 |
3×3 |
1 |
1 |
158 kB |
2 |
Conv |
32 |
64 |
3×3 |
1 |
0 |
3 |
Conv |
64 |
64 |
2×2 |
1 |
0 |
3 |
1 |
Conv |
8 |
32 |
3×3 |
1 |
1 |
295 kB |
2 |
Conv |
32 |
128 |
3×3 |
1 |
0 |
3 |
Conv |
128 |
64 |
2×2 |
1 |
0 |
4 |
1 |
Conv |
8 |
16 |
3×3 |
1 |
1 |
66 kB |
2 |
Conv |
16 |
32 |
3×3 |
2 |
1 |
3 |
Conv |
32 |
64 |
2×2 |
1 |
0 |
5 |
1 |
Conv |
8 |
16 |
3×3 |
1 |
1 |
66 kB |
2 |
Conv |
16 |
32 |
3×3 |
2 |
1 |
3 |
Conv |
32 |
64 |
2×2 |
2 |
0 |
As shown in Table. 3, both the first convolutional layer and the second convolutional layer of
Set. 2 have more output channels than those of
Set. 1. Similarly, those of
Set. 3 have more channels than those of
Set. 2. As a result, the number of parameters in
Set. 2 and
Set. 3 are respectively 2.4 and 4.5 times that of
Set. 1. However, it can be seen from Figs. 5a and 6a that the
score of
Set. 1 and
Set. 2 are very close to each other. This demonstrates that trained under both sets of parameters, the network model converges well to reliable classification performance. Meanwhile, as shown in Figs. 5b and 6b, test results indicate that the classification accuracy of beech smoke is higher in
Set. 1 than in
Set. 2, oil fume is slightly lower than in
Set. 2, and the other three classes of smoke have almost the same classification accuracy under
Set. 1 and
Set. 2. Thus, it is considered that the increased number of parameters in
Set. 2 is meaningless. Moreover, as shown in
Figure 7a, the
score of oil fume PR curve in
Set. 3 is only 0.9141, which is lower than that in
Set. 1 and
Set. 2. Combined with
Figure 7b, it can be seen that while the parameters selected in
Set. 3 resulted in an increase in the classification accuracy of cotton, polyurethane, and n-heptane smokes, the missed alarm rate for oil fume increased dramatically. Therefore, the number of input and output channels of each layer in
Set. 1 is optimal.
Figure 5.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 1 network parameters. a Training result, b test result.
Figure 5.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 1 network parameters. a Training result, b test result.
Figure 6.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 2 network parameters. a Training result, b test result.
Figure 6.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 2 network parameters. a Training result, b test result.
Figure 7.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 3 network parameters. a Training result, b test result.
Figure 7.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 3 network parameters. a Training result, b test result.
To further reduce the number of parameters, the stride and padding are adjusted, as
Set. 4 and
5 in Table. 3. The convolution process is essentially a multiplication of two matrices, after the convolution process, the original input matrix will have a certain degree of shrinkage, when the stride is 1, the matrix length and width will be reduced by 2. When the length and width of the original input matrix are small (e.g., 4 × 4 of the 2D-TS matrix), padding the data with zeros is required before the convolution operation in order to keep the output matrix meaningful after each layer of convolution. As a result, the stride and padding usually are adjusted simultaneously, and the larger the stride the more zeros need to be padded. As shown in Table. 3,
Set. 4 increases the stride of only the second convolutional layer, while
Set. 5 increases the stride of both the second and third convolutional layers. Nevertheless, the number of parameters in
Set. 4 and
Set. 5 are not reduced and remain at 66 kB. However, it can be seen in Figs. 8 and 9 that the classification accuracy of oil fume in
Set. 4 and n-heptane in
Set. 5 is significantly reduced compared to those in
Set. 1 as the classification results in
Figure 5.
Figure 8.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 4 network parameters. a Training result, b test result.
Figure 8.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 4 network parameters. a Training result, b test result.
Figure 9.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 5 network parameters. a Training result, b test result.
Figure 9.
Classification results of five classes of real fire smokes and interferential aerosol (oil fume) under EST-CNN model using Set. 5 network parameters. a Training result, b test result.
Combining
Table 3 and Figs. 5-9, the average classification accuracies of fire smokes and oil fume under the five sets of model parameters are 95.6%, 96.0%, 89.4%, 92.6% and 93.0%, respectively, corresponding to the parametric quantities of 66 kB, 158 kB, 295 kB, 66 kB and 66 kB. In summary, due to the small size of the scattering feature matrix (2D-TS matrix), the parameters that can be adjusted in the convolutional layer are limited. The experimental results show that adjusting the parameters of the convolutional operation (stride and padding) has almost no influence on the number of parameters of the network model while decreasing the classification accuracy, and adjusting the number of input and output channels is an effective method to reduce the number of parameters. Therefore, the
Set. 1 parameters, which have the best comprehensive performance, are recognized as the actual parameters chosen for EST-CNN.