Continuous Wavelet Transform Peak-Seeking Attention Mechanism Conventional Neural Network: A Lightweight Feature Extraction Network with Attention Mechanism Based on the Continuous Wave Transform Peak-Seeking Method for Aero-Engine Hot Jet Fourier Transform Infrared Classification
Abstract
:1. Introduction
- This paper provides a classification method for aircraft engines based on infrared spectroscopy technology. The selective absorption of infrared radiation by different molecules is an important method for determining different substances. Similarly, the different spectral features in the infrared spectra of hot jet gases from different types of aircraft engines help to classify them.
- Due to the scarcity of infrared spectral data for aircraft engine thermal jets, this paper constructed a new benchmark data set. The data set covers the infrared spectrum in the wavelength range of 2.5~12 μm, including six types of different aero-engine models (including turbine engine and turbofan engine).
- This paper provides a deep learning framework for the classification of aero-engine hot jet infrared spectra. A convolutional Neural Network based on a peak-seeking attention mechanism was designed. The backbone network consisted of three feature extraction blocks with the same structure, batch normalization layer and maximum pooling layer. In the part of attention mechanism based on peak seeking, the spectral peak value was detected by continuous wavelet transform method, and the peak wave number of high-frequency occurrences was counted. The attention mechanism weighted the peak value obtained by statistics and acted on the feature map of the trunk CNN. The structure of the network was light, and the classification accuracy and operation efficiency could be taken into account.
2. Spectral Classification Network Structure Design
2.1. Overall Network Design
2.2. Backbone Network Design
2.2.1. One-Dimensional Convolutional Layer (Cov1D Layer)
2.2.2. Batch Normalization (BN)
2.2.3. Maximum Pooling Layer
2.2.4. Flatten Layer
2.2.5. Fully Connected Layer (FC Layer)
2.3. Attention Mechanism Based on Peak Seeking
2.3.1. Peak-Seeking Algorithm Block
Algorithm 1: Peak-seeking algorithm and peak statistics |
Input: Spectral data. |
Output: Peak data. |
|
2.3.2. Attention Mechanism
Algorithm 2: CNN with attention Mechanism |
Input: Spectral data, peak data. |
Output: Prediction label for prediction data set. |
|
2.4. Network Training Method
2.4.1. Optimizer
2.4.2. Loss Function
2.4.3. Activation Function
3. Spectral Data Set
3.1. Design of Aero-Engine Spectrum Measurement Experiment
3.2. Data Preprocessing
3.3. Data Set Production
4. Experiments and Results
4.1. Performance Measures and Experimental Results
- ①
- Accuracy: the ratio of correctly classified samples to the total number of samples.
- ②
- Precision: the ratio of the number of true positive samples to the total number of samples predicted as positive.
- ③
- Recall: the ratio of the number of samples correctly predicted to be in the positive category to the number of samples in the true positive category.
- ④
- F1-score: a metric that quantifies the overall performance of a model by combining the harmonic mean of precision and recall.
- ⑤
- Confusion matrix: The confusion matrix provides a comprehensive evaluation of the classifier’s performance in classifying various categories. It displays the discrepancy between actual value and predicted values. The diagonal elements of the matrix indicate the number of accurate predictions generated by the classifier for each category. Table 6 displays the confusion matrix.
4.2. Comparative Experimental Results of Traditional Classification Methods
4.3. Comparative Experimental Results of Deep Learning Classification Methods
4.4. Analysis of Ablation Study
4.4.1. Effectiveness of Peak Features
4.4.2. Effectiveness of AM
4.4.3. Comparison of Network Design
- ①
- ②
- Network depth: In deep learning algorithms, network depth carries out a decisive role in network expression. The deeper the depth is, the better the network expression is, because network depth determines the quality of features from aspects such as invariance and abstraction. Therefore, we conducted an experimental comparison of networks with different layer structures. Each layer used a feature extraction block and used the same loss function, optimizer and learning rate to obtain Table 19 and Figure 12.
- ③
- ④
4.4.4. Running Time
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Razeghi, M.; Nguyen, B.-M. Advances in mid-infrared detection and imaging: A key issues review. Rep. Prog. Phys. 2014, 77, 082401. [Google Scholar] [CrossRef] [PubMed]
- Chikkaraddy, R.; Arul, R.; Jakob, L.A.; Baumberg, J.J. Single-molecule mid-IR detection through vibrationally-assisted luminescence. Nat. Photonics 2022, 7, 865–871. [Google Scholar]
- Knez, D.; Toulson, B.W.; Chen, A.; Ettenberg, M.H.; Nguyen, H.; Potma, E.O.; Fishman, D.A. Spectral imaging at high definition and high speed in the mid-infrared. Sci. Adv. 2022, 8, eade4247. [Google Scholar] [CrossRef]
- Zhang, J.; Gong, Y. Automated identification of infrared spectra of hazardous clouds by passive FTIR remote sensing. In Multispectral & Hyperspectral Image Acquisition & Processing; International Society for Optics and Photonics: Bellingham, DC, USA, 2001. [Google Scholar]
- Roh, S.B.; Oh, S.K. Identification of Plastic Wastes by Using Fuzzy Radial Basis Function Neural Networks Classifier with Conditional Fuzzy C-Means Clustering. J. Electr. Eng. Technol. 2016, 11, 103–116. [Google Scholar] [CrossRef]
- Kumar, V.; Kashyap, M.; Gautam, S.; Shukla, P.; Joshi, K.B.; Vinayak, V. Fast Fourier infrared spectroscopy to characterize the biochemical composition in diatoms. J. Biosci. 2018, 3, 717–729. [Google Scholar] [CrossRef]
- Han, X.; Li, X.; Gao, M.; Tong, J.; Wei, X.; Li, S.; Ye, S.; Li, Y. Emissions of Airport Monitoring with Solar Occultation Flux-Fourier Transform Infrared Spectrometer. J. Spectrosc. 2018, 2018, 1–10. [Google Scholar] [CrossRef]
- Cięszczyk, S. Passive open-path FTIR measurements and spectral interpretations for in situ gas monitoring and process diagnostics. Acta Phys. Pol. A 2014, 126, 673–678. [Google Scholar] [CrossRef]
- Schütze, C.; Lau, S.; Reiche, N.; Sauer, U.; Borsdorf, H.; Dietrich, P. Ground-based remote sensing with open-path Fourier-transform infrared (OP-FTIR) spectroscopy for large-scale monitoring of greenhouse gases. Energy Procedia 2013, 37, 4276–4282. [Google Scholar] [CrossRef]
- Yang, L.; Tao, Z. Aircraft image recognition in airport flight area based on deep transfer learning. In Proceedings of the International Conference on Smart Transportation and City Engineering 2021, Chongqing, China, 6–8 August 2021. [Google Scholar]
- Shen, H.; Huo, K.; Qiao, X.; Li, C. Aircraft target type recognition technology based on deep learning and structure feature matching. Intell. Fuzzy Syst. 2023, 45, 5685–5696. [Google Scholar] [CrossRef]
- Li, X.; Li, Z.; Qiu, H.; Hou, G.; Fan, P. An overview of hyperspectral image feature extraction, classification methods and the methods based on small samples. Appl. Spectrosc. Rev. 2021, 58, 367–400. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Audebert, N.; Le Saux, B.; Lefevre, S. Deep Learning for Classification of Hyperspectral Data: A Comparative Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
- Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution from Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
- Lu, W.; Wang, X.; Sun, L.; Zheng, Y. Spectral–Spatial Feature Extraction for Hyperspectral Image Classification Using Enhanced Transformer with Large-Kernel Attention. Remote Sens. 2024, 16, 67. [Google Scholar] [CrossRef]
- Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
- Kemker, R.; Kanan, C. Self-Taught Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2693–2705. [Google Scholar] [CrossRef]
- Ding, Y.; Zhao, X.; Zhang, Z.; Cai, W.; Yang, N. Graph Sample and Aggregate-Attention Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Li, Z.; Zhao, B.; Wang, W. An Efficient Spectral Feature Extraction Framework for Hyperspectral Images. Remote Sens. 2020, 12, 3967. [Google Scholar] [CrossRef]
- Ashraf, M.; Zhou, X.; Vivone, G.; Chen, L.; Chen, R.; Majdard, R.S. Spatial-Spectral BERT for Hyperspectral Image Classification. Remote Sens. 2024, 16, 539. [Google Scholar] [CrossRef]
- Hamouda, M.; Ettabaa, K.S.; Bouhlel, M.S. Smart feature extraction and classification of hyperspectral images based on convolutional neural networks. IET Image Process. 2020, 14, 1999–2005. [Google Scholar] [CrossRef]
- Zhang, J.; You, S.; Liu, A.; Xie, L.; Huang, C.; Han, X.; Li, P.; Wu, Y.; Deng, J. Winter Wheat Mapping Method Based on Pseudo-Labels and U-Net Model for Training Sample Shortage. Remote Sens. 2024, 16, 2553. [Google Scholar] [CrossRef]
- Dou, P.; Huang, C.; Han, W.; Hou, J.; Zhang, Y.; Gu, J. Remote sensing image classification using an ensemble framework without multiple classifiers. ISPRS J. Photogramm. Remote Sens. 2024, 208, 190–209. [Google Scholar] [CrossRef]
- Wang, M.; Zhang, X.; Niu, X.; Wang, F.; Zhang, X. Scene Classification of High-Resolution Remotely Sensed Image Based on ResNet. J. Geovisualization Spat. Anal. 2019, 3. [Google Scholar] [CrossRef]
- Faqe Ibrahim, G.R.; Rasul, A.; Abdullah, H. Improving Crop Classification Accuracy with Integrated Sentinel-1 and Sentinel-2 Data: A Case Study of Barley and Wheat. J. Geovisualization Spat. Anal. 2023, 7, 22. [Google Scholar] [CrossRef]
- Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
- Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Li, X.; Hu, X.; Yang, J. Spatial Group-wise Enhance: Improving Semantic Feature Learning in Convolutional Networks. arXiv 2019, arXiv:1905.09646. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Doubenskaia, M.; Pavlov, M.; Grigoriev, S.; Smurov, I. Definition of brightness temperature and restoration of true temperature in laser cladding using infrared camera. Surf. Coat. Technol. 2013, 220, 244–247. [Google Scholar] [CrossRef]
- Homan, D.C.; Cohen, M.H.; Hovatta, T.; Kellermann, K.I.; Kovalev, Y.Y.; Lister, M.L.; Popkov, A.V.; Pushkarev, A.B.; Ros, E.; Savolainen, T. MOJAVE. XIX. Brightness Temperatures and Intrinsic Properties of Blazar Jets. Astrophys. J. 2021, 923, 67. [Google Scholar] [CrossRef]
- Du, S.; Han, W.; Shi, Z.; Liao, Y.; Li, Z. An Aero-Engine Classification Method Based on Fourier Transform Infrared Spectrometer Spectral Feature Vectors. Electronics 2024, 13, 915. [Google Scholar] [CrossRef]
Name | Manufacturer | Measurement Pattern | Spectral Resolution (cm−1) | Spectral Measurement Range (µm) | Full Field of View Angle |
---|---|---|---|---|---|
EM27 | Bruker | Active/Passive | Active: 0.5/1 Passive: 0.5/1/4 | 2.5~12 | 30 mrad (no telescope) (1.7°) |
Telemetry Fourier Transform Infrared Spectrometer | Aerospace Information Research Institute | Passive | 1 | 2.5~12 | 1.5° |
Aero-Engine Serial Number | Environmental Temperature | Environmental Humidity | Detection Distance |
---|---|---|---|
Turbofan engine 1 | 19°C | 58.5%Rh | 5 m |
Turbofan engine 2 | 16°C | 67%Rh | 5 m |
Turbojet engine | 14°C | 40%Rh | 5 m |
Turbojet UAV | 30°C | 43.5%Rh | 11.8 m |
Turbojet UAV with propeller at tail | 20°C | 71.5%Rh | 5 m |
Turbojet manned aircraft | 19°C | 73.5%Rh | 10 m |
Label | Type | Number of Data Pieces | Number of Error Data | Full Band Data Volume | Medium Wave Range Data Volume |
---|---|---|---|---|---|
1 | Turbofan engine 1 | 792 | 17 | 16,384 (1 cm−1)/32,768 (0.5 cm−1) | 7464/14,928 |
2 | Turbofan engine 2 | 258 | 2 | 16,384 (1 cm−1)/32,768 (0.5 cm−1) | 7464/14,928 |
3 | Turbojet engine | 384 | 4 | 16,384 (1 cm−1)/32,768 (0.5 cm−1) | 7464/14,928 |
Label | Type | Number of Data Pieces | Number of Error Data | Full Band Data Volume | Medium Wave Range Data Volume |
---|---|---|---|---|---|
1 | Turbojet UAV | 193 | 0 | 16,384 | 7464 |
2 | Turbojet UAV with propeller at tail | 48 | 0 | 16,384 | 7464 |
3 | Turbojet manned aircraft | 202 | 3 | 16,384 | 7464 |
Label | Type | Number of Data Pieces | Number of Error Data | Full Band Data Volume | Medium Wave Range Data Volume |
---|---|---|---|---|---|
1 | Turbojet UAV | 193 | 0 | 16,384 | 7464 |
2 | Turbojet UAV with propeller at tail | 48 | 0 | 16,384 | 7464 |
3 | Turbojet manned aircraft | 202 | 3 | 16,384 | 7464 |
4 | Turbofan engine 1 | 792 | 17 | 16,384 | 7464 |
5 | Turbofan engine 2 | 258 | 2 | 16,384 | 7464 |
6 | Turbojet engine | 384 | 4 | 16,384 | 7464 |
Forecast Results | |||
---|---|---|---|
Positive Samples | Negative Samples | ||
Real results | Positive samples | TP | TN |
Negative samples | FP | FN |
Methods | Parameter Settings |
---|---|
CWT-AM-CNN | Conv1D (32, 3), Conv1D (64, 3), Conv1D (128, 3), activation = ‘ReLU’ |
BatchNormalization() | |
MaxPooling1D(2) | |
Dense(128, activation = ‘ReLU’), activation = ‘softmax’ | |
Optimizers = Adam, lr = 0.00001 | |
loss = ‘sparse_categorical_crossentropy’, metrics = [‘accuracy’]) | |
epochs = 500 |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Data Sets | ||||||
Data set A | 97.44% | 94.08% | 85.11% | [11 8 0] [0 77 0] [1 0 38] | 88.24% | |
Data set B | 100.00% | 100.00% | 100.00% | [19 0 0] [0 8 0] [0 0 17] | 100.00% | |
Data set C | 100% | 98.72% | 94.70% | [17 0 0 0 0 0] [0 7 0 0 0 0] [0 0 16 0 0 0] [0 0 0 84 0 0] [0 0 0 7 15 0] [0 0 0 0 0 33] | 96.18% |
Characteristic Peak Type | Emission Peak (cm−1) | Absorption Peak (cm−1) | ||
---|---|---|---|---|
Peak standard features | 2350 | 2390 | 720 | 667 |
Characteristic peak range values | 2350.5–2348 | 2377–2392 | 722–718 | 666.7–670.5 |
Methods | Parameter Settings |
---|---|
SVM | decision_function_shape = ‘ovr’, kernel = ‘rbf’ |
XGBoost | objective = ‘multi:softmax’, num_classes = num_classes |
CatBoost | loss_function = ‘MultiClass’ |
Adaboost | n_estimators = 200 |
Random Forest | n_estimators = 300 |
LightGBM | ‘objective’: ‘multiclass’, ‘num_class’: num_classes |
Neural Network | hidden_layer_sizes = (100), activation = ‘ReLU’, solver = ‘adam’, max_iter = 200 |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Classification Methods | ||||||
Feature vector + SVM | 57.04% | 33.33% | 19.01% | [0 0 0] [19 77 39] [0 0 0] | 24.21% | |
Feature vector + XGBoost | 96.30% | 96.09% | 94.36% | [18 3 0] [1 74 1] [0 0 38] | 95.14% | |
Feature vector + CatBoost | 97.04% | 96.53% | 95.80% | [18 2 0] [1 75 1] [0 0 38] | 96.14% | |
Feature vector + AdaBoost | 74.81% | 74.29% | 71.93% | [11 25 0] [8 52 1] [0 0 38] | 71.35% | |
Feature vector + Random Forest | 97.04% | 96.53% | 95.80% | [18 2 0] [1 75 1] [0 0 38] | 96.14% | |
Feature vector + LightGBM | 96.30% | 96.09% | 94.36% | [18 3 0] [1 74 1] [0 0 38] | 95.14% | |
Feature vector + Neural Networks | 86.67% | 68.42% | 92.64% | [1 0 0] [16 77 0] [2 0 39] | 66.03% |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Classification Methods | ||||||
Feature vector + SVM | 86.36% | 88.24% | 92.00% | [19 0 6] [0 8 0] [0 0 11] | 88.31% | |
Feature vector + XGBoost | 84.09% | 86.48% | 88.89% | [18 0 6] [0 8 0] [1 0 11] | 86.53% | |
Feature vector + CatBoost | 86.36% | 88.24% | 92.00% | [19 0 6] [0 8 0] [0 0 11] | 88.31% | |
Feature vector + AdaBoost | 77.27% | 80.60% | 85.19% | [18 0 9] [0 8 0] [1 0 8] | 79.93% | |
Feature vector + Random Forest | 86.36% | 88.24% | 92.00% | [19 0 6] [0 8 0] [0 0 11] | 88.31% | |
Feature vector + LightGBM | 84.09% | 86.48% | 88.89% | [18 0 6] [0 8 0] [1 0 11] | 86.53% | |
Feature vector + Neural Networks | 88.64% | 90.20% | 93.06% | [19 0 5] [0 8 0] [0 0 12] | 90.38% |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Classification Methods | ||||||
Feature vector + SVM | 59.78% | 44.15% | 47.67% | [8 0 3 0 0 0] [0 3 0 0 0 0] [9 1 12 0 0 0] [0 3 1 84 22 33] [0 0 0 0 0 0] [0 0 0 0 0 0] | 42.38% | |
Feature vector + XGBoost | 94.97% | 92.44% | 93.59% | [15 0 3 0 0 0] [0 7 0 0 0 0] [2 0 13 0 0 0] [0 0 0 83 3 0] [0 0 0 1 19 0] [0 0 0 0 0 33] | 92.95% | |
Feature vector + CatBoost | 94.41% | 90.35% | 93.52% | [15 0 2 0 0 0] [0 6 0 0 0 0] [2 0 14 0 0 0] [0 0 0 83 4 0] [0 1 0 1 18 0] [0 0 0 0 0 33] | 91.81% | |
Feature vector + AdaBoost | 79.89% | 63.66% | 71.49% | [17 5 6 0 0 0] [0 2 0 0 0 0] [0 0 10 0 0 0] [0 0 0 84 18 3] [0 0 0 0 0 0] [0 0 0 0 4 30] | 62.56% | |
Feature vector + Random Forest | 94.41% | 91.40% | 92.70% | [15 0 4 0 0 0] [0 7 0 0 0 0] [2 0 12 0 0 0] [0 0 0 83 3 0] [0 0 0 1 19 0] [0 0 0 0 0 33] | 91.91% | |
Feature vector + LightGBM | 94.41% | 90.68% | 92.40% | [14 0 2 0 0 0] [0 6 0 0 0 0] [3 0 14 0 0 0] [0 0 0 82 2 0] [0 1 0 2 20 0] [0 0 0 0 0 33] | 91.42% | |
Feature vector + Neural Networks | 84.92% | 76.79% | 76.57% | [17 0 2 0 0 0] [0 6 0 0 0 0] [0 0 12 0 0 0] [0 0 2 84 18 0] [0 1 0 0 0 0] [0 0 0 0 4 33] | 76.02% |
Methods | Parameter Settings |
---|---|
AE | Dense(encoding_dim,activation = “ ReLU “) Dense(input_dim, activation = “sigmoid”) Dense(num_classes, activation = “softmax”) epochs = 500, optimizer = Adam(lr = 0.00001),loss = ‘sparse_categorical_crossentropy’, metrics = [‘accuracy’] |
RNN | SimpleRNN(4, return_sequences = True) BatchNormalization() Dense(4, activation = ‘ReLU’) Dense(num_classes, activation = ‘softmax’) epochs = 500, optimizer = Adam(lr = 0.00001), loss = ‘sparse_categorical_crossentropy’, metrics = [‘accuracy’] |
LSTM | LSTM(8, return_sequences = True),BatchNormalization() LSTM(8), BatchNormalization() Dense(8, activation = ‘ReLU’)) Dense(num_classes, activation = ‘softmax’) epochs = 500, optimizer = Adam(lr = 0.00001), loss = ‘sparse_categorical_crossentropy’, metrics = [‘accuracy’] |
Methods | Data Set | Accuracy | Precision Score | Recall | Confusion Matrix | F1-score |
---|---|---|---|---|---|---|
AE | A | 58.52% | 52.63% | 36.84% | [2 17 0] [0 77 0] [0 39 0] | 30.79% |
B | 38.64% | 12.88% | 33.33% | [0 0 19] [0 0 8] [0 0 17] | 18.58% | |
C | 46.93% | 7.82% | 16.67% | [0 0 0 17 0 0] [0 0 0 7 0 0] [0 0 0 16 0 0] [0 0 0 84 0 0] [0 0 0 22 0 0] [0 0 0 33 0 0] | 10.65% | |
RNN | A | 38.64% | 12.88% | 33.33% | [0 0 19] [0 0 8] [0 0 17] | 18.58% |
B | 57.03% | 19.01% | 33.33% | [0 19 0] [0 77 0] [0 39 0] | 24.21% | |
C | 46.92% | 7.80% | 16.66% | [0 0 0 17 0 0] [0 0 0 7 0 0] [0 0 0 16 0 0] [0 0 0 84 0 0] [0 0 0 22 0 0] [0 0 0 33 0 0] | 10.64% | |
LSTM | A | 38.63% | 12.88% | 33.33% | [0 0 19] [0 0 8] [0 0 17] | 18.58% |
B | 57.03% | 19.01% | 33.33% | [0 19 0] [0 77 0] [0 39 0] | 24.21% | |
C | 62.57% | 48.72% | 41.91% | [4 0 13 0 0 0] [0 0 7 0 0 0] [0 0 16 0 0 0] [0 0 0 82 0 2] [0 0 0 22 0 0] [0 0 0 23 0 10] | 36.97% |
XGBoost | Accuracy | Precision | Recall | Confusion Matrix | F1-Score | Running Time |
---|---|---|---|---|---|---|
Data set A | 100.00% | 100.00% | 100.00% | [19 0 0] [0 8 0] [0 0 17] | 100.00% | 0.1359 |
Data set B | 99.26% | 99.15% | 99.57% | [19 0 0] [0 77 1] [0 0 38] | 99.35% | 0.2040 |
Data set C | 98.88% | 95.24% | 98.15% | [17 0 0 0 0 0] [0 5 0 0 0 0] [0 2 16 0 0 0] [0 0 0 84 0 0] [0 0 0 0 22 0] [0 0 0 0 0 33] | 96.24% | 0.3402 |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Data Sets | ||||||
Data set A | 94.07% | 96.86% | 85.96% | [11 8 0] [0 77 0] [0 0 39] | 89.47% | |
Data set B | 100% | 100% | 100% | [19 0 0] [0 8 0] [0 0 17] | 100% | |
Data set C | 96.09% | 98.72% | 94.70% | [17 0 0 0 0 0] [0 7 0 0 0 0] [0 0 16 0 0 0] [0 0 0 84 0 0] [0 0 0 7 15 0] [0 0 0 0 0 33] | 96.18% |
Evaluation Criterion | Accuracy | Precision Score | Recall | Confusion Matrix | F1-Score | |
---|---|---|---|---|---|---|
Data Sets | ||||||
Data set A | 92.59% | 91.70% | 84.68% | [11 8 0] [1 76 0] [1 0 38] | 87.29% | |
Data set B | 100% | 100% | 100% | [19 0 0] [0 8 0] [0 0 17] | 100% | |
Data set C | 92.18% | 94.11% | 89.94% | [17 0 0 0 0 0] [0 7 0 0 0 0] [4 0 12 0 0 0] [0 0 0 81 0 3] [0 0 0 7 15 0] [0 0 0 0 0 33] | 91.02% |
Data Set | Number of Layers | 1 | 2 | 3 | 4 | 5 | 6 | |
---|---|---|---|---|---|---|---|---|
Evaluation | ||||||||
Data set A | Accuracy | 63% | 66% | 83% | 81% | 79% | 82% | |
Training Time/s | 315.83 | 939.22 | 1332.54 | 1527.18 | 1735.24 | 2032.12 | ||
Evaluation Time/s | 0.14 | 0.18 | 0.22 | 0.33 | 0.35 | 0.32 | ||
Data set B | Accuracy | 93% | 100% | 100% | 100% | 100% | 100% | |
Training Time/s | 81.90 | 148.38 | 258.92 | 347.15 | 408.00 | 431.55 | ||
Evaluation Time/s | 0.12 | 0.13 | 0.18 | 0.25 | 0.22 | 0.25 | ||
Data set C | Accuracy | 63% | 74% | 77% | 73% | 78% | 82% | |
Training Time/s | 421.56 | 1088.86 | 1522.65 | 2014.09 | 2411.60 | 2850.66 | ||
Evaluation Time/s | 0.16 | 0.15 | 0.21 | 0.23 | 0.30 | 0.36 |
Optimizers | Prediction Accuracy | Training Time/s | Prediction Time/s |
---|---|---|---|
SGD | 93% | 1663.36 | 0.25 |
SGDM | 93% | 2074.59 | 0.23 |
Adagrad | 94% | 2133.88 | 0.24 |
RMSProp | 89% | 2194.60 | 0.27 |
Adam | 94% | 2165.09 | 0.24 |
Learning Rate | Prediction Accuracy | Training Time/s | Prediction Time/s |
---|---|---|---|
0.01 | 0.47 | 878.21 | 0.26 |
0.001 | 0.75 | 1215.80 | 0.20 |
0.0001 | 0.42 | 1246.89 | 0.21 |
0.00001 | 0.95 | 1241.00 | 0.22 |
0.000001 | 0.95 | 1221.39 | 0.21 |
Method | Running Time/s | ||
---|---|---|---|
Data Set A | Data Set B | Data Set C | |
CNN | 5 | 4 | 6 |
CNN-BN | 5 | 4 | 5 |
CWT-AM-CNN | 6 | 5 | 6 |
RNN | 980 | 243 | 1151 |
LSTM | 14 | 4 | 17 |
AE | 0.025 | 0.025 | 0.026 |
Feature vector + SVM | 0.08 | 0.01 | 0.12 |
Feature vector + XGBoost | 0.17 | 0.24 | 0.30 |
Feature vector + CatBoost | 3.09 | 2.61 | 4.74 |
Feature vector + AdaBoost | 0.30 | 0.26 | 0.39 |
Feature vector + Random Forest | 0.48 | 0.44 | 0.56 |
Feature vector + LightGBM | 0.20 | 0.17 | 0.44 |
Feature vector + Neural Networks | 0.29 | 0.31 | 0.85 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Du, S.; Han, W.; Kang, Z.; Lu, X.; Liao, Y.; Li, Z. Continuous Wavelet Transform Peak-Seeking Attention Mechanism Conventional Neural Network: A Lightweight Feature Extraction Network with Attention Mechanism Based on the Continuous Wave Transform Peak-Seeking Method for Aero-Engine Hot Jet Fourier Transform Infrared Classification. Remote Sens. 2024, 16, 3097. https://doi.org/10.3390/rs16163097
Du S, Han W, Kang Z, Lu X, Liao Y, Li Z. Continuous Wavelet Transform Peak-Seeking Attention Mechanism Conventional Neural Network: A Lightweight Feature Extraction Network with Attention Mechanism Based on the Continuous Wave Transform Peak-Seeking Method for Aero-Engine Hot Jet Fourier Transform Infrared Classification. Remote Sensing. 2024; 16(16):3097. https://doi.org/10.3390/rs16163097
Chicago/Turabian StyleDu, Shuhan, Wei Han, Zhenping Kang, Xiangning Lu, Yurong Liao, and Zhaoming Li. 2024. "Continuous Wavelet Transform Peak-Seeking Attention Mechanism Conventional Neural Network: A Lightweight Feature Extraction Network with Attention Mechanism Based on the Continuous Wave Transform Peak-Seeking Method for Aero-Engine Hot Jet Fourier Transform Infrared Classification" Remote Sensing 16, no. 16: 3097. https://doi.org/10.3390/rs16163097
APA StyleDu, S., Han, W., Kang, Z., Lu, X., Liao, Y., & Li, Z. (2024). Continuous Wavelet Transform Peak-Seeking Attention Mechanism Conventional Neural Network: A Lightweight Feature Extraction Network with Attention Mechanism Based on the Continuous Wave Transform Peak-Seeking Method for Aero-Engine Hot Jet Fourier Transform Infrared Classification. Remote Sensing, 16(16), 3097. https://doi.org/10.3390/rs16163097