Submitted:
28 September 2023
Posted:
29 September 2023
You are already at the latest version
Abstract

Keywords:
1. Introduction
- Section 2 - Analyzing the Effects of Vibration on Radiator Integrity: In this section, an in-depth analysis of frequency response signals obtained from a real operational facility is undertaken to simulate the impact of vibrations on the structural integrity of radiators. A randomized durability vibration bench test is conducted, employing state-of-the-art acceleration sensors for real-time signal capture. The rationale behind the selection of GMM and LSTM autoencoders for fault diagnosis is elaborated upon.
- Section 3 - Feature Engineering and Data Refinement: This section is dedicated to the critical task of feature engineering. Here, the focus is on extracting time-domain statistical features from raw data, refining them through PCA. The insights gained from PCA guide the subsequent training of the GMM. Findings related to fault diagnosis for stages 2 and 4, along with an in-depth exploration of anomaly detection for the unlabeled stage 3, are presented.
- Section 4 - Leveraging LSTM Autoencoders: In this section, attention turns to the practical implementation of LSTM autoencoders for fault diagnosis and anomaly detection. The advantages of this approach are highlighted, showcasing its superiority over traditional methods, especially when dealing with time-series data.
- Section 5 - Evaluation and Technical Contributions: The final section provides a comprehensive evaluation of the proposed combination of GMM and LSTM autoencoders. Here, meticulous detailing of the technical contributions of the methodology to the field of machine learning for fault diagnosis is presented. Emphasis is placed on its capability for precise fault detection and anomaly prediction. The practical implications of this approach for real-world applications, such as enhancing equipment uptime, preventing critical failures, and minimizing maintenance costs, are explored.
2. Radiator dataset and research objective
2.1. Random durability vibration bench test and dataset
2.2. Methods and evluation metrics
3. GMM-based fault diagnosis
3.1. Feature engineering and PCA of radiator dataset
3.2. Fault Diagnosis and anomaly detection result using GMM
- Train the Gaussian Mixture Model.
- Establish a threshold value based on the absolute log-likelihood distribution derived from stage 1.
- In stages 2–4, data points whose absolute log-likelihood values exceed the predetermined threshold are flagged as faulty.
- Feature Engineering: Utilized a ten-second window combined with sliding window augmentation.
- Gaussian Distributions: Set to three, based on both the BIC and AIC analysis and expert understanding of fault diagnosis.
4. LSTM Autoencoder-based fault diagnosis
4.1. Rationale for LSTM Autoencoder selection
4.2. Fault diagnosis and anomaly detection result using LSTM Autoencoder
- Feature engineering method: A 10-second window with sliding augmentation.
- LSTM AE structure: Configured as 100/1/100.
- Dropout and L2 Regularization: Not applied.
5. Discussion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Tsui, K.L.; Chen, N.; Zhou, Q.; Hai, Y.; Wang, W. Prognostics and health management: A review on data driven approaches. Mathematical Problems in Engineering 2015. [CrossRef]
- Lee, J.; Wu, F.; Zhao, W.; Ghaffari, M.; Liao, L.; Siegel, D. Prognostics and health management design for rotary machinery systems—Reviews, methodology and applications. Mechanical systems and signal processing 2014, 42, 314–334. [Google Scholar] [CrossRef]
- Meng, H.; Li, Y.F. A review on prognostics and health management (PHM) methods of lithium-ion batteries. Renewable and Sustainable Energy Reviews. 2019, 116, 109405. [Google Scholar] [CrossRef]
- Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mechanical Systems and Signal Processing. 2020, 138, 106587. [Google Scholar] [CrossRef]
- Cao, S.; Hu, Z.; Luo, X.; Wang, H. Research on fault diagnosis technology of centrifugal pump blade crack based on PCA and GMM. Measurement. 2021, 173, 108558. [Google Scholar] [CrossRef]
- Chen, Z.; Peng, Z.; Zou, X.; Sun, H. Deep learning based anomaly detection for muti-dimensional time series: A survey. In China Cyber Security Annual Conference. July 2021; pp. 71–92.
- Zhao, H.; Sun, S.; Jin, B. Sequential fault diagnosis based on LSTM neural network. Ieee Access. 2018, 6, 12929–12939. [Google Scholar] [CrossRef]
- Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural computation. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
- Janiesch, C.; Zschech, P.; Heinrich, K. Machine learning and deep learning. Electronic Markets. 2021, 31, 685–695. [Google Scholar] [CrossRef]
- Kamat, P.; Sugandhi, R. Anomaly detection for predictive maintenance in industry 4.0-A survey. In E3S web of conferences. 2020; Vol. 170, p. 02007.
- Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv preprint arXiv:1901.03407. 2019. arXiv:1901.03407. 2019.
- Dara, S.; Tumma, P. Feature extraction by using deep learning: A survey. In 2018 Second international conference on electronics, communication and aerospace technology (ICECA). March 2018; pp. 1795–1801.
- Bergman, T. L.; Lavine, A. S.; Incropera, F. P.; DeWitt, D. P. Introduction to heat transfer, 5th ed. John Wiley & Sons, pp.632-633, 2005.
- Klyatis, L.M. Accelerated reliability and durability testing technology, Vol. 70, John Wiley & Sons, 2012.
- Elsayed, E.A. Overview of reliability testing. IEEE Transactions on Reliability. 2012, 61, 282–291. [Google Scholar] [CrossRef]
- Kapur, K. C.; Pecht, M. Reliability engineering, Vol. 86, John Wiley & Sons, 2014.
- O'Connor, P.; Kleyner, A. Practical reliability engineering. John Wiley & Sons, 2012.
- Zio, E. Reliability engineering: Old problems and new challenges. Reliability engineering & system safety. 2009, 94, 125–141. [Google Scholar]
- IEC. IEC 60068-2-6: Environmental testing - Part 2-6: Tests - Test Fc: Vibration (sinusoidal). IEC, 2011.
- MIL-STP-810. Environmental Engineering Considerations and Laboratory Tests. no date.
- Breiman, L. Random forests. Machine learning. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems 2017, 30. [Google Scholar]
- Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and regression trees. CRC press. 1984.
- Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, August 2016; pp. 785–794.
- Noble, W.S. What is a support vector machine? Nature biotechnology. 2006, 24, 1565–1567. [Google Scholar] [CrossRef]
- Dreiseitl, S.; Ohno-Machado, L. Logistic regression and artificial neural network classification models: a methodology review. Journal of biomedical informatics. 2002, 35, 352–359. [Google Scholar] [CrossRef]
- Reynolds, D.A. Gaussian mixture models. Encyclopedia of biometrics. 2009, 741, 659–663. [Google Scholar]
- Rasmussen, C. The infinite Gaussian mixture model. Advances in neural information processing systems. 1999, 12. [Google Scholar]
- Nguyen, H.D.; Tran, K.P.; Thomassey, S.; Hamad, M. Forecasting and Anomaly Detection approaches using LSTM and LSTM Autoencoder techniques with the applications in supply chain management. International Journal of Information Management. 2021, 57, 102282. [Google Scholar] [CrossRef]
- Said Elsayed, M.; Le-Khac, N.A.; Dev, S.; Jurcut, A.D. Network anomaly detection using LSTM based autoencoder. In Proceedings of the 16th ACM Symposium on QoS and Security for Wireless and Mobile Networks. 2020, pp. 37-45.
- Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. Journal of the Royal Statistical Society Series B: Statistical Methodology. 1999, 61, 611–622. [Google Scholar] [CrossRef]
- Holland, S.M. Principal components analysis (PCA). Department of Geology, University of Georgia, Athens, GA. 2008, 30602-2501.
- Xuan, G.; Zhang, W.; Chai, P. EM algorithms of Gaussian mixture model and hidden Markov model. In Proceedings 2001 international conference on image processing (Cat. No. 01CH37205), October 2001; Vol. 1, pp. 145–148. IEEE.
- Kuha, J. AIC and BIC: Comparisons of assumptions and performance. Sociological methods & research. 2004, 33, 188–229. [Google Scholar]
- Aurélien, G. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd ed. O'Reilly Media, Inc, 2019.
- Graves, A.; Graves, A. Long short-term memory. Supervised sequence labelling with recurrent neural networks. 2012, 37–45. [Google Scholar]
- Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks. 1994, 5, 157–166. [Google Scholar] [CrossRef]
- Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training recurrent neural networks. In International conference on machine learning (PMLR). May 2013; pp.1310-1318.
- Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems. 1998, 6, 107–116. [Google Scholar] [CrossRef]
- Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural computation. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
- Bao, W.; Yue, J.; Rao, Y. A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PloS one. 2017, 12, e0180944. [Google Scholar] [CrossRef]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research. 2014, 15, 1929–1958. [Google Scholar]
- Cortes, C.; Mohri, M.; Rostamizadeh, A. L2 regularization for learning kernels. arXiv2012, arXiv preprint arXiv:1205.2653.
















| State | Label | Time | Observation result |
|---|---|---|---|
| Stage 1 | Normal 1 | 0–100 min. | No coolant leakage in 100 minutes |
| Stage 2 | Normal 2 | 100–300 min. | No coolant leakage in 300 minutes |
| Stage 3 | Unknown | 300–400 min. | Estimate a coolant leakage between 300 and 400 minutes. |
| Stage 4 | Abnormal | 400–808 min. | Coolant leakage at 400 minutes |
| Window size | The number of Gaussian components | ||||
| 1 | 2 | 3 | 4 | 5 | |
| 1sec | 0.6286 | 0.7018 | 0.7094 | 0.7095 | 0.7076 |
| 2sec | 0.6182 | 0.7536 | 0.7649 | 0.7630 | 0.7609 |
| 3sec | 0.6117 | 0.7872 | 0.8016 | 0.7971 | 0.7979 |
| 4sec | 0.6070 | 0.8183 | 0.8306 | 0.8291 | 0.8293 |
| 5sec | 0.6125 | 0.8384 | 0.8483 | 0.8532 | 0.8517 |
| 6sec | 0.6141 | 0.8572 | 0.8674 | 0.8704 | 0.8717 |
| 7sec | 0.6153 | 0.8696 | 0.8828 | 0.8822 | 0.8823 |
| 8sec | 0.6240 | 0.8642 | 0.8833 | 0.8852 | 0.8871 |
| 9sec | 0.6275 | 0.8724 | 0.8915 | 0.8904 | 0.8899 |
| 10sec | 0.6226 | 0.8873 | 0.9000 | 0.8995 | 0.9009 |
| 10sec(sliding) | 0.6310 | 0.8873 | 0.8998 | 0.8985 | 0.8998 |
| Variables | Window size | Hyperparameters | AUC | ||
| LSTM AE structure | Dropoutrate | L2 Regularization | |||
| Window size | 1sec | 100/1/100 | None | None | 0.8579 |
| 2sec | 100/1/100 | None | None | 0.9151 | |
| 3sec | 100/1/100 | None | None | 0.9631 | |
| 4sec | 100/1/100 | None | None | 0.9639 | |
| 5sec | 100/1/100 | None | None | 0.9572 | |
| 6sec | 100/1/100 | None | None | 0.9656 | |
| 7sec | 100/1/100 | None | None | 0.9672 | |
| 8sec | 100/1/100 | None | None | 0.9647 | |
| 9sec | 100/1/100 | None | None | 0.9683 | |
| 10sec | 100/1/100 | None | None | 0.9728 | |
| 10sec(sliding) | 100/1/100 | None | None | 0.9893 | |
| The number of nodes | 10sec(sliding) | 200/1/200 | None | None | 0.9902 |
| 10sec(sliding) | 300/1/300 | None | None | 0.9912 | |
| Dropout rate | 10sec(sliding) | 100/1/100 | 0.3 | None | 0.9868 |
| 10sec(sliding) | 100/1/100 | 0.5 | None | 0.9845 | |
| 10sec(sliding) | 100/1/100 | 0.7 | None | 0.9782 | |
| L2 Regularization | 10sec(sliding) | 100/1/100 | None | 0.1 | 0.9729 |
| 10sec(sliding) | 100/1/100 | None | 0.01 | 0.9853 | |
| 10sec(sliding) | 100/1/100 | None | 0.001 | 0.9702 | |
| Dropout rate +L2 Regularization | 10sec(sliding) | 100/1/100 | 0.5 | 0.01 | 0.9751 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).