Submitted:
22 January 2024
Posted:
23 January 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Literature Review
2.1. Product Quality Control
2.1.1. Defect Detection
2.1.2. Defect Prediction
2.2. Predictive Maintenance
2.2.1. Failure Prediction
2.2.2. Remaining Useful Life Prediction
2.3. AI Models for Product Quality Control and Predictive Maintenance
2.3.1. Machine Learning Methods
- Support Vector Machine: It is an ML technique commonly employed for classification and regression tasks [34]. In classification, SVM creates hyperplanes to effectively separate different classes using support vectors [34]. In regression, the objective is to identify a function that closely matches data points within a defined margin [34]. SVM was employed in various studies including predicting part quality in plastic injection [5], identifying defects in laser additive manufacturing [35], predicting tool wear during milling operations [24], and estimating the RUL of machine components [36].
- K-Nearest Neighbors: [37] is a machine learning technique used for classification and regression tasks. To classify a new data point, the algorithm looks at the K closest data points from the training set and assigns the majority class among those neighbors to the new data point [37]. For regression, it predicts the value based on the average of the K nearest data points values [37]. It does not require training but its performance can be affected by the chosen K value and distance metric. in [38] KNN was used for fabric defect detection based on features extracted from thermal camera images . In the context of additive manufacturing quality control, KNN demonstrated effective porosity prediction, as presented in [39]. The effectiveness of KNN was also highlighted in [40], a comparative study on predictive maintenance.
- Naive Bayes: [41] is a probabilistic machine learning algorithm that relies on Bayes’ theorem. It assumes feature independence given the class label. The algorithm calculates probabilities for a data point belonging to each class, and predicts the class with the highest probability. Naive Bayes can also be used for regression tasks [42]. The Naive Bayes algorithm combined with Particle Swarm Optimization (PSO) [43] in [44] effectively detected product defects. In [45], a Naive Bayes approach using vibration signals successfully identified specific bearing faults.
- Regression (Linear and Logistic): Linear regression [46] is an ML algorithm that searches for a linear relationship between input features and the target variable by fitting a straight line to the data points. The effectiveness of this algorithm depends on the linearity assumption of the data. Logistic regression [47] on the other hand, is used for classification by calculating the probability of belonging to a class using a logistic function that produces values between 0 and 1. In [48], a logistic regression model is proposed for predicting product quality in the rolling process. Linear regression is applied in [49] to forecast machine failure in a turbine generator for maintenance scheduling in oil and gas platforms. The study in [50] employs Multiple Linear Regression (MLR) [51] to estimate the RUL of bearings based on vibration data.
- Decision Tree: [52] is an ML technique used for both classification and regression tasks. It constructs a model in the form of a tree where each internal node represents a feature and each leaf node represents a class or a predicted value. It recursively splits the data based on the feature that provides the best separation at each node up to a certain stopping criterion. In [53], the authors used a J48 decision tree model to predict part quality in the injection molding process. The study in [54] employed Decision Trees in combination with a genetic algorithm to predict the RUL of an aircraft air conditioning system.
- Ensemble Learning: [55] is a category of machine learning techniques that combines predictions from multiple models to predict the target variable. Predictions are often combined using voting for classification tasks and averaging for numerical value prediction. There are different ensemble methods such as bagging, XGBoost [56], Random Forest [57], and Gradient Boosting Machine [58]. XGBoost successfully predicts manufacturing defects in [15,18]. In [59], a Gradient Boosting model is suggested for predicting steel product quality. In predictive maintenance, Random Forest models are used to predict the RUL of machines in [26,60].
2.3.2. Deep Learning Models
- Multi-Layer Perceptron: MLP [63] models are the most well-known deep learning models. MLP consists of interconnected layers of neurons. Neurons in intermediate layers assign weights to inputs from the previous layer, sum them with a bias, and apply an activation function. The last layer gives the output value depending on the nature of the problem, whether it is a classification or regression task. The study in [19] applied MLP in an online defect detection system for the injection molding process. In [31], MLP was employed to predict the RUL of aircraft turbofan engines. MLP model demonstrated strong performance in predicting machine component failures in [21].
- Convolutional Neural network: Convolutional Neural Network (CNN) [64] is a widely employed DL method for image processing such as image classification, object detection, and image segmentation. Its key component, the convolutional layer, uses filters to detect patterns in input images through convolutions. CNNs can learn features automatically without the need for manual features extraction. It is common in the literature to use Transfer Learning, which involves taking a pre-trained CNN model on one task and adapting it to a new, similar task. Many pre-trained model architectures have been proposed [65]. In [12] a CNN was used for background texture-based defect detection. In [13], a modified YOLO [66] transformed into a fully convolutional model, was introduced for real-time surface defect detection in steel strips. The work in [14] reviews and presents other examples of defect detection in images of products. Regarding predictive maintenance, [29,67] applied CNNs to predict RUL.
- Recurrent Neural Network: an RNN [68] is a type of network introduced to process sequential data such as time series data or text. RNNs take an ordered sequence of data as input to predict one or more output values. They are designed to extract important information from sequential data and use it for prediction. There are different types of RNN architectures including vanilla RNN [68], Long Short-Term Memory (LSTM) networks [69], and Gated Recurrent Units (GRU)[70]. [71] proposed an approach to predict the RUL of aircraft engines using a basic LSTM model. The study in [22] demonstrates the effectiveness of a system based on a Gated Recurrent Unit (GRU) model for predicting tool wear, gearbox faults, and bearing faults.
- Generative Adversarial Network: a GAN [72] is a type of deep artificial network that enables the generation of synthetic data from a given real dataset. This model primarily consists of two components: a generator and a discriminator, each with a specific role during the model training process. The generator transforms a random noise vector into synthetic data that resembles the original dataset. On the other hand, the discriminator is used to differentiate between the synthetic and real sample, by classifying them accordingly. A steel surface defect detection method utilizing GANs was introduced in [73]. Works such as [74,75,76] propose approaches using GANs for predicting the RUL of machines using data extracted from multiple sensors.
- Autoencoder: It [77] is an artificial neural network consisting of two main components: the encoder and the decoder. The encoder processes input data and transforms it into a lower-dimensional encoded representation within a latent space. The decoder performs the inverse operation by taking the encoded representation and decoding it to reconstruct the original data. Its role is to recreate a version of the initial input as close as possible. The main objective of the autoencoder is to minimize the difference between the input data and the reconstructed data. An autoencoder-based model is used in [17] for quality prediction in the injection molding process. The study in [78] proposed a deep learning model composed of a Variational Autoencoder (VAE) [79] and a Recurrent Neural Network (RNN) for predicting the RUL of machines. The VAE is used to reduce the dimensionality of the data and extract features from sensor data, while the RNN is used for RUL prediction.
- Transformer: It [80] was initially created for natural language processing tasks such as language translation and text summarization. Similar to Autoencoders, Transformers mainly consist of two parts: an encoder and a decoder, both composed of multiple layers of self-attention and feed-forward neural networks. The transformer is designed to learn to produce outputs by focusing only on relevant information extracted from the input data using the attention mechanism. Recently, several studies have explored the application of Transformers in other tasks, such as image processing [81] and time series data analysis [82]. Transformers enable parallel processing of data, overcoming the sequential processing limitations of RNNs [83]. Studies like [84] [85] have used Transformers for surface defect detection. In addition, a system employing the Transformer was proposed for predicting RUL of Li-ion batteries in [86].
2.4. Conclusion and Discussion
- Data Imbalance: It is a prevalent challenge in applying AI to predictive maintenance and product quality control. Machine failure instances and defective product examples, are often rare compared to normal cases, resulting in imbalanced datasets. AI models are designed to minimize overall error rates and may perform well on normal cases, but struggle with predicting machine failures and identifying defects. Various techniques [87] have been employed to address this issue, but data imbalance remains a significant challenge for AI applications in many fields.
- Explainability and Interpretability: Some traditional machine learning models, such as decision trees, are by default explainable. However, more complex models like ensemble and deep learning models are not inherently explainable. Deep learning models are often viewed as complex black-box models, presenting a challenge in understanding decision-making processes and the internal logic. This opacity may cause regulatory compliance issues regarding the accountability and transparency of their decisions. Explainability and interpretability techniques [88] should be applied to establish confidence in the decision of the model. Furthermore, these techniques can help identify the root causes of a product defect or a machine failure. A better understanding of these causes could aid in optimizing the production process by adjusting parameters related to these incidents. Most of the existing studies on product quality control and predictive maintenance did not address the explainability of the proposed solutions.
- Real-time Detection Instead of Prediction: Most of the works predict equipment failure at the time of prediction rather than in future states. These approaches perform real-time failure detection rather than making a prediction. A reliable predictive maintenance system should be capable of predicting future failures based on the current state, equipment history, and other information related to the environmental conditions of the equipment.
- Domain Dependency and the Need for Industry-Specific Datasets: Domain dependence is one of the main challenges of applying AI for predictive maintenance and product quality control. It can be observed in the literature that the field of predictive maintenance is not only relevant to the manufacturing industry but also to other sectors. Public datasets like C-MAPSS [32] and PHM08 [89] have been used to evaluate predictive maintenance proposals, but these datasets are not specifically related to the manufacturing industry. It is challenging to ensure that the proposed approaches will be effective when applied to real-world cases in industry. Transfer learning [90] techniques allow adapting a model built in one domain to another, but the input data of the model could not be the same. Domain dependence is a persistent issue in the application of AI in industry, and this can impact the performance of the model. A model may perform well on benchmarking data but exhibit very poor performance in a real-world industrial application.
- Overlooking Component Interactions: The degradation of one component may be linked to other components, further complicating the identification of failure causes. Failing to account for these interactions can lead to inaccurate failure predictions. It is crucial to monitor the overall machine state and consider interactions between components in the design of AI models for predictive maintenance.
- Single-quality Criterion Consideration: One limitation of existing works on defect prediction is that they typically focus on a single quality criterion. To establish an effective quality control system, the dataset should include data from partial quality inspections that address all relevant criteria. Sources of defects may differ across various quality criteria, introducing an additional challenge in the application of AI for predicting manufacturing defects.
3. Experimental Study
3.1. Evaluation metrics
- Precision: measures the proportion of correctly predicted instances among those predicted as positive [91].
- Recall: evaluates the proportion of correctly predicted positive instances among all truly positive instances [91].
- F-score: combines precision and recall into a single measure allowing the assessment of balance between the two [91].
3.2. Quality Prediction in Plastic Injection
3.3. Machine Component Failure Prediction
- Telemetry: includes measurements of machine pressure, vibration, rotation, and voltage.
- Errors: log of recorded machine errors.
- Machine: provides machine characteristics such as age and model.
- Maintenance: contains the history of all machine component replacements.
- Failures: information on the history of failed component replacements.
4. AI Solution Development for Product Quality Control and Predictive Maintenance
4.1. Data Collection and Feature Engineering
4.2. Data Preprocessing
4.3. Data Analysis
4.4. Model Development
4.5. Model Explanation
4.6. Model Deployment
5. Conclusions
Acknowledgments
Abbreviations
| AI | Artificial Intelligence |
| IoT | Internet of Things |
| SPC | Statistical Process Control |
| ML | Machine Learning |
| DL | Deep Learning |
| QC | QUality Control |
| CV | Computer Vision |
| RUL | Remaining Useful Life |
| C-MAPSS | Commercial Modular Aero-Propulsion System Simulation |
| SVM | Support Vector Machine |
| KNN | K Nearest Neighbors |
| MLR | Multiple Linear Regression |
| MLP | Multilayer Perceptron |
| CNN | Convolutional Neural Network |
| RNN | Recurrent Neural Network |
| LSTM | Long Short-term Memory |
| GRU | Gated Recurrent Unit |
| GAN | Generative Adversarial Network |
| VAE | Variational Autoencoder |
References
- Moeuf, A.; Lamouri, S.; Pellerin, R.; Tamayo-Giraldo, S.; Tobon-Valencia, E.; Eburdy, R. Identification of critical success factors, risks and opportunities of Industry 4.0 in SMEs. International Journal of Production Research 2020, 58, 1384–1400. [Google Scholar] [CrossRef]
- Kotsiopoulos, T.; Sarigiannidis, P.; Ioannidis, D.; Tzovaras, D. Machine learning and deep learning in smart manufacturing: The smart grid paradigm. Computer Science Review 2021, 40, 100341. [Google Scholar] [CrossRef]
- Dash, R.; McMurtrey, M.; Rebman, C.; Kar, U.K. Application of artificial intelligence in automation of supply chain management. Journal of Strategic Innovation and Sustainability 2019, 14, 43–53. [Google Scholar]
- Laayati, O.; Bouzi, M.; Chebak, A. Smart energy management: Energy consumption metering, monitoring and prediction for mining industry. 2020 IEEE 2nd International Conference on Electronics, Control, Optimization and Computer Science (ICECOCS). IEEE, 2020, pp. 1–5.
- Silva, B.; Sousa, J.; Alenya, G. Machine learning methods for quality prediction in thermoplastics injection molding. 2021 International Conference on Electrical, Computer and Energy Technologies (ICECET). IEEE, 2021, pp. 1–6.
- Abidi, M.H.; Mohammed, M.K.; Alkhalefah, H. Predictive maintenance planning for industry 4.0 using machine learning for sustainable manufacturing. Sustainability 2022, 14, 3387. [Google Scholar] [CrossRef]
- Colantonio, L.; Equeter, L.; Dehombreux, P.; Ducobu, F. A systematic literature review of cutting tool wear monitoring in turning by using artificial intelligence techniques. Machines 2021, 9, 351. [Google Scholar] [CrossRef]
- Oakland, J.; Oakland, J.S. Statistical process control; Routledge, 2007. [Google Scholar]
- Dutoit, C.; Dehombreux, P.; Lorphèvre, E.R.; Equeter, L. Statistical process control and maintenance policies for continuous production systems subjected to different failure impact models: literature review. Procedia CIRP 2019, 86, 55–60. [Google Scholar] [CrossRef]
- Ramezani, J.; Jassbi, J. Quality 4.0 in action: smart hybrid fault diagnosis system in plaster production. Processes 2020, 8, 634. [Google Scholar] [CrossRef]
- Wang, K.J.; Fan-Jiang, H.; Lee, Y.X. A multiple-stage defect detection model by convolutional neural network. Computers & Industrial Engineering 2022, 168, 108096. [Google Scholar]
- Wang, T.; Chen, Y.; Qiao, M.; Snoussi, H. A fast and robust convolutional neural network-based defect detection model in product quality control. The International Journal of Advanced Manufacturing Technology 2018, 94, 3465–3471. [Google Scholar] [CrossRef]
- Li, J.; Su, Z.; Geng, J.; Yin, Y. Real-time detection of steel strip surface defects based on improved yolo detection network. IFAC-PapersOnLine 2018, 51, 76–81. [Google Scholar] [CrossRef]
- Chen, Y.; Ding, Y.; Zhao, F.; Zhang, E.; Wu, Z.; Shao, L. Surface defect detection methods for industrial products: A review. Applied Sciences 2021, 11, 7657. [Google Scholar] [CrossRef]
- Uyan, T.Ç.; Otto, K.; Silva, M.S.; Vilaça, P.; Armakan, E. Industry 4.0 foundry data management and supervised machine learning in low-pressure die casting quality improvement. International Journal of Metalcasting 2022, 1–16. [Google Scholar] [CrossRef]
- García, V.; Sánchez, J.S.; Rodríguez-Picón, L.A.; Méndez-González, L.C.; Ochoa-Domínguez, H.d.J. Using regression models for predicting the product quality in a tubing extrusion process. Journal of Intelligent Manufacturing 2019, 30, 2535–2544. [Google Scholar] [CrossRef]
- Jung, H.; Jeon, J.; Choi, D.; Park, J.Y. Application of machine learning techniques in injection molding quality prediction: Implications on sustainable manufacturing industry. Sustainability 2021, 13, 4120. [Google Scholar] [CrossRef]
- Obregon, J.; Hong, J.; Jung, J.Y. Rule-based explanations based on ensemble machine learning for detecting sink mark defects in the injection moulding process. Journal of Manufacturing Systems 2021, 60, 392–405. [Google Scholar] [CrossRef]
- Chen, J.C.; Guo, G.; Wang, W.N. Artificial neural network-based online defect detection system with in-mold temperature and pressure sensors for high precision injection molding. The International Journal of Advanced Manufacturing Technology 2020, 110, 2023–2033. [Google Scholar] [CrossRef]
- Mobley, R.K. An introduction to predictive maintenance; Elsevier, 2002. [Google Scholar]
- Cardoso, D.; Ferreira, L. Application of predictive maintenance concepts using artificial intelligence tools. Applied Sciences 2020, 11, 18. [Google Scholar] [CrossRef]
- Zhao, R.; Wang, D.; Yan, R.; Mao, K.; Shen, F.; Wang, J. Machine health monitoring using local feature-based gated recurrent unit networks. IEEE Transactions on Industrial Electronics 2017, 65, 1539–1548. [Google Scholar] [CrossRef]
- Janssens, O.; Slavkovikj, V.; Vervisch, B.; Stockman, K.; Loccufier, M.; Verstockt, S.; Van de Walle, R.; Van Hoecke, S. Convolutional neural network based fault detection for rotating machinery. Journal of Sound and Vibration 377. [CrossRef]
- Lee, W.J.; Wu, H.; Yun, H.; Kim, H.; Jun, M.B.; Sutherland, J.W. Predictive maintenance of machine tool systems using artificial intelligence techniques applied to machine condition data. Procedia Cirp 2019, 80, 506–511. [Google Scholar] [CrossRef]
- Wang, B.; Lei, Y.; Yan, T.; Li, N.; Guo, L. Recurrent convolutional neural network: A new framework for remaining useful life prediction of machinery. Neurocomputing 2020, 379, 117–129. [Google Scholar] [CrossRef]
- ASLANTAŞ, G.; ÖZSARAÇ, M.; RUMELLİ, M.; ALAYGUT, T.; BAKIRLI, G.; BIRANT, D. Prediction of Remaining Useful Life for Plastic Injection Molding Machines Using Artificial Intelligence Methods. Journal of Artificial Intelligence and Data Science 2, 8–15.
- CEYLAN, U.; Yakup, G. Siamese Inception Time Network for Remaining Useful Life Estimation. Journal of Artificial Intelligence and Data Science 1, 165–175.
- Ellefsen, A.L.; Ushakov, S.; Æsøy, V.; Zhang, H. Validation of data-driven labeling approaches using a novel deep network structure for remaining useful life predictions. IEEE Access 2019, 7, 71563–71575. [Google Scholar] [CrossRef]
- Li, H.; Zhao, W.; Zhang, Y.; Zio, E. Remaining useful life prediction using multi-scale deep convolutional neural network. Applied Soft Computing 2020, 89, 106113. [Google Scholar] [CrossRef]
- Chen, Z.; Wu, M.; Zhao, R.; Guretno, F.; Yan, R.; Li, X. Machine remaining useful life prediction via an attention-based deep learning approach. IEEE Transactions on Industrial Electronics 2020, 68, 2521–2531. [Google Scholar] [CrossRef]
- Kang, Z.; Catal, C.; Tekinerdogan, B. Remaining useful life (RUL) prediction of equipment in production lines using artificial neural networks. Sensors 2021, 21, 932. [Google Scholar] [CrossRef]
- Ramasso, E.; Saxena, A. Performance Benchmarking and Analysis of Prognostic Methods for CMAPSS Datasets. International Journal of Prognostics and Health Management 2014, 5, 1–15. [Google Scholar] [CrossRef]
- Nozari, H.; Sadeghi, M.E.; others. Artificial intelligence and Machine Learning for Real-world problems (A survey). International Journal of Innovation in Engineering 2021, 1, 38–47. [Google Scholar]
- Gunn, S.R.; others. Support vector machines for classification and regression. ISIS technical report 1998, 14, 5–16. [Google Scholar]
- Liu, T.; Huang, L.; Chen, B. Real-time defect detection of laser additive manufacturing based on support vector machine. Journal of Physics: Conference Series. IOP Publishing, 2019, Vol. 1213, p. 052043. [CrossRef]
- Huang, H.Z.; Wang, H.K.; Li, Y.F.; Zhang, L.; Liu, Z. Support vector machine based estimation of remaining useful life: current research status and future trends. Journal of Mechanical Science and Technology 2015, 29, 151–163. [Google Scholar] [CrossRef]
- Kramer, O.; Kramer, O. K-nearest neighbors. Dimensionality reduction with unsupervised nearest neighbors 2013, pp. 13–23.
- Yıldız, K.; Buldu, A.; Demetgul, M. A thermal-based defect classification method in textile fabrics with K-nearest neighbor algorithm. Journal of Industrial Textiles 2016, 45, 780–795. [Google Scholar] [CrossRef]
- Khanzadeh, M.; Chowdhury, S.; Marufuzzaman, M.; Tschopp, M.A.; Bian, L. Porosity prediction: Supervised-learning of thermal history for direct laser deposition. Journal of manufacturing systems 2018, 47, 69–82. [Google Scholar] [CrossRef]
- Ouadah, A.; Zemmouchi-Ghomari, L.; Salhi, N. Selecting an appropriate supervised machine learning algorithm for predictive maintenance. The International Journal of Advanced Manufacturing Technology 2022, 119, 4277–4301. [Google Scholar] [CrossRef]
- Rish, I.; others. An empirical study of the naive Bayes classifier. IJCAI 2001 workshop on empirical methods in artificial intelligence, 2001, Vol. 3, pp. 41–46.
- Frank, E.; Trigg, L.; Holmes, G.; Witten, I.H. Naive Bayes for regression. Machine Learning 2000, 41, 5–25. [Google Scholar] [CrossRef]
- Shami, T.M.; El-Saleh, A.A.; Alswaitti, M.; Al-Tashi, Q.; Summakieh, M.A.; Mirjalili, S. Particle swarm optimization: A comprehensive survey. IEEE Access 2022, 10, 10031–10061. [Google Scholar] [CrossRef]
- Romli, I.; Pardamean, T.; Butsianto, S.; Wiyatno, T.N.; bin Mohamad, E. Naive bayes algorithm implementation based on particle swarm optimization in analyzing the defect product. Journal of Physics: Conference Series. IOP Publishing, 2021, Vol. 1845, p. 012020. [CrossRef]
- Zhang, N.; Wu, L.; Yang, J.; Guan, Y. Naive bayes bearing fault diagnosis based on enhanced independence of data. Sensors 2018, 18, 463. [Google Scholar] [CrossRef]
- Montgomery, D.C.; Peck, E.A.; Vining, G.G. Introduction to linear regression analysis; John Wiley & Sons, 2021. [Google Scholar]
- Stoltzfus, J.C. Logistic regression: a brief primer. Academic emergency medicine 2011, 18, 1099–1104. [Google Scholar] [CrossRef]
- Jin, R.; Li, J.; Shi, J. Quality prediction and control in rolling processes using logistic regression. Transactions of NAMRI/SME 2007, 35, 113–120. [Google Scholar]
- Akhir, E.A.P.; Ayuni, N. Predictive analytics of machine failure using linear regression on KNIME platform. 2021 5th International Conference on Artificial Intelligence and Virtual Reality (AIVR), 2021, pp. 59–64.
- Zhao, M.; Tang, B.; Tan, Q. Bearing remaining useful life estimation based on time–frequency representation and supervised dimensionality reduction. Measurement 2016, 86, 41–55. [Google Scholar] [CrossRef]
- Uyanık, G.K.; Güler, N. A study on multiple linear regression analysis. Procedia-Social and Behavioral Sciences 2013, 106, 234–240. [Google Scholar] [CrossRef]
- Kotsiantis, S.B. Decision trees: a recent overview. Artificial Intelligence Review 2013, 39, 261–283. [Google Scholar] [CrossRef]
- Ogorodnyk, O.; Lyngstad, O.V.; Larsen, M.; Wang, K.; Martinsen, K. Application of machine learning methods for prediction of parts quality in thermoplastics injection molding. International workshop of advanced manufacturing and automation. Springer, 2018, pp. 237–244.
- Gerdes, M.; Galar, D.; Scholz, D. Genetic algorithms and decision trees for condition monitoring and prognosis of A320 aircraft air conditioning. Insight-Non-Destructive Testing and Condition Monitoring 2017, 59, 424–433. [Google Scholar] [CrossRef]
- Dietterich, T.G. Ensemble methods in machine learning. Multiple Classifier Systems: First International Workshop, MCS 2000 Cagliari, Italy, June 21–23, 2000 Proceedings 1. Springer, 2000, pp. 1–15.
- Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, 2016, pp. 785–794.
- Breiman, L. Random forests. Machine learning 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Frontiers in neurorobotics 2013, 7, 21. [Google Scholar] [CrossRef] [PubMed]
- Takalo-Mattila, J.; Heiskanen, M.; Kyllönen, V.; Määttä, L.; Bogdanoff, A. Explainable Steel Quality Prediction System Based on Gradient Boosting Decision Trees. IEEE Access 2022. [Google Scholar] [CrossRef]
- Ayvaz, S.; Alpay, K. Predictive maintenance system for production lines in manufacturing: A machine learning approach using IoT data in real-time. Expert Systems with Applications 2021, 173, 114598. [Google Scholar] [CrossRef]
- Hao, X.; Zhang, G.; Ma, S. Deep learning. International Journal of Semantic Computing 2016, 10, 417–439. [Google Scholar] [CrossRef]
- Dey, L.; Meisheri, H.; Verma, I. Predictive Analytics with Structured and Unstructured data-a Deep Learning based Approach. IEEE Intell. Informatics Bull. 2017, 18, 27–34. [Google Scholar]
- Taud, H.; Mas, J. Multilayer perceptron (MLP). Geomatic approaches for modeling land change scenarios 2018, pp. 451–455.
- Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN variants for computer vision: history, architecture, application, challenges and future scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
- Ribani, R.; Marengoni, M. A survey of transfer learning for convolutional neural networks. 2019 32nd SIBGRAPI conference on graphics, patterns and images tutorials (SIBGRAPI-T). IEEE, 2019, pp. 47–57.
- Shafiee, M.J.; Chywl, B.; Li, F.; Wong, A. Fast YOLO: A fast you only look once system for real-time embedded object detection in video. arXiv preprint arXiv:1709.05943 2017.
- Sateesh Babu, G.; Zhao, P.; Li, X.L. Deep convolutional neural network based regression approach for estimation of remaining useful life. Database Systems for Advanced Applications: 21st International Conference, DASFAA 2016, Dallas, TX, USA, April 16-19, 2016, Proceedings, Part I 21. Springer, 2016, pp. 214–228.
- Medsker, L.R.; Jain, L. Recurrent neural networks. Design and Applications 2001, 5, 64–67. [Google Scholar]
- Staudemeyer, R.C.; Morris, E.R. Understanding LSTM–a tutorial into long short-term memory recurrent neural networks. arXiv preprint arXiv:1909.09586 2019.
- Dey, R.; Salem, F.M. Gate-variants of gated recurrent unit (GRU) neural networks. 2017 IEEE 60th international midwest symposium on circuits and systems (MWSCAS). IEEE, 2017, pp. 1597–1600.
- Wu, Y.; Yuan, M.; Dong, S.; Lin, L.; Liu, Y. Remaining useful life estimation of engineered systems using vanilla LSTM neural networks. Neurocomputing 2018, 275, 167–179. [Google Scholar] [CrossRef]
- Creswell, A.; White, T.; Dumoulin, V.; Arulkumaran, K.; Sengupta, B.; Bharath, A.A. Generative adversarial networks: An overview. IEEE signal processing magazine 2018, 35, 53–65. [Google Scholar] [CrossRef]
- Liu, K.; Li, A.; Wen, X.; Chen, H.; Yang, P. Steel surface defect detection using GAN and one-class classifier. 2019 25th international conference on automation and computing (icac). IEEE, 2019, pp. 1–6.
- Verstraete, D.; Droguett, E.; Modarres, M. A deep adversarial approach based on multi-sensor fusion for semi-supervised remaining useful life prognostics. Sensors 2019, 20, 176. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Zhang, W.; Ma, H.; Luo, Z.; Li, X. Data alignments in machinery remaining useful life prediction using deep adversarial neural networks. Knowledge-Based Systems 2020, 197, 105843. [Google Scholar] [CrossRef]
- Khan, S.A.; Prosvirin, A.E.; Kim, J.M. Towards bearing health prognosis using generative adversarial networks: Modeling bearing degradation. 2018 International Conference on Advancements in Computational Sciences (ICACS). IEEE, 2018, pp. 1–6.
- Bank, D.; Koenigstein, N.; Giryes, R. Autoencoders. arXiv preprint arXiv:2003.05991 2020.
- Su, C.; Li, L.; Wen, Z. Remaining useful life prediction via a variational autoencoder and a time-window-based sequence neural network. Quality and Reliability Engineering International 2020, 36, 1639–1656. [Google Scholar] [CrossRef]
- Doersch, C. Tutorial on variational autoencoders. arXiv preprint arXiv:1606.05908 2016.
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Advances in neural information processing systems 2017, 30. [Google Scholar]
- Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in vision: A survey. ACM computing surveys (CSUR) 2022, 54, 1–41. [Google Scholar] [CrossRef]
- Wen, Q.; Zhou, T.; Zhang, C.; Chen, W.; Ma, Z.; Yan, J.; Sun, L. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125 2022.
- Zaheer, M.; Guruganesh, G.; Dubey, K.A.; Ainslie, J.; Alberti, C.; Ontanon, S.; Pham, P.; Ravula, A.; Wang, Q.; Yang, L.; others. Big bird: Transformers for longer sequences. Advances in neural information processing systems 2020, 33, 17283–17297. [Google Scholar]
- Wang, J.; Xu, G.; Yan, F.; Wang, J.; Wang, Z. Defect transformer: An efficient hybrid transformer architecture for surface defect detection. Measurement 2023, 211, 112614. [Google Scholar] [CrossRef]
- Shang, H.; Sun, C.; Liu, J.; Chen, X.; Yan, R. Defect-aware transformer network for intelligent visual surface defect detection. Advanced Engineering Informatics 2023, 55, 101882. [Google Scholar] [CrossRef]
- Chen, D.; Hong, W.; Zhou, X. Transformer network for remaining useful life prediction of lithium-ion batteries. Ieee Access 2022, 10, 19621–19628. [Google Scholar] [CrossRef]
- Johnson, J.M.; Khoshgoftaar, T.M. Survey on deep learning with class imbalance. Journal of Big Data 2019, 6, 1–54. [Google Scholar] [CrossRef]
- Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable ai: A review of machine learning interpretability methods. Entropy 2020, 23, 18. [Google Scholar] [CrossRef] [PubMed]
- Saxena, A.; Goebel, K. PHM08 challenge data set, NASA AMES prognostics data repository. Moffett Field, CA, Tech. Rep 2008. [Google Scholar]
- Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A comprehensive survey on transfer learning. Proceedings of the IEEE 2020, 109, 43–76. [Google Scholar] [CrossRef]
- Vujović, Ž.; others. Classification model evaluation metrics. International Journal of Advanced Computer Science and Applications 2021, 12, 599–606. [Google Scholar] [CrossRef]
- Polenta, A.; Tomassini, S.; Falcionelli, N.; Contardo, P.; Dragoni, A.F.; Sernani, P. A Comparison of Machine Learning Techniques for the Quality Classification of Molded Products. Information 2022, 13, 272. [Google Scholar] [CrossRef]
- Microsoft. Predictive Maintenance Modelling Guide. 2018. https://github.com/ashishpatel26/Predictive_Maintenance_using_Machine-Learning_Microsoft_Casestudy/tree/master/data (Consulted on March 23, 2023).
- Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM computing surveys (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef]
- Patro, S.; Sahu, K.K. Normalization: A preprocessing stage. arXiv preprint arXiv:1503.06462 2015.
- Gogtay, N.J.; Thatte, U.M. Principles of correlation analysis. Journal of the Association of Physicians of India 2017, 65, 78–81. [Google Scholar] [PubMed]
- Li, E.; Zeng, L.; Zhou, Z.; Chen, X. Edge AI: On-demand accelerating deep neural network inference via edge computing. IEEE Transactions on Wireless Communications 2019, 19, 447–457. [Google Scholar] [CrossRef]







| Data Partition | Size |
|---|---|
| Training set | 870 |
| Validation set | 290 |
| Test set | 291 |
| Total | 1451 |
| Model | Precision | Recall | F1-Score |
|---|---|---|---|
| SVM | 0.87 | 0.86 | 0.86 |
| Naive Bayes | 0.85 | 0.83 | 0.83 |
| KNN | 0.92 | 0.91 | 0.91 |
| Logistic Regression | 0.80 | 0.79 | 0.79 |
| Decision Tree | 0.91 | 0.91 | 0.91 |
| Random Forest | 0.98 | 0.98 | 0.98 |
| XGBoost | 0.94 | 0.93 | 0.93 |
| Gradient Boosting | 0.95 | 0.94 | 0.94 |
| Extra Trees | 0.94 | 0.93 | 0.93 |
| MLP | 0.95 | 0.95 | 0.95 |
| Layer (type) | Output Shape | Param # |
|---|---|---|
| dense (Dense) | (None, 80) | 1120 |
| dense_1 (Dense) | (None, 50) | 4050 |
| dropout (Dropout) | (None, 50) | 0 |
| dense_2 (Dense) | (None, 4) | 204 |
| Total params: 5,374 | ||
| Trainable params: 5,374 | ||
| Non-trainable params: 0 | ||
| Data partition | Size (RNN models) | Size (Other models) |
| Training set | 189517 | 193528 |
| Validation set | 46346 | 47804 |
| Test set | 47047 | 47047 |
| Total | 281910 | 288379 |
| Model | Precision | Recall | F1-Score |
|---|---|---|---|
| SVM | 0.95 | 0.96 | 0.96 |
| Naive Bayes | 0.61 | 0.95 | 0.71 |
| KNN | 0.59 | 0.34 | 0.39 |
| Logistic Regression | 0.93 | 0.84 | 0.88 |
| Decision Tree | 0.99 | 0.94 | 0.96 |
| Random Forest | 0.99 | 0.96 | 0.97 |
| XGBoost | 0.99 | 0.97 | 0.98 |
| Gradient Boosting | 0.98 | 0.76 | 0.77 |
| Extra Trees | 0.98 | 0.95 | 0.96 |
| MLP | 0.96 | 0.97 | 0.96 |
| SimpleRNN | 0.98 | 0.94 | 0.96 |
| LSTM | 0.98 | 0.95 | 0.96 |
| GRU | 0.99 | 0.97 | 0.98 |
| Layer (Type) | Output Shape | Param # |
|---|---|---|
| gru (GRU) | (None, 56) | 14,784 |
| dense (Dense) | (None, 5) | 285 |
| Total params: 15,069 | ||
| Trainable params: 15,069 | ||
| Non-trainable params: 0 | ||
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
