Preprint
Review

This version is not peer-reviewed.

Enabling Intelligent Industrial Automation: A Review of Machine Learning Applications with Digital Twin and Edge AI Integration

A peer-reviewed article of this preprint also exists.

Submitted:

24 June 2025

Posted:

25 June 2025

You are already at the latest version

Abstract
The integration of machine learning (ML) into industrial automation is fundamentally reshaping how manufacturing systems are monitored, inspected, and optimized. This review explores the transformative role of ML across three key domains: Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO). By applying machine learning to real-time sensor data and operational histories, advanced models enable proactive fault prediction, intelligent inspection, and dynamic process control—directly enhancing system reliability, product quality, and efficiency. This review also analyzes how Digital Twin (DT) and Edge AI technologies are being leveraged to address domain-specific challenges in Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO). DTs enable virtual replication and simulation of industrial systems, supporting predictive analytics and optimization in low-risk environments. Meanwhile, Edge AI supports low-latency, on-device inference, allowing ML models to operate in real time, even in bandwidth-constrained or cloud-independent settings. Together, these technologies are expanding the practical impact of ML across industrial automation tasks. The paper also catalogs the datasets used, the tools and sensors employed for data collection, and the industrial software platforms supporting ML deployment in practice. It positions ML, DT, and Edge AI as central to the evolution of intelligent and connected manufacturing systems.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

Industrial automation is undergoing a significant transformation driven by the integration of advanced technologies, with Machine Learning (ML) at the forefront. Traditional automation systems, based on rigid rule sets and fixed logic, often struggled to adapt to the variability and complexity inherent in modern manufacturing environments. Over the past decade, ML-driven intelligent automation has emerged as a game-changer—offering systems the ability to learn from historical and real-time data, recognize patterns, make predictions, and continuously optimize operations without explicit reprogramming [1,2,3]. This shift has profoundly impacted several core areas of manufacturing. In Predictive Maintenance (PdM), ML models analyze sensor data to predict equipment failures before they occur, thereby reducing downtime and extending asset lifespan. In Quality Control (QC), image-based deep learning and signal analysis enable real-time defect detection with higher accuracy than human inspection. In Process Optimization (PO), ML algorithms adaptively fine-tune operational parameters to increase yield, minimize waste, and reduce energy consumption [4,5]. The evolution of Industry 4.0 has further accelerated this transformation, introducing a cyber-physical framework characterized by interconnectivity, decentralized decision-making, and real-time data analytics. This revolution is powered by the convergence of Industrial Internet of Things (IIoT), cloud computing, sensor networks, and advanced control systems, producing massive volumes of heterogeneous data [6]. However, generating data alone is not sufficient; it must be meaningfully processed and applied. Here, ML and AI emerge as critical tools for extracting actionable insights from complex datasets [7].
Beyond centralized AI, the trend is shifting toward Edge AI and Digital Twin (DT) technologies, which collectively enhance responsiveness, interpretability, and system resilience. Edge AI refers to deploying ML models directly on edge devices—such as embedded systems and industrial controllers—allowing for low-latency, real-time decision-making without relying on cloud infrastructure [8]. This is particularly valuable for time-sensitive tasks like defect detection and anomaly response on the production line. Digital Twins, on the other hand, are virtual replicas of physical assets or systems that continuously synchronize with real-time operational data. They serve as simulation and decision-support tools, enabling predictive analytics, what-if testing, and system optimization without interfering with live operations. When combined with ML and Edge AI, Digital Twins become intelligent, adaptive models capable of autonomous decision support and system improvement [9,10]. Together, these technologies underpin the four core design principles of Industry 4.0: interconnection, information transparency, technical assistance, and decentralized decision-making [11]. In this paper, we explore how AI technologies are advancing Predictive Maintenance, Quality Control, and Process Optimization, and how the integration of Digital Twin and Edge AI is enabling scalable, responsive, and intelligent automation across industrial settings.
To ground this review in current research, papers were systematically collected from leading academic and industrial databases including Web of Science, Scopus, and IEEE Xplore. Priority was given to works demonstrating practical deployments, novel ML architectures, or integration with enabling technologies such as Digital Twin (DT) and Edge AI. This paper is structured to answer three key research questions:
RQ1: How are machine learning models transforming industrial automation tasks in PdM, QC, and PO?
RQ2: In what ways do Digital Twin and Edge AI technologies enhance or enable those domains in real-world industrial contexts?
RQ3: What datasets, tools, and platforms are being used in practice to support intelligent automation, and what gaps remain in their adoption and deployment?
Together, these questions guide a comprehensive review of the role ML, DT, and Edge AI play in advancing intelligent, resilient, and data-driven manufacturing systems. The remainder of the paper is organized as follows: Section 2 introduces core machine learning theories; Section 3 reviews recent ML applications in PdM, QC, and PO; Section 4 highlights the enabling roles of Digital Twin and Edge AI technologies; Section 5 surveys datasets, data acquisition tools, and industrial platforms; Section 6 concludes with insights and directions for future work.

2. Machine Learning Algorithms

Machine Learning is broadly categorized into Supervised Learning (SL), Unsupervised Learning (UL), and Reinforcement Learning (RL). SL uses labeled data to train models for classification (categorical outputs) and regression (continuous outputs) using algorithms like Neural Networks, SVM, Decision Trees, and Logistic Regression. UL works with unlabeled data to discover patterns through clustering, dimensionality reduction (e.g., PCA, Autoencoders), and density estimation. RL involves agents learning by interacting with environments to maximize rewards over time, often using Q-learning or Deep Reinforcement Learning, with applications in process optimization and supply chains. Figure 1 presents a structured overview of machine learning algorithm classifications.

3. ML Application in Predicitve Maintenance, Quality Control, and Process Optimization

In this study, we examine the application of machine learning (ML) techniques across three key domains of industrial automation:
  • Predictive Maintenance (PdM)
  • Quality Control (QC), and
  • Process Optimization (PO)
These domains were selected due to their critical roles in enhancing operational reliability, reducing production costs, and improving product consistency. Each represents a distinct yet complementary area where ML has demonstrated significant potential in augmenting traditional industrial practices. The following sections provide a structured analysis of recent advancements in ML-based approaches within these domains, highlighting the algorithms, methodologies, and deployment strategies that are shaping the future of intelligent manufacturing systems.

3.1. Predictive Maintenance

Predictive Maintenance (PdM) is one of the most vibrant and rapidly evolving application areas in industrial automation which utilizes sensor technologies, data analytics and machine learning (ML) algorithms to anticipate machine failure before they occur. The growing interest is driven by its direct impact on reducing unplanned downtimes, optimizing maintenance cycles, and ultimately improving system reliability and cost-efficiency [12]. Recent contributions highlight notable advances in automation, interpretability, and adaptability of PdM systems, pushing the boundaries of traditional maintenance paradigms. Table 1 shows the applications of ML in PdM. A particularly compelling development is the integration of AutoML frameworks in PdM pipelines. Tools like PyCaret and AutoKeras can significantly streamline the process of selecting and tuning machine learning models for fault classification in ball bearings. The unique contribution of this work is the elimination of intensive manual feature engineering, making it especially appealing for small-to-medium enterprises (SMEs) seeking low-code predictive capabilities [13]. Image-based deep learning has been used for time-series data to improve accuracy in fault detection. Multivariate sensor data is converted into time-series images and then used convolutional neural networks (CNNs) to detect faults of conveyor belt motors. This approach leveraged spatial representation for enhanced accuracy and addressed the challenge of handling high-dimensional, asynchronous signals in industrial settings [14].
The evolution of PdM is marked by a shift from traditional signal analysis to advanced LSTM-GAN models for accurate Remaining Useful Life (RUL) prediction in bearing systems. By combining LSTM (as a generative model) and autoencoders (as discriminators), the method tackles the common issue of error accumulation in long-term degradation forecasts, providing more stable and accurate predictions [15]. Domain-specific optimization in PdM is gaining traction, as seen in the use of SVR models to predict lubricant wear from robot sensor data, reducing reliance on manual testing [16]. This signals a broader movement toward context-aware PdM, where models are tailored not just to generalize, but to extract meaningful insights in operationally relevant terms. Cost-sensitive maintenance decision-making continues to be a priority. Unsupervised anomaly detection (e.g., Isolation Forest) has been combined with reliability models like Generalized Fault Trees to develop robust hybrid systems for components like injection molds. These approaches don’t just detect faults—they quantify the economic trade-offs between preventive and corrective maintenance, making them invaluable for strategic planning [17].
Corrective maintenance costs a lot of money. It can be avoided with the help of Machine Learning techniques. Artificial Neural Network can be used to predict the future condition of a motor and figure out the time of failure [18]. Line-start permanent magnet synchronous motors (LS-PMSM) are widely used in the industry. But broken rotor bars are one of the most significant faults of this kind of motor, leading to other secondary failures and eventually collapses. Random forest algorithms can diagnose this kind of fault in LS-PMSM by extracting features from the startup transient current signal [19]. General Electric (GE) has used predictive maintenance in its manufacturing facilities. By analyzing sensor data from industrial equipment, they can predict potential issues and schedule maintenance before a breakdown occurs. They focus on critical areas such as fault diagnostics, prognostics, process automation, and prescriptive analytics that can quantify uncertainty and assess risk to provide safe and optimal actions or policies to improve business outcomes [20].
Table 1. ML Application in Predictive Maintenance.
Table 1. ML Application in Predictive Maintenance.
Sub-Area Publication Year & Reference Algorithm Task, Methodology, & Outcome
Fault Prediction 2023, [21] Supervised Learning (LSTM, KNN, KG) Task: Robot state prediction and PdM strategy generation. Methodology: LSTM for state detection, KNN for fault prediction, and KG for decision support. Outcome: Closed-loop PdM system for welding robots.
Fault Prediction 2023, [22] Supervised Learning (RF, GB, DL) Task : Predict failures in a manufacturing plant. Methodology: ML models trained on factory equipment data. Outcome: Improved failure prediction, reduced downtime.
Fault Prediction 2022, [23] Supervised Learning(SVM,BNN, RF) Task: Fault detection and classification in LV motors. Methodology: Two-phase ML approach (abnormal behavior detection + fault type prediction). Outcome:Reduced detection time, accurate fault diagnosis.
Fault Prediction 2018, [24] Supervised Learning (LSTM) Task: Build a smart predictive maintenance system for early fault detection and technician support. Methodology: Used IoT sensors for data collection, LSTM/GRU for failure prediction, and AR tools (HoloLens/tablet) to guide maintenance actions. Outcome: Improved fault prediction and reduced downtime. AR support made maintenance faster and easier for operators.
Fault Prediction 2018, [25] Supervised Learning (BN) Task:Develop a fault modeling and diagnosis system. Methodology: A Bayesian Network (BN) framework was used to represent causal relationships between process parameters and faults. A hybrid learning system was created to improve fault prediction and root cause analysis. Outcome: The system demonstrated improved fault modeling and interpretability.
Fault Prediction 2018, [26] Supervised Learning, Unsupervised Learning (RF) Task: Develop a real-time fault detection and diagnosis system in smart factory environments. Methodology: Employed a big data pipeline integrating data acquisition, storage, preprocessing, and analytics. Used Principal Component Analysis (PCA) and k-Nearest Neighbors (k-NN) for dimensionality reduction and classification. Applied decision trees for fault reasoning. Outcome: Achieved over 90% accuracy in fault classification across multiple use cases.
Fault Prediction 2017, [27] Supervised Learning (ANN) Task: Enable predictive maintenance in machine centers. Methodology: Proposed a five-step framework integrating sensors, AI, CPS, and ANN for fault diagnosis and prognosis. Demonstrated using machine data to predict backlash errors. Outcome: Successfully predicted faults weeks in advance, enabling proactive maintenance.
Fault Prediction 2023, [28] Supervised Learning (CF) Task: Cooling system monitoring. Methodology: Open-source R-based DSS with data preprocessing and predictive models. Outcome: Cost-effective PdM for SMEs.
Condition Monitoring 2021, [29] Supervised Learning (ET) Task: Develop scalable PdM framework. Methodology: Modular edge-cloud architecture with plug-and-play sensor integration and time-series ML. Outcome: Demonstrated early condition degradation in HPC components.
Condition Monitoring 2019, [30] Supervised Learning, Unsupervised Learning (PCA, DTree, RF, KNN, SVM) Task: Predict tool wear in CNC end-milling operations using multi-sensor data. Methodology: Time and frequency domain features were extracted and fused. Machine learning models (Random Forest, SVM, MLP) were trained and validated. Outcome: Random Forest achieved the best performance. Sensor fusion enhanced prediction accuracy over individual sensors.
Condition Monitoring 2018, [31] Supervised Learning (LDA, Clustering) Task: Improve fault diagnosis in Fused Deposition Modeling (FDM) using acoustic emission (AE) data to monitor extruder health. Methodology: Extracted time/frequency domain features were reduced via Linear Discriminant Analysis (LDA). Unsupervised clustering (CFSFDP) was used to identify states without prior labels. Outcome: Achieved 90.2% classification accuracy across five states using 2D feature space. CFSFDP outperformed other clustering methods in F1 score and accuracy.
Lifetime Prediction 2023, [32] Supervised Learning (RF, XGBoost, MLP, SVR) Task: Remaining Useful Life estimation. Methodology: Comparative ML modeling with filtering, clustering, and feature engineering. Outcome: RF achieved best results; prevented 42% of failures.
Lifetime Prediction 2021, [33] Supervised Learning Task: RUL prediction for robot reducer. Methodology: Use motor current signature analysis (MCSA) features in ML model. Outcome: Effective health state classification.
Cost Minimization 2022, [34] Supervised Learning Task: Develop PdM for wiring firms. Methodology: Expert system using ML to reduce downtime. Outcome: Identifies AI as cost-effective alternative to PM.
Cost Minimization 2019, [35] Supervised Learning Task: Optimize maintenance timing in parallel production lines. Methodology: Used multi-agent PPO-based reinforcement learning in a simulated environment to decide maintenance based on machine state and buffer load. Outcome: Reduced breakdowns by 80%, improved throughput by up to 28%, and cut maintenance costs by 19%.

3.2. Quality Control

Quality Control (QC) is a systematic process employed in manufacturing and other industries to ensure that products or services meet defined quality standards and specifications. It involves the inspection and testing of products to identify defects or deviations from desired quality levels, thereby ensuring consistency and reliability in output [36]. Quality Control (QC) represents a rapidly evolving domain within industrial automation where machine learning (ML) techniques, as shown in Table 2, are delivering measurable benefits in defect detection, real-time monitoring, and predictive modeling. Recent advancements are characterized by the integration of deep learning architectures, the emergence of soft sensor-based inspections, and the fusion of visual and contextual data streams for high-precision assessments.
In the printing industry, deep convolutional networks have been used for real-time industrial inspection. A Deep Learning-based soft sensor with a high-resolution optical camera is integrated in the industrial automation for gravure cylinder surface inspection [37]. This system not only improved defect detection accuracy but also significantly reduced manual inspection time, highlighting the potential of computer vision in quality-intensive sectors. Similarly, in the food packaging industry, pre-trained convolutional networks such as-DenseNet161, ResNet50, etc. were employed to automate defect detection in thermoformed trays. The model, trained on domain-specific image datasets, showed high reliability in identifying sealing and closure anomalies—a task traditionally reliant on human inspection and prone to inconsistency [38].
Another compelling approach involves time-series based predictive quality modeling in automotive manufacturing. A combination of supervised models (LSTM, Random Forest, Neural Networks) has been used to predict hole positioning in bumper beams during milling operations. By forecasting potential deviations early in the process, their system facilitated in-process adjustments, reducing tolerance violations and production scrap [39]. Emerging work on online quality assessment systems introduces lightweight models suitable for deployment in real-time production environments. WDCNN with Follow-the-Regularized-Leader, an online learning algorithm, is used for condition assessment in cars and general manufacturing. These methods adapt to evolving process data, making them especially relevant for high-mix, low-volume production contexts [40].
Across recent contributions to quality control research, three key patterns emerge. First, visual inspection using deep learning is becoming standard for complex surface and packaging tasks, where high-resolution, image-based data is readily available. Second, predictive modeling with time-series data facilitates early detection of dimensional or positional deviations, making it well-suited for continuous manufacturing environments. Finally, online and adaptive learning approaches are being adopted to enhance model robustness in dynamic industrial settings, supporting the development of real-time QC systems for both small and large-scale operations. The Quality Control landscape is marked by a shift from retrospective quality checks to real-time, predictive, and autonomous decision-making systems. Future research will likely focus on explainable AI (XAI), domain adaptation, and low-latency deployment architectures to scale these innovations across industries.
Table 2. ML Application in Quality Control.
Table 2. ML Application in Quality Control.
Sub-Area Publication Year & Reference Algorithm Task, Methodology, & Outcome
Defect Detection 2024, [41] Supervised Learning, Unsupervised Learning (YOLOv5, OCR,CNN) Task: Real-time defect detection in tuna cans. Methodology: Used YOLOv5 for can inspection, OCR for label detection, integrated with IoT stack (Node-RED, Grafana). Outcome: High-speed classification, automated sorting via robotic arm.
Defect Detection 2023, [42] Supervised Learning (LSTM, RF, NN) Task : Predict hole locations in bumper beams to preempt quality issues. Methodology: Trained time-series models using previous beam measurements. Outcome: Improved early detection of tolerance violations, enhancing QC and reducing scrap.
Defect Detection 2023, [43] Supervised Learning, Unsupervised Learning (Custom CNN) Task: Visual defect detection in casting. Methodology: Developed custom CNN and deployed on shop floor via user-friendly app. Outcome:Achieved 99.86% accuracy in image-based inspection for castings.
Defect Detection 2022, [44] Supervised Learning, Unsupervised Learning (CNN) Task: Visual flaw detection with explainability. Methodology: Combined CNN for image analysis with ILP for rule-based reasoning, integrated human-in-the-loop feedback. Outcome: Created a system offering human-verifiable justifications.
Defect Detection 2019, [45] Supervised Learning (CNN) Task: On-line defect recognition in Selective Laser Melting (SLM) during additive manufacturing. Methodology: Developed a bi-stream Deep Convolutional Neural Network (DCNN) to analyze layer-wise in-process images (powder layers and part slices) and detect defects caused by improper SLM parameters. Outcome: Achieved 99.4% defect classification accuracy; model outperformed traditional approaches; supports adaptive SLM process control and real-time quality assurance.
Defect Detection 2018, [46] Supervised Learning (DTree) Task: Detect keyholing porosity and balling instabilities in Laser Powder Bed Fusion (LPBF). Methodology: Applied Scale-Invariant Feature Transform (SIFT) to extract melt pool features, encoded using Bag-of-Words representation, followed by classification with Support Vector Machine (SVM). Outcome: Enabled accurate identification of melt pool defects, supporting quality control in LPBF processes.
Image Recognition 2019, [39] Supervised Learning, Unsupervised Learning (SIFTS, SVM) Task: Monitor and predict tool wear conditions in milling operations. Methodology: Tool condition classification was performed using a Support Vector Machine (SVM). A cloud dashboard was used for monitoring and visualization. Outcome: Enabled efficient and scalable monitoring of tool conditions, supporting timely maintenance decisions.
Image Recognition 2018, [47] Supervised Learning (SVM) Task: Detect anomalies and failures in industrial manufacturing processes. Methodology: Employed an intelligent agent with a threshold-based decision algorithm and trained it using operational data. Outcome: Enabled proactive fault detection and efficient process management, reducing unexpected downtimes.
Image Recognition 2018, [48] Supervised Learning (CNN) Task: Predict track width and continuity in Laser Powder Bed Fusion (LPBF) using video analysis. Methodology: Trained a CNN using supervised learning on 10 ms in situ video clips of the LPBF process. Outcome: Enabled accurate prediction of track features from video, supporting real-time quality monitoring.
Online Quality Control 2023, [40] Supervised Learning (WDCNN, FTRL) Task: Real-time quality assessment of cars and bearings. Methodology: Applied online learning (incremental updates) with identity parsing on streaming data using river in Python. Outcome: Achieved real-time classification with stable accuracy.
Online Quality Control 2021, [38] Supervised Learning, Unsupervised Learning (CNN) Task: Detect sealing and closure defects in food trays inline. Methodology: Built a modular CV system using CNNs trained on domain-specific image datasets. Outcome: Achieved near 100% defect detection rate inline, with <0.3
Online Quality Control 2019, [49] Supervised Learning (SVM) Task: Enable cost-efficient real-time QC in automotive manufacturing. Methodology: Applied an SVM considering inspection costs and error types; performance assessed via Design of Experiment. Outcome: Effective QC with improved cost-sensitivity and error handling.
Online Quality Control 2019, [50] Supervised Learning, Unsupervised Learning (SDAE) Task: Perform robust pattern recognition from noisy signals. Methodology: Used SDAE for unsupervised feature extraction and supervised regression fine-tuning. Outcome: Improved generalization and feature robustness for classification tasks.

3.3. Process Optimization

Process Optimization represents a critical domain in industrial automation where machine learning (ML) is increasingly leveraged to improve efficiency, adapt to dynamic conditions, and minimize resource usage. As shown in Table 3, recent contributions demonstrate an encouraging trend toward the integration of ML with physical modeling, virtual environments, and human-machine interactions—highlighting a shift from static parameter tuning to adaptive, intelligent process control.
A compelling development is seen in the context of thermoplastic composites manufacturing, where a hybrid ML framework was proposed to optimize the Automated Fiber Placement (AFP) process. This approach integrates Artificial Neural Networks (ANNs), Finite Element Analysis (FEA), and Virtual Sample Generation (VSG) to overcome the small-data bottleneck common in advanced manufacturing. The study demonstrated how ML models trained on both synthetic and experimental data could significantly reduce defect rates while optimizing key parameters like speed, heat, and compaction pressure [14]. The concept of an “Artificial Neural Twin” was introduced for plastic recycling industry. This virtual twin environment combined differentiable data fusion and model predictive control (MPC) with supervised and unsupervised learning to optimize decentralized process chains. Notably, the use of simulation-based training within a Unity-developed environment allowed for safe, iterative learning—especially useful for distributed manufacturing systems where real-time experimentation may be costly or risky [51].
Human-centered optimization has also gained traction. A data-driven adaptation model is utilized for industrial HMIs, mining 151 days of interaction logs to identify user patterns and generate adaptive interface rules. The goal was to improve efficiency and decision-making in repetitive operator tasks, pointing toward a broader trend of integrating behavioral analytics and rule learning into interface design [52]. In more traditional sectors, such as water treatment and desalination, ML models such as ANFIS, BPNN, SVR, and RBF networks were applied to optimize membrane-based processes like electrodialysis and reverse osmosis. These models helped predict pollutant removal efficiency and system throughput under variable conditions, demonstrating ML’s capability to capture nonlinear, multi-parameter dependencies in physical systems [53].
Across these works, several recurring themes highlight the evolving landscape of process optimization. Hybrid modeling is increasingly adopted to combine machine learning with domain-specific physics, particularly in scenarios where experimental data is limited or costly [54]. Digital twin and simulation environments are enabling low-risk, data-driven optimization of complex systems, especially in distributed or reconfigurable manufacturing settings. Additionally, modeling operator behavior and adapting interfaces reflect a growing emphasis on human-in-the-loop optimization, where systems dynamically support human decisions through learned behavioral patterns. Collectively, these trends mark a shift from static parameter tuning toward self-adaptive, learning-based systems that evolve with the production environment. Looking ahead, future research is expected to prioritize explainable control strategies, edge deployment for latency-sensitive tasks, and the integration of reinforcement learning for both continuous and episodic control scenarios [55].
Table 3. ML Application in Process Optimization.
Table 3. ML Application in Process Optimization.
Sub-Area Publication Year & Reference Algorithm Task, Methodology, & Outcome
Performance Prediction 2024, [51] Supervised Learning, Unsupervised Learning (MPC) Task:: Optimize process chains via decentralized learning. Methodology: Uses a quasi-neural network model with gradient-based continual learning across distributed nodes. Outcome: Enables continual optimization without compromising data sovereignty.
Performance Prediction 2022, [56] Supervised Learning (Bayesian Optimization with Probabilistic Constraints) Task : Improve solar cell efficiency using data-efficient optimization. Methodology: BO with human-in-the-loop feedback and prior knowledge constraints. Outcome: Achieved 18.5% PCE with only 100 tests—faster than conventional methods.
Performance Prediction 2019, [53] Supervised Learning (ANN, GA, RBF, BPNN, ANFIS, SVR) Task: Optimize and model desalination and treatment processes. Methodology: Benchmarked ANN/GA vs classical models for ion rejection, flux prediction, pollutant removal, etc. Outcome:ANN-based tools achieved superior prediction accuracy and process adaptability.
Performance Prediction 2019, [57] Supervised Learning (NN) Task: Predict temperature and density evolution from laser trajectories. Methodology: Used three neural networks with a localized trajectory decomposition technique. Outcome: Enabled spatially-aware predictions for process monitoring.
Performance Prediction 2018, [58] Supervised Learning (CNN) Task: Identify geometries that are hard to manufacture. Methodology: Applied a 3D CNN with a secondary interpretability method to analyze feature contribution. Outcome: Accurately predicted and explained manufacturability issues.
Performance Prediction 2018, [59] Supervised Learning (RF, SVM) Task: Predict lead time in variable-demand flow shops. Methodology: Employed a Twin model with frequent retraining and online learning. Outcome: Achieved adaptive and accurate lead time forecasts.
Process Control 2025, [60] Supervised Learning (RSM-GA, ANN-GA, ANFIS-GA) Task: Maximize tensile, flexural, and compressive strengths in FDM parts. Methodology: Used hybrid optimization combining RSM and AI methods on experimental design. Outcome: Hybrid models improved strength by up to 8.86% across mechanical tests.
Process Control 2024, [61] Reinforcement Learning, Supervised Learning (TD3, PPO) Task: Develop autonomous process control in injection molding. Methodology: Combines supervised learning + DRL in a digital twin framework. Outcome: Real-time optimization with reduced human involvement and improved quality/cost-efficiency balance.
Process Control 2022, [54] Supervised Learning (ANN) Task: Optimize AFP process to reduce defects and improve ILSS. Methodology: Combined ANN with photonic sensors, VSG, and FEA simulations. Outcome: Developed a decision-support tool to automate parameter tuning and defect minimization.
Process Control 2020, [62] Reinforcement Learning (QLrn) Task: Optimize control in nonlinear, uncertain manufacturing processes. Methodology: Applied Q-learning for independent decision-making under partial observability. Outcome: Achieved adaptive control despite randomness and incomplete information.
Process Control 2019, [63] Supervised Learning (SVM) Task: Improve grinding parameters for helical flutes. Methodology: Combined simulation, SVM prediction, and simulated annealing to optimize feed rate and grinder speed. Outcome: Enhanced surface quality and process efficiency.
Scheduling 2019, [64] Reinforcement Learning (QLrn) Task: Minimize makespan in robotic assembly lines. Methodology: Used multi-agent reinforcement learning for dynamic planning and task scheduling. Outcome: Improved scheduling efficiency in multi-robot systems.
Scheduling 2018, [65] Supervised Learning (Bagging, Boosting) Task: Optimize job shop scheduling via dispatching rule selection. Methodology: Evaluated bagging, boosting, and stacking for rule selection. Outcome: Reduced mean tardiness and flow time.

4. Machine Learning-Driven Digital Twins and Edge AI for Industrial Automation

Digital Twin (DT) and Edge Artificial Intelligence (Edge AI) are key technological enablers in advancing intelligent industrial systems. DT facilitates the creation of virtual replicas of physical assets, enabling real-time monitoring, simulation, and predictive decision-making. Meanwhile, Edge AI enables localized data processing at the source, minimizing latency and bandwidth consumption while supporting fast, autonomous responses. Together, these technologies form a powerful synergy for optimizing operations, reducing downtime, and enhancing productivity across industrial domains. This section presents a comprehensive review of how DT and Edge AI are being applied across various functions—ranging from predictive maintenance to adaptive quality control—highlighting their integration, implementation strategies, and industrial impact.

4.1. Digital Twin

Digital Twin (DT) technology has emerged as one of the most transformative innovations in industrial automation. A digital twin is a virtual replica of a physical system, process, or asset—ranging from individual machines and manufacturing cells to entire production lines or facilities. It enables bidirectional communication between the physical and virtual domains, creating a closed-loop system that enhances visibility, predictability, and control. At the core of a digital twin lies the physical-to-virtual connection, which is facilitated by real-time data streams collected from sensors and industrial IoT devices. These data streams ensure that the virtual representation continuously mirrors the real-world system’s operational status. More importantly, DT systems are not merely passive representations; they are designed to process and analyze real-time data, enabling live monitoring and simulation of system behavior. This empowers engineers and operators to detect anomalies, anticipate faults, and make informed decisions proactively.
What distinguishes a full-fledged digital twin from basic simulations or static models is its virtual-to-physical feedback loop. In this loop, insights or predictions generated in the digital environment are transmitted back to the physical system for action [67]. For instance, in predictive maintenance applications, the digital twin can identify early signs of equipment degradation and trigger maintenance operations before actual failure occurs. This tight integration dramatically enhances operational efficiency and system reliability [68]. In parallel, Artificial Intelligence (AI) has become a critical enabler of industrial automation. AI techniques, particularly machine learning (ML) and deep learning (DL), offer powerful tools to learn from operational data, predict system behavior, detect faults, and optimize production parameters. AI systems thrive on large volumes of data—precisely the kind of rich, multi-dimensional datasets provided by digital twins [27]. The convergence of AI and digital twin technologies represents a significant leap forward in the evolution of smart manufacturing. Since both systems depend on real-time and historical data streams, their integration enables mutually reinforcing intelligence. The digital twin provides a continuously updated, structured representation of the system, while AI contributes predictive and prescriptive capabilities by identifying complex patterns, correlations, and anomalies within the data. Figure 2 illustrates a schematic representation of a DT integrated with AI components. By leveraging vast historical datasets and AI models, especially those based on supervised learning, reinforcement learning, or unsupervised anomaly detection, the predictive performance of digital twins can be significantly enhanced. Machine learning models can be embedded within the digital twin to predict the remaining useful life (RUL) of components [69], detect process drifts and anomalies [70], or even learn to select optimal actions through simulated interactions [71]. This transforms the digital twin into a dynamic, self-learning system capable of adaptive decision-making without direct reliance on the physical system.
This synergistic relationship is particularly evident in three critical areas of industrial automation, Predictive Maintenance, Quality Control, and Process Optimization. In this section, we explore the integration of digital twins and AI in the context of Predictive Maintenance, Quality Control, and Process Optimization—three domains where this convergence is driving measurable gains in efficiency, reliability, and autonomy in industrial systems.

4.1.1. AI- Driven Digital Twin Applications for PdM, QC, and PO

Digital Twin (DT) technology has become a pivotal tool in advancing industrial AI applications by enabling real-time replication, monitoring, and control of physical assets. When augmented with machine learning (ML), deep learning (DL), or reinforcement learning (RL), these digital representations evolve from passive models to adaptive, predictive engines. Table 4 shows how AI-powered DTs are increasingly leveraged to address core challenges in Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO).
In the Predictive Maintenance domain, DTs enable real-time diagnostics and long-term health-state monitoring. One application in photovoltaic (PV) systems used a DT to integrate ensemble ML models and neural networks, delivering high-accuracy fault diagnostics by comparing DT outputs with field data. This cloud-based platform could process live and synthetic inputs from geographically distributed PV systems, significantly improving response times and predictive capability [72]. Similarly, another PV-focused DT embedded eXtreme Gradient Boosting (XGBoost) within a cloud-monitoring framework to estimate system health from sensor telemetry—enabling timely interventions and minimizing downtime [73]. A particularly innovative PdM application involved smart forging, where a DT was used as an environment to train a deep reinforcement learning (DRL) agent (Proximal Policy Optimization). The DRL system adjusted induction heating coil power dynamically during the forging process to reduce temperature variation and minimize defects—advancing zero-defect manufacturing in heavy industry [74].
For Quality Control, the QUILT framework stands out as an example of leveraging side-channel data (acoustic, magnetic, vibration) within a DT for legacy 3D printing systems. By applying supervised ML models on these unconventional signals, the system could detect anomalies and ensure print quality without modifying the underlying hardware—pushing the boundary of QC into the realm of non-invasive diagnostics [75]. A multi-domain digital twin framework has been used in [76], for real-time detection of fatigue, abnormal behavior, cyber threats, and energy anomalies in a textile factory.
In Process Optimization, DTs have become enablers of intelligent simulation and control. A standout example is the Artificial Neural Twin, which fused supervised and unsupervised learning with Model Predictive Control (MPC) to optimize decentralized manufacturing lines using Unity-based simulation environments [51]. Another study used a DT-driven CatBoost model to optimize energy consumption in industrial drying, achieving annual savings of 3.7 MWh in a real-world deployment [77]. Likewise, Ant Colony Optimization was employed within a robotic DT to learn collision-free paths for industrial arms, showcasing DTs as a training ground for heuristic RL agents [78].
A broader view reveals a convergence of DTs with Sim-to-Real transfer, where models trained in DTs are deployed in physical systems with minimal adaptation. For instance, DRL agents trained on DTs of robotic grasping tasks successfully transferred policies to real-world robot arms, achieving high success rates while ensuring operational safety through risk-monitoring modules [79]. Finally, AI-integrated DTs are expanding into cross-domain orchestration, such as holistic security in smart factories. One system—the SMS-DT—combined supervised and unsupervised models across edge, platform, and enterprise tiers to ensure cyber-physical security and operational resilience. It not only predicted faults and anomalies but also correlated them across diverse factory systems including human behavior, machine status, and network security [76].
Several shared themes emerge across the application of digital twins (DTs) in industrial domains. Simulation-enhanced learning allows DTs to generate synthetic yet representative data, supporting the training of machine learning models when real-world data is limited or incomplete [74,80]. Real-time adaptability is achieved through continuous live data streaming, enabling AI models to evolve alongside changing system states. In safety-critical environments, the combination of DTs with explainable AI (XAI) and physics-based simulations enhances system transparency and trust [81]. Additionally, the move toward lightweight, edge-compatible DTs facilitates their deployment in decentralized factory settings. Collectively, these advancements position AI-empowered digital twins as active agents in industrial intelligence—providing predictive insight, real-time control, and continuous learning across predictive maintenance (PdM), quality control (QC), and process optimization (PO). Integrations of DTs with federated learning, generative simulation, and cyber-physical feedback to unlock greater autonomy in smart manufacturing could be some of the focuses for the future research.
Table 4. AI- Driven Digital Twin in Industrial Automation.
Table 4. AI- Driven Digital Twin in Industrial Automation.
Area Sub-Area Publication Year & Reference Algorithm Task, Methodology, & Outcome
Predictive Maintenance Fault Prediction 2024, [82] Supervised Learning (LSTM, CNN) Task: Predict early failure of SiC/GaN semiconductors. Methodology: Built a DT for thermal monitoring and used ML for degradation prediction. Outcome: Enabled early detection and extended device lifespan.
Predictive Maintenance Fault Prediction 2022, [83] Supervised Learning (BPNN) Task: Improve fault prediction and diagnosis for large-diameter auger rigs in coal mining. Methodology: Developed a digital twin model with geometric, physical, and behavioral layers using Unity3D and ANSYS. Trained a BP neural network on fault data (4 fault types) with expert-assisted feedback correction. Outcome: Model showed strong performance in identifying drill pipe bend/fracture, bearing fault, and overpressure events.
Predictive Maintenance Fault Prediction 2021, [84] Supervised Learning, Unsupervised Learning (IRF, HC, TL) Task: Improve fault detection and classification on intelligent production lines. Methodology: Proposed IRF by filtering RF trees via hierarchical clustering (high accuracy + diversity), then applied transfer learning to fine-tune with physical data. Outcome: Achieved 97.8% accuracy (vs. 88.7% for RF); outperformed KNN, ANN, LSTM, SVM; effective in diagnosing conveyor, tightening, and alignment faults with low-latency online analysis.
Predictive Maintenance Fault Prediction 2021, [85] Supervised Learning Task: Predict surface defects in HPDC castings. Methodology: Converted HPDC process images into pixel-based tabular data; applied SVD and edge detection for dimensionality reduction. Trained a Random Forest classifier and mapped tree paths into distributed CEP engine rules for real-time inference. Outcome: : RF + SVD + edge detection achieved 99.99% accuracy; crack location precisely identified in test images. CEP model enabled lightweight, distributed, low-latency defect prediction without large-scale computation.
Predictive Maintenance Fault Prediction 2020, [86] Supervised Learning (NN, RF, RR) Task: Predict generator oil temperature and detect early anomalies to prevent aircraft No-Go events. Methodology: Segmented time-series data from 606 anomaly-free flights and applied Fourier/Haar basis expansion. NN chosen for best generalization. Anomalies detected by monitoring divergence from reference MSE over consecutive flights. Outcome: Detected failures 5 to 9 flights before actual events; NN-Fourier DT achieved MSE  0.10 and showed good anomaly sensitivity with minimal false positives.
Predictive Maintenance Fault Prediction 2019, [87] Supervised Learning, Unsupervised Learning (DNN) Task: Perform real-time fault diagnosis under data-scarce and distribution-shifting conditions in smart manufacturing. Methodology: Proposed DFDD (Digital-Twin Fault Diagnosis using DTL), trained DNN model in digital space using SSAE + Softmax, then used DTL with MMD to adapt model to physical space. Integrated Process Visibility System (PVS) retrieved shop-floor operation data without extra sensors. Outcome:DFDD achieved 97.96% accuracy, outperforming DNN trained only on virtual (67.59%) or physical (91.54%) data. Robust against imbalanced and distribution-shifted test sets.
Predictive Maintenance Lifetime Prediction 2022, [88] Supervised Learning (LASSO, SVR, XGBoost) Task: Achieve full-lifecycle monitoring and predictive maintenance for locomotives. Methodology: Proposed a 3-layer ML-integrated DT architecture. Applied ML to a Digital Twin in Maintenance (DTMT) for health monitoring and fault prediction using bearing temperature data. Developed a series combination model (LASOO + SVR + XGBoost) to forecast axle temperature trends. Outcome:Detected locomotive bearing faults  1 week in advance. Enabled proactive fault alerts and lifecycle optimization.
Predictive Maintenance Lifetime Prediction 2021, [89] Supervised Learning, Unsupervised Learning (LSTM) Task: Enhance predictive maintenance of aero-engines through data-driven digital twin modeling. Methodology: Developed an implicit digital twin (IDT) using sensor data and historical operation data, integrated with LSTM for RUL prediction. Applied S-G filter for denoising. Defined Health Index (HI) to evaluate degradation and predict RUL using IDT-LSTM. Outcome: Achieved RMSE of 13.12 for RUL prediction, outperforming other methods; optimal performance at 80% training data.
Predictive Maintenance Health Monitoring 2024, [82] Supervised Learning (DNN) Task: Monitor WBG semiconductor health using a digital twin. Methodology: Combined thermal-electrical simulation and AI models to predict degradation. Outcome: Enabled accurate lifetime estimation and failure prediction using hybrid DT-AI approach.
Quality Control Defect Detection 2022, [90] Supervised Learning, Unsupervised Learning (SVR, GPR) Task: Identify bearing crack type and size under variable speed. Methodology: Modeled AE signals using autoregression, SVR, and GPR combined with Laguerre filters. Estimated unknown signals using a strict-feedback backstepping DT with fuzzy logic. Generated residuals and used RMS features for classification via SVM. Outcome: Achieved 97.13% accuracy in crack type diagnosis and 96.9% in crack size classification across eight bearing conditions and multiple speeds.
Quality Control Defect Detection 2022, [91] Supervised Learning, Unsupervised Learning (LR, K-means Clustering) Task: Detect anomalies in a pasteurization system at a food plant using ML-enhanced Digital Twin. Methodology: Built a LabVIEW-Python based DT of a pilot pasteurizer using real-time pressure and flow data. Trained 3 ML models: linear regressor (P1 prediction), MLP classifier (machine status: ok, warning, failure), and K-means (unsupervised status clustering). Outcome: MLP reached 96–99% accuracy across fluids; K-means accurately clustered operational states. DT enabled remote monitoring and decision support.
Quality Control Image Recognition 2022, [92] Supervised Learning, Unsupervised Learning (CNN) Task: Monitor and classify the quality of banana fruit. Methodology: Developed a Digital Twin system using thermal images (FLIR One camera) labeled into four classes. Trained a CNN with SAP Intelligent Technologies and used cloud-edge architecture for real-time data collection and alerts. Outcome: Enabled real-time classification and inventory decision-making.
Quality Control Image Recognition 2020, [93] Supervised Learning (Inception-v3 CNN with Transfer Learning) Task: Classify orientation ("up" or "down") of 3D-printed parts in robotic pick-and-place system. Methodology: Synthetic images generated using DT simulations in Blender. Labeled with Python script. Inception-v3 CNN retrained using TensorFlow. Outcome: Achieved 100% accuracy on real-world images; validated DT-generated data for robust model training.
Quality Control Image Recognition 2020, [94] Supervised Learning, Unsupervised Learning (CNN) Task: Monitor and control weld joint growth and penetration. Methodology: Built DT using weld images processed by CNN for BSBW and image processing for TSBW. Unity GUI for visualization. System adaptively adjusted welding time to meet penetration specs. Outcome: Real-time monitoring via visualization.
Quality Control Image Recognition 2020, [95] Supervised Learning, Unsupervised Learning (MobileNet, UNet, Transfer Learning) Task: Enable low-cost, high-precision plant disease/nutrient deficiency detection. Methodology: LoRaWAN WSN collected sensor data; used MobileNet and UNet on PlantVillage dataset. Simulated WSN in OMNeT++ and FLoRa; image downsampling for efficiency. Outcome: 95.67% validation accuracy; enabled rural deployment via energy-efficient LoRa-based WSN.
Quality Control Online Quality Control 2022, [96] Supervised Learning, Unsupervised Learning (PointNet) Task: Real-time object detection and pose estimation in robotic DT system. Methodology: Built DT with ROS and Unity for ABB IRB 120. Used LineMod and PointNet for object recognition/pose estimation. Collected data with Blensor and RealSense D435i. Outcome: 100% classification accuracy,  3° pose error; real-time DT sync with <0.1 ms delay.
Quality Control Online Quality Control 2022, [97] Supervised Learning, Unsupervised Learning (YOLOv4-M2, OpenPose) Task: Improve small object detection in complex smart manufacturing. Methodology: Designed a hybrid model using MobileNetv2+YOLOv4 for object detection and OpenPose for long-range human posture detection. Outcome: Achieved 91.8% accuracy, 78.2% mAP at 8–10 m.
Quality Control Online Quality Control 2021, [98] Supervised Learning (FFT, PCA, SVM) Task: Enhance welder training and performance using VR-based DT. Methodology: Captured motion via VR, transmitted to UR5 robot. Used FFT-PCA-SVM to classify welding skill. Outcome: 94.44% classification accuracy; enabled immersive feedback and performance monitoring.
Process Optimization Performance Prediction 2023, [99] Supervised Learning (ANN, k-NN, Symbolic Regression) Task: Predict and optimize workstation productivity using DT. Methodology: Combined PPC and ML to forecast throughput from failure/downtime data. Outcome: Symbolic Regression: R 2 =0.96 (train); ANN: R 2 =0.95 (test); enabled adaptive PPC decisions.
Process Optimization Performance Prediction 2022, [100] Supervised Learning, Unsupervised Learning (CNN, Spatio-Temporal GCN) Task: Predict road behavior and secure data transfer in autonomous cars. Methodology: Combined CNN and DT with spatio-temporal GCN and load balancing. Outcome: 92.7% prediction accuracy, 80% delivery rate, low delay and leakage.
Process Optimization Performance Prediction 2022, [101], Reinforcement Learning (BCDDPG, LSTM) Task: Enable robust and energy-efficient flocking of UAV swarms. Methodology: Developed DT-enabled framework using BCDDPG and LSTM for dynamic feature learning. Trained in simulation and deployed to UAVs. Outcome: Outperformed baselines in 8 metrics including arrival rate >80% and energy efficiency.
Process Optimization Task Modelling 2022, [102] Reinforcement Learning (DDQN) Task: Minimize energy in UAV-based Mobile Edge Computing. Methodology: DT-based offloading with DDQN, closed-form power solutions, and iterative CPU allocation. Outcome: Reduced energy and delay vs. baselines; scalable under dynamic loads.
Process Optimization Process Control 2024, [103] Supervised Learning (CNN, YOLOv3) Task: Object detection in factories. Methodology: Trained YOLOv3 on synthetic data from factory DT. Outcome: Enabled robust object recognition without real datasets.
Process Optimization Process Control 2022, [104] Supervised Learning, Unsupervised Learning (VGG-16) Task: Enable intuitive robot programming. Methodology: DT system with Hololens MR, Unity simulation, and CNN for object pose estimation. Outcome: Real-time gesture control with ±1–2 cm error.
Process Optimization Process Control 2022, [105] Reinforcement Learning (PDQN, DQN) Task: Optimize smart conveyor control. Methodology: Built DT-ACS and introduced PDQN to improve control performance. Outcome: Faster convergence, better robustness, reduced cost under dynamic loads.
Process Optimization Process Control 2021, [106] Supervised Learning, Unsupervised Learning (K-Means, KNN) Task: Improve monitoring and prediction in chemical plants. Methodology: Preprocessed data (IQR, normalization), clustered via K-Means, and built KNN models. Deployed model to cloud with WebSocket interface. Outcome: 16.6% data reduction, 99.74% classification accuracy, R2 = 0.96 for regression.
Process Optimization Process Control 2019, [107] Supervised Learning (LightGBM, XGBoost, RF, AdaBoost, CART) Task: Optimize yield in catalytic cracking units. Methodology: 5-step DT framework using IoT + ML; trained 4 models with ensemble methods; online deployment with MES. Outcome: Real-world deployment increased light oil yield by 0.5%.
Process Optimization Scheduling 2022, [108] Reinforcement Learning (Parallel RL, Q-Learning, SARSA, DNN) Task: Improve shipyard scheduling and QoS management. Methodology: Built 3-layer DTN; trained DNN for latency prediction; tested RL variants. Outcome: Parallel RL had best performance; DT enabled real-time decisions and resource efficiency.
Process Optimization Scheduling 2021, [109] Supervised Learning (ANN) Task: Enhance planning in fast fashion lines. Methodology: DT system with ANN for demand forecast, DES for simulating operations, and dashboard visualization. Outcome: Lead time reduced by 28%, operator use up 37%, staffing optimized.
Process Optimization Scheduling 2020, [110] Reinforcement Learning Task: Optimize scheduling in manual assembly. Methodology: Built Python-based adaptive simulation using FPY/HPU data, RL for recommendation refinement. Outcome: Identified bottlenecks and improved efficiency; RL adapted dynamically.

4.2. Edge AI

Industrial automation generates massive volumes of data, traditionally handled through cloud-based solutions. However, relying solely on cloud infrastructure introduces latency, bandwidth limitations, and security risks. To address these challenges, edge computing has emerged as a complementary solution. It leverages computing-capable edge devices—located near the data source—to perform processing locally, thereby minimizing the need for continuous cloud communication [111,112]. When artificial intelligence (AI) models are deployed directly on these edge devices, the paradigm is referred to as Edge AI [113]. This approach eliminates the dependency on remote servers for inference, enabling faster response times, reduced data transmission, enhanced privacy, and lower power consumption [114,115].
Edge AI represents a significant advancement in the field of intelligent systems. Devices such as the NVIDIA Jetson Nano, Raspberry Pi 5, and Texas Instruments F28P55x microcontrollers offer cost-effective and high-performance platforms capable of running AI models at the edge . These platforms empower manufacturers to implement intelligent decision-making for tasks such as anomaly detection, quality inspection, and real-time process optimization without the need for high-end GPUs or extensive cloud infrastructure [4,116,117]. Recognized by industry analysts like Deloitte and Gartner as one of the fastest-growing technologies in artificial intelligence, Edge AI is rapidly gaining traction in smart manufacturing [118,119]. It enables real-time, localized automation while keeping system costs and complexity under control—making it a transformative force in the evolution of Industry 4.0.

4.2.1. Edge AI in PdM, QC, and PO

Edge Artificial Intelligence (Edge AI) has rapidly emerged as a pivotal enabler for real-time industrial intelligence by pushing computation closer to the data source. Unlike traditional cloud-based approaches, Edge AI systems offer low-latency, high-reliability inference at the device level, making them especially suitable for time-sensitive tasks in manufacturing, maintenance, and quality inspection. Table 5 demonstrates how Edge AI is being operationalized across multiple domains—most notably Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO).
In the domain of Predictive Maintenance, one of the most impactful applications of Edge AI is remaining useful life (RUL) estimation for machinery components. Using the NASA C-MAPSS dataset, LSTM, CNN-RNN, and hybrid models were implemented on edge hardware (e.g., Raspberry Pi and NVIDIA Jetson) to predict degradation states of aircraft engines. The study demonstrated that inference times remained within practical limits even on constrained hardware, showing how deep learning models can be adapted for real-time RUL assessment in industrial settings [120]. Another application focuses on thermal error prediction in CNC machines, where LSTM networks and hybrid models (e.g., improved Grey Relational Analysis) were deployed on FPGA-based edge systems [5]. Genetic Algorithm (GA) is used on IIoT sensor data to enable predictive maintenance and energy scheduling to reduce energy use by 28.1% in [121]. In Quality Control, Edge AI has proven instrumental in acoustic defect detection. A real-world deployment in a beverage factory used LSTM and SVM-based models running on edge servers to classify acoustic signatures of glass bottles on a high-speed conveyor. The system achieved high accuracy in identifying defective units and replaced human-based inspection with an autonomous, high-throughput solution—illustrating the scalability of Edge AI in visually and acoustically intensive QC tasks [122]. Similarly, in railway bearing monitoring, fuzzy logic-based decision systems combined with FFT processing were executed on fog/edge computing platforms to detect overheating in real time. By processing thermal and vibration signals locally, the system enabled timely fault classification and reduced the reliance on centralized diagnostic hubs [123].
From a process optimization perspective, Edge AI supports real-time feedback and dynamic decision-making. The iRobot-Factory study showcased an intelligent robotic system that integrates ML and DL models at the fog layer to manage distributed robotic arms and manufacturing cells. Tactile and audiovisual inputs were processed at the edge to determine operator emotions and optimize task assignments—highlighting Edge AI’s role not just in performance optimization but also in human-machine interaction enhancement [124]. Several key trends highlight the transformative role of Edge AI in smart manufacturing. Low-latency learning enables time-sensitive predictions and decisions to be made directly at the data source, minimizing reliance on centralized processing [4,122,123]. Hardware-aware model design is gaining traction, with compact architectures and deployment strategies optimized for FPGA and ARM-based edge devices to meet stringent computational and energy constraints [4,5,124]. Multimodal sensor integration is becoming standard, allowing edge systems to process diverse input streams—such as thermal, audio, and vibration data—for richer machine learning inference. Additionally, autonomous operation is increasingly common, with many systems functioning independently of cloud connectivity, enhancing resilience and preserving data privacy. Collectively, Edge AI is redefining the operational frontier of smart manufacturing by enabling intelligent, autonomous, and adaptive systems directly on the shop floor. Future research should emphasize federated edge learning, energy-efficient model architectures, and lifecycle-aware retraining strategies to ensure sustained performance in dynamic production environments.
Table 5. Edge AI in Industrial Automation.
Table 5. Edge AI in Industrial Automation.
Area Sub-Area Publication Year & Reference Algorithm Task, Methodology, & Outcome
Predictive Maintenance Fault Prediction 2024, [116] Supervised Learning (SVM, RF, KNN, CNN, LightBGM) Task: : Detect tool wear in milling. Methodology: Developed an Edge AI system running 5 SL models on low-cost hardware. Outcome: CNN outperformed others in wear classification, enabling efficient on-device inference.
Predictive Maintenance Fault Prediction 2020, [125] Supervised Learning, Unsupervised Learning (GBRBM, DNN) Task: Accurately detect faults in IIoT manufacturing facilities using edge AI with minimal latency. Methodology: Transforms fault detection into a classification task using a multi-block GBRBM (Gaussian-Bernoulli Restricted Boltzmann Machine) for feature extraction and deep autoencoder for training. The architecture enables low-latency classification directly at the edge. Outcome: Achieved 88.39% accuracy; significantly outperformed SVM, LDA, LR, QDA, and FNN baselines.
Predictive Maintenance Fault Prediction 2020, [126] Supervised Learning, Unsupervised Learning (1D-CNN) Task: Accurately detect gear and bearing faults in gearboxes under multiple operating conditions using deep learning on edge equipment. Methodology: Proposed a multi-task 1D-CNN model trained with shared and task-specific layers. Model deployed on edge devices for low-latency real-time diagnosis. Outcome: Achieved 95.76% joint accuracy; after applying triplet loss, test accuracy reached 90.13
Predictive Maintenance Anomaly Detection 2020, [127] Supervised Learning, Unsupervised Learning (CNN-VA, SCVAE) Task: Perform unsupervised anomaly detection on time-series manufacturing sensor data. Methodology: Proposes SCVAE (compressed CNN-VAE using Fire Modules) trained on labeled UCI datasets and unlabeled CNC machine data. Compares SCVAE with other anomaly detection methods. Outcome: : SCVAE achieved high anomaly detection accuracy while reducing model size and inference time significantly, making it suitable for edge deployment.
Quality Control Defect Detection 2020, [128] Supervised Learning, Unsupervised Learning (R-CNN, ResNet101) Task: Detect surface defects on complex-shaped manufactured parts (turbo blades). Methodology: Faster R-CNN is deployed at edge nodes for low-latency detection, while cloud servers support training and updates. The smart system integrates cloud-edge collaboration for continuous model evolution. Outcome: Achieved 81% precision and 72% recall on test set; edge computing improved speed over cloud or embedded-only setups.
Quality Control Defect Detection 2021, [129] Supervised Learning, Unsupervised Learning (CNN) Task: Automate visual defect detection in injection-molded tampon applicators using deep learning and edge computing. Methodology: A CNN model processes grayscale images acquired from vision sensors mounted on rotating rails.The system performs real-time defect classification on edge boxes connected to PLCs for automated sorting. Outcome:Achieved 92.62% accuracy and 0.839 MCC with fast inference, validating industrial applicability.
Quality Control Defect Detection 2020, [130] Unsupervised Learning (K-means Clustering) Task: Develop a real-time, low-latency fabric defect detection system. Methodology:Modified DenseNet is optimized with a custom loss function, data augmentation (6 strategies), and pruning for edge deployment. Trained and deployed on Cambricon 1H8 edge device with factory data. Outcome:Achieved 18% AUC gain, 50% reduction in data transmission, and 32% lower latency vs cloud, validating robust, real-time performance for 11 defect classes.
Quality Control Image Recognition 2023, [131] Supervised Learning, Unsupervised Learning (TADS) Task: Optimize execution time of DNN-based quality inspection tasks in smart manufacturing. Methodology: Proposes TADS (Task-Aware DNN Splitting), a scheme that selects optimal DNN layer split points based on task number, type (concurrent/periodic), inter-arrival time, and bandwidth. Outcome:Achieved up to 97% task time reduction vs baseline schemes; validated through both simulations and real-world deployment.
Quality Control Image Recognition 2021, [132] Supervised Learning (MobileNetV1, ResNet) Task: Improve operator safety and operational tracking in a shipyard workshop. Methodology: A mist computing architecture using smart IIoT cameras performs real-time human detection and machinery tracking locally without uploading image data to the cloud. Outcome: Demonstrated extremely low yearly energy consumption (0.35–0.36 kWh/device) and scalable carbon footprint analysis across regions using different energy sources.
Quality Control Image Recognition 2020, [133] Supervised Learning (SVM) Task: Automate detection of edge and surface defects in logistics packaging boxes. Methodology: Images are preprocessed with grayscale, denoising, and morphological operations. Features are extracted using SIFT and classified using SVM (RBF kernel). Outcome: Achieved 91% accuracy in classifying edge and surface defects, outperforming CNN in both accuracy and speed under edge computing conditions.
Quality Control Online Quality Control 2020, [134] Supervised Learning (GBT, SVM, DT, NB, LR) Task: Replace traditional X-ray inspections in PCB manufacturing. Methodology: Historical SPI data were used to train supervised models (GBT selected). Prediction occurs on solder-joint level; deployment strategy filters X-ray usage based on predicted FOV defect status. Outcome: 29% average X-ray inspection volume reduced without sacrificing defect detection accuracy.
Process Optimization Process Control 2020, [135] Supervised Learning, Unsupervised Learning (ResNet34, RFBNet, Key Point Regression) Task: Estimate and calibrate the 3D pose of robotic arms with five key points (base, shoulder, elbow, wrist, end). Methodology: Two-stage pipeline—robot arm detection with RFBNet and key point regression using a lightweight CNN (ResNet34 backbone). Trained on RGB-D data from Webots simulator, deployed on NVIDIA Jetson AGX. Outcome: Achieved 1.28 cm joint error, 0.70 cm base error; 14 FPS on edge device with low GPU memory.
Process Optimization Scheduling 2020, [136] Supervised Learning, Unsupervised Learning (LSTM, FCM clustering) Task: Detect anomalies in discrete manufacturing processes and perform energy-aware production rescheduling. Methodology: Energy data is collected from CNC tools and preprocessed (cleaning, clustering by FCM). An LSTM model predicts tool wear and machine degradation. If an anomaly occurs, an edge-triggered rescheduling mechanism (RSR/TR) is initiated. Outcome: 3.5% detection error; energy and production efficiency improved by 21.3% and 13.7% respectively.

5. Dataset, Data Acquisition Tools, and Industrial Platforms

The development and deployment of effective AI models in industrial automation critically depend on the quality and characteristics of the datasets used. This section systematically reviews the publicly available and proprietary datasets, data acquisition devices, and the input-output variables employed across AI applications in predictive maintenance, quality control, and process optimization. By analyzing dataset sources, feature types, and labeling methods, this section aims to guide researchers and practitioners in understanding the current data landscape and identifying gaps in data-driven industrial automation research.

5.1. Dataset

A critical dimension of AI development in industrial automation is the nature and richness of the data used to train, validate, and deploy models. Table 6 provides a comprehensive summary of datasets utilized in recent research studies. In Predictive Maintenance (PdM), the most prevalent data type is time-series sensor data, which captures the temporal evolution of equipment behavior. This includes vibration signals [137,138], thermal readings [5], and acoustic emissions [123]. In Quality Control (QC), the dominant data type is visual imagery, derived from high-resolution cameras or industrial scanners. Visual datasets are used to detect surface defects, cracks, or scratches in manufactured parts and to classify product orientation, packaging compliance, or misassembly [4]. Some applications enhance visual inspection by integrating thermal imaging [139] or acoustic signals [122] to detect internal or non-visible defects—highlighting the move toward multimodal quality sensing. In Process Optimization (PO), datasets are more varied and often involve multivariate process parameters, such as temperature, pressure, feed rate, or cutting speed, collected during operations [5,53]. The use of robot kinematic and trajectory data [78], for motion planning, energy optimization, and collision avoidance in industrial robotics has also been found in some applications. Synthetic or simulation-based datasets are used to supplement limited real-world data, particularly when experiments are costly or disruptive to production [51,54,140,141,142]. Digital twin systems, serving as dynamic virtual replicas of physical assets, utilize a blend of real sensor data and synthetic inputs to enable predictive modeling and optimization. For instance, synthetic CAD-generated image datasets are used to pretrain CNN classifiers for part orientation, requiring only a small number of real samples for fine-tuning—an approach that highlights the data efficiency of simulation-augmented learning [93]. On the other hand, Edge AI systems focus on low-latency, on-device inference and operate under hardware constraints that shape their data needs [143]. Designed for autonomous operation, Edge AI datasets are structured to support intermittent connectivity, and in some cases, local or federated learning updates occur without central aggregation [144].
Table 6. Overview of Dataset Characteristics.
Table 6. Overview of Dataset Characteristics.
Area Reference Dataset Used Devices Used Input Variables Output Variables Number of Samples
Predictive Maintenance [88] Real-world axle temperature data from CDD5B1 locomotives Onboard sensors Axle temperature, ambient temp, GPS speed, generator temp Predicted axle temp, residual error, failure alert 0,000
Predictive Maintenance [113] Custom dataset (6 sensors, 6 units) 4 low-power embedded edge devices Accelerometer, gyro, magnetometer, mic Aging classification 939
Predictive Maintenance [126] Custom DDS vibration data (gear & bearing) Edge-ready hardware (lightweight CNNs), DDS simulator, 1D sensors, FFT preprocessor Time-series vibration signals (gear, bearing) Fault category of gear and bearing (multi-label output) 192,000
Predictive Maintenance [145] Time-series current signals from solar panel systems TIDA-010955 AFE board with C2000 control card , current transformers. ADC samples, FFT features. Binary classification: Arc (1) or Normal (0). Not specified
Predictive Maintenance [146] Vibration data (3-axis), collected from motors under various fault conditions. Vibration sensors, motor controller, dual GaN inverters, and EMJ04-APB22 PMSM motors Time-series vibration data, FFT or raw signals. Fault types (e.g., normal, flaking, erosion, localized damage). Not Specified
Quality Control [129] Real factory image dataset from SMEs GigE Vision Cameras, Edge Box (NVIDIA GTX 1080 Ti), PLC, rotating rail Grayscale product images (300×300 px) Binary defect classification (OK/Defective) 3428
Quality Control [130] Alibaba Tianchi fabric dataset (real industrial images) Intelligent edge camera (Cambrian 1H8), ARM Cortex A7 High-res fabric images Defect classification 2022
Quality Control [133] Custom dataset from logistics warehouse TXG12 industrial camera, LED lights, conveyor with PLC Grayscale carton images (500×653 px) Binary classification (OK, Edge Defect, Surface Defect) 3000
Quality Control [147] Custom image dataset (12 defect categories) Sensors, fog nodes, cameras Image features from product sensors Binary/Multiclass defect classification 2400
Process Optimization [4] Custom manufacturing images NVIDIA Jetson Nano Product images, object categories Defect detection, inventory state Not specified
Process Optimization [106] 64,789 records of process data IoT devices Process temps, fan pressure/speed, raw material consumption Operating mode, fault diagnosis, predicted material consumption 61,753
Process Optimization [136] Milling shop energy logs Electric meters, edge server, PLCs, CNC lathes, milling machines Energy consumption metrics Anomaly class (normal, tool wear, degradation), reschedule strategy 1,000
Process Optimization [148] Real CNC motion data Fagor 8070 CNC controller Control loop parameters, speed, load torque, backlash, friction factors Position error, control effort, peak error Not specified

5.2. Industrial Platforms and Software

The AI-driven solutions in industrial automation depends on robust software tools and platforms that support data acquisition, model development, real-time inference, and system optimization. In the context of smart manufacturing, a wide array of commercial, open-source, and hardware-integrated tools are used across three core domains: Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO). These tools facilitate the transition from traditional rule-based systems to intelligent, adaptive, and data-driven manufacturing ecosystems.
In Predictive Maintenance (PdM), software platforms are designed to monitor equipment health, detect anomalies, and predict failures based on sensor data and historical logs. Industry-leading platforms such as ABB Ability™ Condition Monitoring for Motors, ABB Ability™ Predictive Maintenance, IBM Maximo Application Suite, and PTC ThingWorx enable integration of IoT data streams with AI-powered diagnostic models [149,150,151,152]. Azure IoT Suite and Uptake Fusion further enhance PdM capabilities by combining real-time telemetry with cloud analytics [153,154]. For algorithm development and simulation, tools like MATLAB’s Predictive Maintenance Toolbox allow engineers to extract condition indicators, perform signal analysis, and estimate Remaining Useful Life (RUL) [155]. These systems often operate in conjunction with edge devices (e.g., NVIDIA Jetson, Raspberry Pi, Siemens IoT2040) to support low-latency, on-device inference for remote or latency-sensitive environments.
In the domain of Quality Control (QC), computer vision and signal processing software are widely used for high-precision inspection tasks. Tools such as Cognex VisionPro [156], Keyence Vision Systems [157], and NI LabVIEW with the Vision Development Module [158] provide industrial-grade image acquisition and analysis capabilities. Matrox Imaging Library (MIL) is also commonly employed for 2D/3D defect detection [159]. In addition, platforms like ZEISS PiWeb enable statistical process control (SPC) using dimensional measurement data [160]. For AI-based inspection, deep learning models such as YOLO [161], ResNet [162], and EfficientNet [163] are trained using frameworks like OpenCV and deployed on edge AI platforms such as Edge Impulse [164], AWS Panorama [165], or Google Coral [166] enabling real-time classification of defects at the production line.
Process Optimization (PO) leverages simulation-based and algorithmic tools for optimizing process flows, resource allocation, and control strategies. Platforms like Siemens Tecnomatix Plant Simulation, Rockwell Automation Arena, and AnyLogic provide discrete event and agent-based modeling environments for simulating factory operations [167,168,169]. In the chemical and energy sectors, AspenTech’s Aspen Plus and HYSYS are widely adopted for modeling and optimizing thermodynamic and kinetic processes [170,171]. Furthermore, Ansys Twin Builder offers integrated simulation and digital twin environments for real-time feedback and optimization [172]. On the AI side, libraries such as Scikit-learn, Pyomo [173], Optuna [174], and SimOpt [175] support advanced data-driven optimization tasks, including hyperparameter tuning, constraint satisfaction, and decision modeling. Reinforcement learning frameworks like Stable-Baselines3 and Ray RLlib are increasingly being applied to dynamic PO problems where adaptive control policies are required [176,177].
Several tools operate across these domains, reflecting the convergence of AI, IoT, and simulation technologies. Digital Twin platforms like GE Predix [178], Siemens Insights Hub [179], and ANSYS Twin Builder [172] integrate real-time sensor data with virtual system models to support diagnostics, forecasting, and autonomous decision-making. Edge AI runtimes, including OpenVINO [180], TensorRT [181], and NVIDIA Dynamo [182], allow high-performance inference on low-power devices. Additionally, cloud-integrated IIoT services such as AWS IoT Greengrass [183] and Microsoft Azure Digital Twins [184] support secure data flow, remote monitoring, and coordinated control across decentralized production environments.

6. Conclusions

The integration of machine learning (ML) into industrial automation is no longer a theoretical promise but a growing reality, driving measurable improvements across Predictive Maintenance (PdM), Quality Control (QC), and Process Optimization (PO). This review has demonstrated that ML techniques—spanning supervised, unsupervised, and reinforcement learning—enable more intelligent, adaptive, and responsive manufacturing systems. From image-based defect detection to time-series-based predictive modeling and self-optimizing control systems, ML is reshaping the landscape of industrial decision-making.
However, the transformation is not solely driven by algorithms. Emerging technologies such as Digital Twin (DT) and Edge AI are amplifying the practical applicability of ML, especially in decentralized, latency-sensitive, or simulation-dependent environments. DTs enable real-time mirroring and predictive simulation of physical systems, while Edge AI facilitates low-latency inference at the source of data generation. Together, they overcome limitations related to data privacy, network dependency, and real-time responsiveness.
Beyond algorithmic performance, this paper also highlights the role of datasets, data acquisition tools, and industrial software platforms that form the operational backbone of AI-driven automation. Understanding the nature of the data used, the devices involved in its collection, and the tools available for implementation is essential for the successful deployment of ML in manufacturing settings. Despite significant progress, challenges remain. These include the need for explainability in high-stakes environments, the integration of AI with legacy infrastructure, and the development of transferable models for data-scarce applications. Future research should continue to explore hybrid modeling, simulation-to-real transfer, and federated learning to unlock scalable and trustworthy AI deployment in industry.
Overall, this review underscores that ML, DT, and Edge AI are not just complementary technologies but interdependent enablers of a smarter, more resilient industrial future. Their convergence signals a paradigm shift toward self-adaptive, data-driven manufacturing systems aligned with the core vision of Industry 4.0.

Author Contributions

Conceptualization, Mohammad Abidur Rahman and Kamran Iqbal; methodology, Mohammad Abidur Rahman; validation, Kamran Iqbal and Ali Abushaiba; formal analysis, Mohammad Abidur Rahman; investigation, Mohammad Abidur Rahman and Md Farhan Shahrior; resources, Mohammad Abidur Rahman and Md Farhan Shahrior; data curation, Mohammad Abidur Rahman and Md Farhan Shahrior; writing—original draft preparation, Mohammad Abidur Rahman; writing—review and editing, Mohammad Abidur Rahman and Md Farhan Shahrior; visualization, Mohammad Abidur Rahman; supervision, Kamran Iqbal and Ali Abushaiba. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AI Artificial Intelligence
ANFIS Adaptive Neuro-Fuzzy Inference System
ANFIS-GA ANFIS with Genetic Algorithm
ANN Artificial Neural Network
ANN-GA Artificial Neural Network with Genetic Algorithm
BCDDPG Behavior-Coupling Deep Deterministic Policy Gradient
BN Batch Normalization
BNN Bayesian Neural Network
BPNN Backpropagation Neural Network
CART Classification and Regression Trees
CF Collaborative Filtering
CNN Convolutional Neural Network
DDQN Double Deep Q-Network
DL Deep Learning
DNN Deep Neural Network
DQN Deep Q-Network
DT Digital Twin
Dtree Decision Tree
EC Edge Computing
ET Extra Trees
FFT Fast Fourier Transform
GA Genetic Algorithm
GB Gradient Boosting
GPR Gaussian Process Regression
HC Hierarchical Clustering
IoT Internet of Things
IRF Iterative Random Forest
LASSO Least Absolute Shrinkage and Selection Operator
LDA Linear Discriminant Analysis
LR Logistic Regression / Linear Regression
LSTM Long Short-Term Memory
MEC Mobile Edge Computing
ML Machine Learning
MLP Multi-Layer Perceptron
MPC Model Predictive Control
MSE Mean Squared Error
NN Neural Network
OCR Optical Character Recognition
PCA Principal Component Analysis
PdM Predictive Maintenance
PDQN Profit-sharing Deep Q-Network
PO Process Optimization
PPO Proximal Policy Optimization
QC Quality Control
QLrn Q-Learning
RF Random Forest
RL Reinforcement Learning
RMS Root Mean Square
RR Ridge Regression
SARSA State-Action-Reward-State-Action
SCADA Supervisory Control and Data Acquisition
SDAE Stacked Denoising Autoencoder
SIFT Scale-Invariant Feature Transform
ST-GCN Spatio-Temporal Graph Convolutional Network
SVR Support Vector Regression
SVM Support Vector Machine
TD3 Twin Delayed Deep Deterministic Policy Gradient
TL Transfer Learning
UNet U-shaped Convolutional Neural Network
VCG Variational Cooperative Game
XGBOOST Extreme Gradient Boosting
XR Extended Reality

References

  1. Lee, J.; Bagheri, B.; Kao, H.-A. A Cyber-Physical Systems Architecture for Industry 4.0-Based Manufacturing Systems. Manufacturing Letters 2015, 3, 18–23. [CrossRef]
  2. Wuest, T.; Weimer, D.; Irgens, C.; Thoben, K.-D. Machine Learning in Manufacturing: Advantages, Challenges, and Applications. Production & Manufacturing Research 2016, 4(1), 23–45. [CrossRef]
  3. Rahman, M.A.; Abushaiba, A.A.; Elrajoubi, A.M. Integration of C2000 Microcontrollers with MATLAB Simulink Embedded Coder: A Real-Time Control Application. In: Proceedings of the 2024 7th International Conference on Electrical Engineering and Green Energy (CEEGE); IEEE: Los Angeles, CA, USA, 2024; pp. 131–136. [CrossRef]
  4. Gawate, E.; Rane, P. Empowering Intelligent Manufacturing with the Potential of Edge Computing with NVIDIA’s Jetson Nano. In: Proceedings of the 2023 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS); IEEE: Greater Noida, India, 2023; pp. 375–380. [CrossRef]
  5. Liang, Y.C.; Li, W.D.; Lou, P.; Hu, J.M. Thermal Error Prediction for Heavy-Duty CNC Machines Enabled by Long Short-Term Memory Networks and Fog-Cloud Architecture. Journal of Manufacturing Systems 2022, 62, 950–963. [CrossRef]
  6. Qin, J.; Liu, Y.; Grosvenor, R. A Categorical Framework of Manufacturing for Industry 4.0 and Beyond. Procedia CIRP 2016, 52, 173–178. [CrossRef]
  7. Wang, J.; Ma, Y.; Zhang, L.; Gao, R.X.; Wu, D. Deep Learning for Smart Manufacturing: Methods and Applications. Journal of Manufacturing Systems 2018, 48, 144–156. [CrossRef]
  8. Shi, W.; Cao, J.; Zhang, Q.; Li, Y.; Xu, L. Edge Computing: Vision and Challenges. IEEE Internet of Things Journal 2016, 3(5), 637–646. [CrossRef]
  9. Tao, F.; Zhang, H.; Liu, A.; Nee, A.Y.C. Digital Twin in Industry: State-of-the-Art. IEEE Transactions on Industrial Informatics 2019, 15(4), 2405–2415. [CrossRef]
  10. Transforming Manufacturing with Digital Twins | McKinsey. https://www.mckinsey.com/capabilities/operations/our-insights/digital-twins-the-next-frontier-of-factory-optimization. Accessed 2025-05-19.
  11. Hermann, M.; Pentek, T.; Otto, B. Design Principles for Industrie 4.0 Scenarios. In: Proceedings of the 2016 49th Hawaii International Conference on System Sciences (HICSS); IEEE: Koloa, HI, USA, 2016; pp. 3928–3937. http://ieeexplore.ieee.org/document/7427673/. Accessed 2025-05-19. [CrossRef]
  12. Meddaoui, A.; Hain, M.; Hachmoud, A. The Benefits of Predictive Maintenance in Manufacturing Excellence: A Case Study to Establish Reliable Methods for Predicting Failures. The International Journal of Advanced Manufacturing Technology 2023, 128(7–8), 3685–3690. [CrossRef]
  13. Hadi, R.H.; Hady, H.N.; Hasan, A.M.; Al-Jodah, A.; Humaidi, A.J. Improved Fault Classification for Predictive Maintenance in Industrial IoT Based on AutoML: A Case Study of Ball-Bearing Faults. Processes 2023, 11(5), 1507. [CrossRef]
  14. Kiangala, K.S.; Wang, Z. An Effective Predictive Maintenance Framework for Conveyor Motors Using Dual Time-Series Imaging and Convolutional Neural Network in an Industry 4.0 Environment. IEEE Access 2020, 8, 121033–121049. [CrossRef]
  15. Lu, B.-L.; Liu, Z.-H.; Wei, H.-L.; Chen, L.; Zhang, H.; Li, X.-H. A Deep Adversarial Learning Prognostics Model for Remaining Useful Life Prediction of Rolling Bearing. IEEE Transactions on Artificial Intelligence 2021, 2(4), 329–340. [CrossRef]
  16. Guo, D.; Chen, X.; Ma, H.; Sun, Z.; Jiang, Z. State Evaluation Method of Robot Lubricating Oil Based on Support Vector Regression. Computational Intelligence and Neuroscience 2021, 2021(1). [CrossRef]
  17. Nunes, P.; Rocha, E.; Santos, J.; Antunes, R. Predictive Maintenance on Injection Molds by Generalized Fault Trees and Anomaly Detection. Procedia Computer Science 2023, 217, 1038–1047. [CrossRef]
  18. Scalabrini Sampaio, G.; Vallim Filho, A.R.D.A.; Santos Da Silva, L.; Augusto Da Silva, L. Prediction of Motor Failure Time Using an Artificial Neural Network. Sensors 2019, 19(19), 4342. [CrossRef]
  19. Quiroz, J.C.; Mariun, N.; Mehrjou, M.R.; Izadi, M.; Misron, N.; Mohd Radzi, M.A. Fault Detection of Broken Rotor Bar in LS-PMSM Using Random Forests. Measurement 2018, 116, 273–280. [CrossRef]
  20. Predictive Maintenance | GE Research. https://www.ge.com/research/project/predictive-maintenance. Accessed 2024-02-26.
  21. Wang, X.; Liu, M.; Liu, C.; Ling, L.; Zhang, X. Data-Driven and Knowledge-Based Predictive Maintenance Method for Industrial Robots for the Production Stability of Intelligent Manufacturing. Expert Systems with Applications 2023, 234, 121136. [CrossRef]
  22. Satwaliya, D.S.; Thethi, H.P.; Dhyani, A.; Kiran, G.R.; Al-Taee, M.; Alazzam, M.B. Predictive Maintenance Using Machine Learning: A Case Study in Manufacturing Management. In: Proceedings of the 2023 3rd International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE); IEEE: Greater Noida, India, 2023; pp. 872–876. https://ieeexplore.ieee.org/document/10183012/. Accessed 2025-05-23. [CrossRef]
  23. Nikfar, M.; Bitencourt, J.; Mykoniatis, K. A Two-Phase Machine Learning Approach for Predictive Maintenance of Low Voltage Industrial Motors. Procedia Computer Science 2022, 200, 111–120. [CrossRef]
  24. Cachada, A.; Moreira, P.M.; Romero, L.; Barbosa, J.; Leitão, P.; Geraldes, C.A.S.; Deusdado, L.; Costa, J.; Teixeira, C.; Teixeira, J.; Moreira, A.H.J. Maintenance 4.0: Intelligent and Predictive Maintenance System Architecture. In: Proceedings of the 2018 IEEE 23rd International Conference on Emerging Technologies and Factory Automation (ETFA); IEEE: Turin, Italy, 2018; pp. 139–146. https://ieeexplore.ieee.org/document/8502489/. Accessed 2024-04-19. [CrossRef]
  25. Carbery, C.M.; Woods, R.; Marshall, A.H. A Bayesian Network Based Learning System for Modelling Faults in Large-Scale Manufacturing. In: 2018 IEEE International Conference on Industrial Technology (ICIT); IEEE: Lyon, France, 2018; pp. 1357–1362. https://ieeexplore.ieee.org/document/8352377/. Accessed 2024-04-19. [CrossRef]
  26. Syafrudin, M.; Alfian, G.; Fitriyani, N.; Rhee, J. Performance Analysis of IoT-Based Sensor, Big Data Processing, and Machine Learning Model for Real-Time Monitoring System in Automotive Manufacturing. Sensors 2018, 18(9), 2946. [CrossRef]
  27. Li, Z.; Wang, Y.; Wang, K.-S. Intelligent Predictive Maintenance for Fault Diagnosis and Prognosis in Machine Centers: Industry 4.0 Scenario. Advances in Manufacturing 2017, 5(4), 377–387. [CrossRef]
  28. Pejić Bach, M.; Topalović, A.; Krstić, I.; Ivec, A. Predictive Maintenance in Industry 4.0 for the SMEs: A Decision Support System Case Study Using Open-Source Software. Designs 2023, 7(4), 98. [CrossRef]
  29. Borghesi, A.; Burrello, A.; Bartolini, A. ExaMon-X: A Predictive Maintenance Framework for Automatic Monitoring in Industrial IoT Systems. IEEE Internet of Things Journal 2023, 10(4), 2995–3005. [CrossRef]
  30. Li, Z.; Liu, R.; Wu, D. Data-Driven Smart Manufacturing: Tool Wear Monitoring with Audio Signals and Machine Learning. Journal of Manufacturing Processes 2019, 48, 66–76. [CrossRef]
  31. Liu, J.; Hu, Y.; Wu, B.; Wang, Y. An Improved Fault Diagnosis Approach for FDM Process with Acoustic Emission. Journal of Manufacturing Processes 2018, 35, 570–579. [CrossRef]
  32. Taşcı, B.; Omar, A.; Ayvaz, S. Remaining Useful Lifetime Prediction for Predictive Maintenance in Manufacturing. Computers & Industrial Engineering 2023, 184, 109566. [CrossRef]
  33. Lulu, J.; Yourui, T.; Jia, W. Remaining Useful Life Prediction for Reducer of Industrial Robots Based on MCSA. In: 2021 Global Reliability and Prognostics and Health Management (PHM-Nanjing); IEEE: Nanjing, China, 2021; pp. 1–7. https://ieeexplore.ieee.org/document/9613006/. Accessed 2025-05-23. [CrossRef]
  34. Eddarhri, M.; Adib, J.; Hain, M.; Marzak, A. Towards Predictive Maintenance: The Case of the Aeronautical Industry. Procedia Computer Science 2022, 203, 769–774. [CrossRef]
  35. Kuhnle, A.; Jakubik, J.; Lanza, G. Reinforcement Learning for Opportunistic Maintenance Optimization. Production Engineering 2019, 13(1), 33–41. [CrossRef]
  36. Bertolini, M.; Mezzogori, D.; Neroni, M.; Zammori, F. Machine Learning for Industrial Applications: A Comprehensive Literature Review. Expert Systems with Applications 2021, 175, 114820. [CrossRef]
  37. Villalba-Diez, J.; Schmidt, D.; Gevers, R.; Ordieres-Meré, J.; Buchwitz, M.; Wellbrock, W. Deep Learning for Industrial Computer Vision Quality Control in the Printing Industry 4.0. Sensors 2019, 19(18), 3987. [CrossRef]
  38. Banús, N.; Boada, I.; Xiberta, P.; Toldrà, P.; Bustins, N. Deep Learning for the Quality Control of Thermoforming Food Packages. Scientific Reports 2021, 11(1), 21887. [CrossRef]
  39. Scime, L.; Beuth, J. Using Machine Learning to Identify In-Situ Melt Pool Signatures Indicative of Flaw Formation in a Laser Powder Bed Fusion Additive Manufacturing Process. Additive Manufacturing 2019, 25, 151–165. [CrossRef]
  40. Yin, Y.; Wan, M.; Xu, P.; Zhang, R.; Liu, Y.; Song, Y. Industrial Product Quality Analysis Based on Online Machine Learning. Sensors 2023, 23(19), 8167. [CrossRef]
  41. González, S.V.; Jimenez, L.C.; Given, W.G.G.; Noboa, B.V.; Enderica, C.S. Automated Quality Control System for Canned Tuna Production Using Artificial Vision. In: 2024 3rd International Conference on Artificial Intelligence for Internet of Things (AIIoT); IEEE: Vellore, India, 2024; pp. 1–6. https://ieeexplore.ieee.org/document/10574669/. Accessed 2025-05-23. [CrossRef]
  42. Msakni, M.K.; Risan, A.; Schütz, P. Using Machine Learning Prediction Models for Quality Control: A Case Study from the Automotive Industry. Computational Management Science 2023, 20(1), 14. [CrossRef]
  43. Sundaram, S.; Zeid, A. Artificial Intelligence-Based Smart Quality Inspection for Manufacturing. Micromachines 2023, 14(3), 570. [CrossRef]
  44. Müller, D.; März, M.; Scheele, S.; Schmid, U. An Interactive Explanatory AI System for Industrial Quality Control. arXiv 2022, Version Number: 1. [CrossRef]
  45. Caggiano, A.; Zhang, J.; Alfieri, V.; Caiazzo, F.; Gao, R.; Teti, R. Machine learning-based image processing for on-line defect recognition in additive manufacturing. CIRP Ann. 2019, 68(1), 451–454. [CrossRef]
  46. Kim, A.; Oh, K.; Jung, J.-Y.; Kim, B. Imbalanced classification of manufacturing quality conditions using cost-sensitive decision tree ensembles. Int. J. Comput. Integr. Manuf. 2018, 31(8), 701–717. [CrossRef]
  47. Gobert, C.; Reutzel, E.W.; Petrich, J.; Nassar, A.R.; Phoha, S. Application of supervised machine learning for defect detection during metallic powder bed fusion additive manufacturing using high resolution imaging. Addit. Manuf. 2018, 21, 517–528. [CrossRef]
  48. Yuan, B.; Guss, G.M.; Wilson, A.C.; Hau-Riege, S.P.; DePond, P.J.; McMains, S.; Matthews, M.J.; Giera, B. Machine-Learning-Based Monitoring of Laser Powder Bed Fusion. Adv. Mater. Technol. 2018, 3(12), 1800136. [CrossRef]
  49. Oh, Y.; Busogi, M.; Ransikarbum, K.; Shin, D.; Kwon, D.; Kim, N. Real-time quality monitoring and control system using an integrated cost effective support vector machine. J. Mech. Sci. Technol. 2019, 33(12), 6009–6020. [CrossRef]
  50. Yu, J.; Zheng, X.; Wang, S. A deep autoencoder feature learning method for process pattern recognition. J. Process Control 2019, 79, 1–15. [CrossRef]
  51. Emmert, J.; Mendez, R.; Dastjerdi, H.M.; Syben, C.; Maier, A. The Artificial Neural Twin — Process optimization and continual learning in distributed process chains. Neural Netw. 2024, 180, 106647. [CrossRef]
  52. Reguera-Bakhache, D.; Garitano, I.; Uribeetxeberria, R.; Cernuda, C.; Zurutuza, U. Data-Driven Industrial Human-Machine Interface Temporal Adaptation for Process Optimization. In Proceedings of the 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vienna, Austria, 2020; pp. 518–525. [CrossRef]
  53. Al Aani, S.; Bonny, T.; Hasan, S.W.; Hilal, N. Can machine language and artificial intelligence revolutionize process automation for water treatment and desalination? Desalination 2019, 458, 84–96. [CrossRef]
  54. Islam, F.; Wanigasekara, C.; Rajan, G.; Swain, A.; Prusty, B.G. An approach for process optimisation of the Automated Fibre Placement (AFP) based thermoplastic composites manufacturing using Machine Learning, photonic sensing and thermo-mechanics modelling. Manuf. Lett. 2022, 32, 10–14. [CrossRef]
  55. Farooq, A.; Iqbal, K. A Survey of Reinforcement Learning for Optimization in Automation. In Proceedings of the 2024 IEEE 20th International Conference on Automation Science and Engineering (CASE), Bari, Italy, 2024; pp. 2487–2494. [CrossRef]
  56. Liu, Z.; Rolston, N.; Flick, A.C.; Colburn, T.W.; Ren, Z.; Dauskardt, R.H.; Buonassisi, T. Machine Learning with Knowledge Constraints for Process Optimization of Open-Air Perovskite Solar Cell Manufacturing. arXiv 2021, Version Number: 4. [CrossRef]
  57. Stathatos, E.; Vosniakos, G.-C. Real-time simulation for long paths in laser-based additive manufacturing: a machine learning approach. Int. J. Adv. Manuf. Technol. 2019, 104(5–8), 1967–1984. [CrossRef]
  58. Ghadai, S.; Balu, A.; Sarkar, S.; Krishnamurthy, A. Learning localized features in 3D CAD models for manufacturability analysis of drilled holes. Comput. Aided Geom. Des. 2018, 62, 263–275. [CrossRef]
  59. Gyulai, D.; Pfeiffer, A.; Nick, G.; Gallina, V.; Sihn, W.; Monostori, L. Lead time prediction in a flow-shop environment with analytical and machine learning approaches. IFAC-Pap. 2018, 51(11), 1029–1034. [CrossRef]
  60. Dev, S.; Srivastava, R. Experimental investigation and optimization of the additive manufacturing process through AI-based hybrid statistical approaches. Prog. Addit. Manuf. 2025, 10(1), 107–126. [CrossRef]
  61. Khdoudi, A.; Masrour, T.; El Hassani, I.; El Mazgualdi, C. A Deep-Reinforcement-Learning-Based Digital Twin for Manufacturing Process Optimization. Systems 2024, 12(2), 38. [CrossRef]
  62. Dornheim, J.; Link, N.; Gumbsch, P. Model-free Adaptive Optimal Control of Episodic Fixed-horizon Manufacturing Processes Using Reinforcement Learning. Int. J. Control Autom. Syst. 2020, 18(6), 1593–1604. [CrossRef]
  63. Denkena, B.; Dittrich, M.-A.; Böß, V.; Wichmann, M.; Friebe, S. Self-optimizing process planning for helical flute grinding. Prod. Eng. 2019, 13(5), 599–606. [CrossRef]
  64. Tan, Q.; Tong, Y.; Wu, S.; Li, D. Modeling, planning, and scheduling of shopfloor assembly process with dynamic cyber-physical interactions: a case study for CPS-based smart industrial robot production. Int. J. Adv. Manuf. Technol. 2019, 105(9), 3979–3989. [CrossRef]
  65. Priore, P.; Ponte, B.; Puente, J.; Gómez, A. Learning-based scheduling of flexible manufacturing systems using ensemble methods. Comput. Ind. Eng. 2018, 126, 282–291. [CrossRef]
  66. Jones, D.; Snider, C.; Nassehi, A.; Yon, J.; Hicks, B. Characterising the Digital Twin: A systematic literature review. CIRP J. Manuf. Sci. Technol. 2020, 29, 36–52. [CrossRef]
  67. Aivaliotis, P.; Georgoulias, K.; Chryssolouris, G. The use of Digital Twin for predictive maintenance in manufacturing. Int. J. Comput. Integr. Manuf. 2019, 32(11), 1067–1080. [CrossRef]
  68. Pan, Y.; Kang, S.; Kong, L.; Wu, J.; Yang, Y.; Zuo, H. Remaining useful life prediction methods of equipment components based on deep learning for sustainable manufacturing: a literature review. Artif. Intell. Eng. Des. Anal. Manuf. 2025, 39, 4. [CrossRef]
  69. Xu, Q.; Ali, S.; Yue, T. Digital Twin-based Anomaly Detection with Curriculum Learning in Cyber-physical Systems. ACM Trans. Softw. Eng. Methodol. 2023, 32(5), 1–32. [CrossRef]
  70. Schena, L.; Marques, P.A.; Poletti, R.; Ahizi, S.; Van Den Berghe, J.; Mendez, M.A. Reinforcement Twinning: From digital twins to model-based reinforcement learning. J. Comput. Sci. 2024, 82, 102421. [CrossRef]
  71. Kreuzer, T.; Papapetrou, P.; Zdravkovic, J. Artificial intelligence in digital twins—A systematic literature review. Data Knowl. Eng. 2024, 151, 102304. [CrossRef]
  72. Liu, J.; Lu, X.; Zhou, Y.; Cui, J.; Wang, S.; Zhao, Z. Design of Photovoltaic Power Station Intelligent Operation and Maintenance System Based on Digital Twin. In: Proceedings of the 2021 6th International Conference on Robotics and Automation Engineering (ICRAE), Guangzhou, China, 20–22 Nov 2021; IEEE: pp. 206–211. [CrossRef]
  73. Livera, A.; Paphitis, G.; Pikolos, L.; Papadopoulos, I.; Montes-Romero, J.; Lopez-Lorente, J.; Makrides, G.; Sutterlueti, J.; Georghiou, G.E. Intelligent Cloud-Based Monitoring and Control Digital Twin for Photovoltaic Power Plants. In: Proceedings of the 2022 IEEE 49th Photovoltaics Specialists Conference (PVSC), Philadelphia, PA, USA, 5–10 June 2022; IEEE: pp. 0267–0274. [CrossRef]
  74. Ma, Y.; Kassler, A.; Ahmed, B.S.; Krakhmalev, P.; Thore, A.; Toyser, A.; Lindbäck, H. Using Deep Reinforcement Learning for Zero Defect Smart Forging. In: Ng, A.H.C., Syberfeldt, A., Högberg, D., Holm, M., Eds.; Advances in Transdisciplinary Engineering; IOS Press: 2022. [CrossRef]
  75. Chhetri, S.R.; Faezi, S.; Canedo, A.; Faruque, M.A.A. QUILT: quality inference from living digital twins in IoT-enabled manufacturing systems. In: Proceedings of the International Conference on Internet of Things Design and Implementation, Montreal, QC, Canada, 15–18 April 2019; ACM: pp. 237–248. [CrossRef]
  76. Maia, E.; Wannous, S.; Dias, T.; Praça, I.; Faria, A. Holistic Security and Safety for Factories of the Future. Sensors 2022, 22(24), 9915. [CrossRef]
  77. Barriga, R.; Romero, M.; Nettleton, D.; Hassan, H. Advanced data modeling for industrial drying machine energy optimization. J. Supercomput. 2022, 78(15), 16820–16840. [CrossRef]
  78. Bansal, R.; Khanesar, M.A.; Branson, D. Ant Colony Optimization Algorithm for Industrial Robot Programming in a Digital Twin. In: Proceedings of the 2019 25th International Conference on Automation and Computing (ICAC), Lancaster, UK, 5–7 Sep 2019; IEEE: pp. 1–5. [CrossRef]
  79. Liu, Y.; Xu, H.; Liu, D.; Wang, L. A digital twin-based sim-to-real transfer for deep reinforcement learning-enabled industrial robot grasping. Robot. Comput. Integr. Manuf. 2022, 78, 102365. [CrossRef]
  80. Li, J.; Pang, D.; Zheng, Y.; Guan, X.; Le, X. A flexible manufacturing assembly system with deep reinforcement learning. Control Eng. Pract. 2022, 118, 104957. [CrossRef]
  81. Kobayashi, K.; Alam, S.B. Explainable, Interpretable & Trustworthy AI for Intelligent Digital Twin: Case Study on Remaining Useful Life. arXiv 2023, arXiv:2301.06676. Version 2. [CrossRef]
  82. Mehrabi, A.; Yari, K.; Van Driel, W.D.; Poelma, R.H. AI-Driven Digital Twin for Health Monitoring of Wide Band Gap Power Semiconductors. In: Proceedings of the 2024 IEEE 10th Electronics System-Integration Technology Conference (ESTC), Berlin, Germany, 11–13 Sep 2024; IEEE: pp. 1–8. [CrossRef]
  83. Li, Y. Fault Prediction and Diagnosis System for Large-diameter Auger Rigs Based on Digital Twin and BP Neural Network. In: Proceedings of the 2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Beijing, China, 7–9 Jan 2022; IEEE: pp. 523–527. [CrossRef]
  84. Guo, K.; Wan, X.; Liu, L.; Gao, Z.; Yang, M. Fault Diagnosis of Intelligent Production Line Based on Digital Twin and Improved Random Forest. Appl. Sci. 2021, 11(16), 7733. [CrossRef]
  85. Chakrabarti, A.; Sukumar, R.P.; Jarke, M.; Rudack, M.; Buske, P.; Holly, C. Efficient Modeling of Digital Shadows for Production Processes: A Case Study for Quality Prediction in High Pressure Die Casting Processes. In: Proceedings of the 2021 IEEE 8th International Conference on Data Science and Advanced Analytics (DSAA), Porto, Portugal, 6–9 Oct 2021; IEEE: pp. 1–9. [CrossRef]
  86. Boulfani, F.; Gendre, X.; Ruiz-Gazen, A.; Salvignol, M. Anomaly detection for aircraft electrical generator using machine learning in a functional data framework. In: Proceedings of the 2020 Global Congress on Electrical Engineering (GC-ElecEng), Valencia, Spain, 4–6 Nov 2020; IEEE: pp. 27–32. [CrossRef]
  87. Xu, Y.; Sun, Y.; Liu, X.; Zheng, Y. A Digital-Twin-Assisted Fault Diagnosis Using Deep Transfer Learning. IEEE Access 2019, 7, 19990–19999. [CrossRef]
  88. Ren, Z.; Wan, J.; Deng, P. Machine-Learning-Driven Digital Twin for Lifecycle Management of Complex Equipment. IEEE Trans. Emerg. Top. Comput. 2022, 10(1), 9–22. [CrossRef]
  89. Xiong, M.; Wang, H.; Fu, Q.; Xu, Y. Digital twin–driven aero-engine intelligent predictive maintenance. Int. J. Adv. Manuf. Technol. 2021, 114(11–12), 3751–3761. [CrossRef]
  90. Piltan, F.; Toma, R.N.; Shon, D.; Im, K.; Choi, H.-K.; Yoo, D.-S.; Kim, J.-M. Strict-Feedback Backstepping Digital Twin and Machine Learning Solution in AE Signals for Bearing Crack Identification. Sensors 2022, 22(2), 539. [CrossRef]
  91. Tancredi, G.P.; Vignali, G.; Bottani, E. Integration of Digital Twin, Machine-Learning and Industry 4.0 Tools for Anomaly Detection: An Application to a Food Plant. Sensors 2022, 22(11), 4143. [CrossRef]
  92. Melesse, T.Y.; Bollo, M.; Pasquale, V.D.; Centro, F.; Riemma, S. Machine Learning-Based Digital Twin for Monitoring Fruit Quality Evolution. Procedia Comput. Sci. 2022, 200, 13–20. [CrossRef]
  93. Alexopoulos, K.; Nikolakis, N.; Chryssolouris, G. Digital twin-driven supervised machine learning for the development of artificial intelligence applications in manufacturing. Int. J. Comput. Integr. Manuf. 2020, 33(5), 429–439. [CrossRef]
  94. Wang, Q.; Jiao, W.; Zhang, Y. Deep learning-empowered digital twin for visualized weld joint growth monitoring and penetration control. J. Manuf. Syst. 2020, 57, 429–439. [CrossRef]
  95. Angin, P.; Anisi, M.H.; Göksel, F.; Gürsoy, C.; Büyükgülcü, A. AgriLoRa: A Digital Twin Framework for Smart Agriculture. J. Wirel. Mob. Netw. Ubiquitous Comput. Dependable Appl. 2020, 11(4), 77–96. [CrossRef]
  96. Zhang, Q.; Li, Y.; Lim, E.; Sun, J. Real Time Object Detection in Digital Twin with Point-Cloud Perception for a Robotic Manufacturing Station. In: 2022 27th International Conference on Automation and Computing (ICAC), Bristol, United Kingdom, 2022; pp. 1–6. [CrossRef]
  97. Zhou, X.; Xu, X.; Liang, W.; Zeng, Z.; Shimizu, S.; Yang, L.T.; Jin, Q. Intelligent Small Object Detection for Digital Twin in Smart Manufacturing With Industrial Cyber-Physical Systems. IEEE Trans. Ind. Inform. 2022, 18(2), 1377–1386. [CrossRef]
  98. Wang, Q.; Jiao, W.; Wang, P.; Zhang, Y. Digital Twin for Human-Robot Interactive Welding and Welder Behavior Analysis. IEEE/CAA J. Autom. Sin. 2021, 8(2), 334–343. [CrossRef]
  99. Chiurco, A.; Elbasheer, M.; Longo, F.; Nicoletti, L.; Solina, V. Data Modeling and ML Practice for Enabling Intelligent Digital Twins in Adaptive Production Planning and Control. Procedia Comput. Sci. 2023, 217, 1908–1917. [CrossRef]
  100. Chen, D.; Lv, Z. Artificial Intelligence Enabled Digital Twins for Training Autonomous Cars. Internet Things Cyber-Phys. Syst. 2022, 2, 31–41. [CrossRef]
  101. Shen, G.; Lei, L.; Li, Z.; Cai, S.; Zhang, L.; Cao, P.; Liu, X. Deep Reinforcement Learning for Flocking Motion of Multi-UAV Systems: Learn From a Digital Twin. IEEE Internet Things J. 2022, 9(13), 11141–11153. [CrossRef]
  102. Li, B.; Liu, Y.; Tan, L.; Pan, H.; Zhang, Y. Digital Twin Assisted Task Offloading for Aerial Edge Computing and Networks. IEEE Trans. Veh. Technol. 2022, 71(10), 10863–10877. [CrossRef]
  103. Urgo, M.; Terkaj, W.; Simonetti, G. Monitoring Manufacturing Systems Using AI: A Method Based on a Digital Factory Twin to Train CNNs on Synthetic Data. CIRP J. Manuf. Sci. Technol. 2024, 50, 249–268. [CrossRef]
  104. Gallala, A.; Kumar, A.A.; Hichri, B.; Plapper, P. Digital Twin for Human–Robot Interactions by Means of Industry 4.0 Enabling Technologies. Sensors 2022, 22(13), 4950. [CrossRef]
  105. Wang, T.; Cheng, J.; Yang, Y.; Esposito, C.; Snoussi, H.; Tao, F. Adaptive Optimization Method in Digital Twin Conveyor Systems via Range-Inspection Control. IEEE Trans. Autom. Sci. Eng. 2022, 19(2), 1296–1304. [CrossRef]
  106. Li, H.; Liu, Z.; Yuan, W.; Chen, G.; Chen, X.; Yang, Y.; Peng, J. The Digital Twin Model of Chemical Production Systems in Smart Factories: A Case Study. In: 2021 IEEE 23rd International Conference on High Performance Computing & Communications (HPCC); 7th Int. Conf. on Data Science & Systems; 19th Int. Conf. on Smart City; 7th Int. Conf. on Dependability in Sensor, Cloud & Big Data Systems & Application (DSS/SmartCity/DependSys), Haikou, Hainan, China, 2021; pp. 1035–1041. [CrossRef]
  107. Min, Q.; Lu, Y.; Liu, Z.; Su, C.; Wang, B. Machine Learning Based Digital Twin Framework for Production Optimization in Petrochemical Industry. Int. J. Inf. Manag. 2019, 49, 502–519. [CrossRef]
  108. Huang, B.; Wang, D.; Li, H.; Zhao, C. Network Selection and QoS Management Algorithm for 5G Converged Shipbuilding Network Based on Digital Twin. In: 2022 10th International Conference on Information and Education Technology (ICIET), Matsue, Japan, 2022; pp. 403–408. [CrossRef]
  109. Dos Santos, C.H.; Gabriel, G.T.; Do Amaral, J.V.S.; Montevechi, J.A.B.; De Queiroz, J.A. Decision-Making in a Fast Fashion Company in the Industry 4.0 Era: A Digital Twin Proposal to Support Operational Planning. Int. J. Adv. Manuf. Technol. 2021, 116(5–6), 1653–1666. [CrossRef]
  110. Latif, H.; Shao, G.; Starly, B. A Case Study of Digital Twin for a Manufacturing Process Involving Human Interactions. In: 2020 Winter Simulation Conference (WSC), Orlando, FL, USA, 2020; pp. 2659–2670. [CrossRef]
  111. Qiu, T.; Chi, J.; Zhou, X.; Ning, Z.; Atiquzzaman, M.; Wu, D.O. Edge Computing in Industrial Internet of Things: Architecture, Advances and Challenges. IEEE Commun. Surv. Tutor. 2020, 22(4), 2462–2488. [CrossRef]
  112. Nain, G.; Pattanaik, K.K.; Sharma, G.K. Towards Edge Computing in Intelligent Manufacturing: Past, Present and Future. J. Manuf. Syst. 2022, 62, 588–611. [CrossRef]
  113. Montes-Sánchez, J.M.; Uwate, Y.; Nishio, Y.; Vicente-Díaz, S.; Jiménez-Fernández, Predictive Maintenance Edge Artificial Intelligence Application Study Using Recurrent Neural Networks for Early Aging Detection in Peristaltic Pumps. IEEE Trans. Reliab. 2024, 1–15. [CrossRef]
  114. Parikh, S.; Dave, D.; Patel, R.; Doshi, N. Security and Privacy Issues in Cloud, Fog and Edge Computing. Procedia Comput. Sci. 2019, 160, 734–739. [CrossRef]
  115. Maciel, P.; Dantas, J.; Melo, C.; Pereira, P.; Oliveira, F.; Araujo, J.; Matos, R. A Survey on Reliability and Availability Modeling of Edge, Fog, and Cloud Computing. J. Reliab. Intell. Environ. 2022, 8(3), 227–245. [CrossRef]
  116. Artiushenko, V.; Lang, S.; Lerez, C.; Reggelin, T.; Hackert-Oschätzchen, M. Resource-Efficient Edge AI Solution for Predictive Maintenance. Procedia Comput. Sci. 2024, 232, 348–357. [CrossRef]
  117. Texas Instruments. TMS320F28P55x Real-Time Microcontrollers Datasheet (Rev. B). 2024. https://www.ti.com/lit/ds/symlink/tms320f28p550sj.pdf.
  118. What’s New in Artificial Intelligence From the 2023 Gartner Hype Cycle™. https://www.gartner.com/en/articles/what-s-new-in-artificial-intelligence-from-the-2023-gartner-hype-cycle.
  119. Bringing AI to the Device: Edge AI Chips Come Into Their Own. https://www2.deloitte.com/us/en/insights/industry/technology/technology-media-and-telecom-predictions/2020/ai-chips.html.
  120. Hsu, H.-Y.; Srivastava, G.; Wu, H.-T.; Chen, M.-Y. Remaining Useful Life Prediction Based on State Assessment Using Edge Computing on Deep Learning. Comput. Commun. 2020, 160, 91–100. [CrossRef]
  121. Teoh, Y.K.; Gill, S.S.; Parlikad, A.K. IoT and Fog-Computing-Based Predictive Maintenance Model for Effective Asset Management in Industry 4.0 Using Machine Learning. IEEE Internet Things J. 2023, 10(3), 2087–2094. [CrossRef]
  122. Zhang, T.; Ding, B.; Zhao, X.; Liu, G.; Pang, Z. LearningADD: Machine Learning Based Acoustic Defect Detection in Factory Automation. J. Manuf. Syst. 2021, 60, 48–58. [CrossRef]
  123. Milić, S.D.; Miladinović, N.M.; Rakić, A. A Wayside Hotbox System with Fuzzy and Fault Detection Algorithms in IIoT Environment. Control Eng. Pract. 2020, 104, 104624. [CrossRef]
  124. Hu, L.; Miao, Y.; Wu, G.; Hassan, M.M.; Humar, I. iRobot-Factory: An Intelligent Robot Factory Based on Cognitive Manufacturing and Edge Computing. Future Gener. Comput. Syst. 2019, 90, 569–577. [CrossRef]
  125. Huang, H.; Ding, S.; Zhao, L.; Huang, H.; Chen, L.; Gao, H.; Ahmed, S.H. Real-Time Fault Detection for IIoT Facilities Using GBRBM-Based DNN. IEEE Internet Things J. 2020, 7(7), 5713–5722. [CrossRef]
  126. Zhao, X.; Lv, K.; Zhang, Z.; Zhang, Y.; Wang, Y. A Multi-Fault Diagnosis Method of Gear-Box Running on Edge Equipment. J. Cloud Comput. 2020, 9(1), 58. [CrossRef]
  127. Kim, D.; Yang, H.; Chung, M.; Cho, S.; Kim, H.; Kim, M.; Kim, K.; Kim, E. Squeezed Convolutional Variational AutoEncoder for Unsupervised Anomaly Detection in Edge Device Industrial Internet of Things. In Proceedings of the 2018 Int. Conf. Inf. Comput. Technol. (ICICT), DeKalb, IL, USA, 23–25 March 2018; pp. 67–71. https://ieeexplore.ieee.org/document/8356842/ Accessed 2025-05-24. [CrossRef]
  128. Wang, Y.; Liu, M.; Zheng, P.; Yang, H.; Zou, J. A Smart Surface Inspection System Using Faster R-CNN in Cloud-Edge Computing Environment. Adv. Eng. Inform. 2020, 43, 101037. [CrossRef]
  129. Ha, H.; Jeong, J. CNN-Based Defect Inspection for Injection Molding Using Edge Computing and Industrial IoT Systems. Appl. Sci. 2021, 11(14), 6378. [CrossRef]
  130. Zhu, Z.; Han, G.; Jia, G.; Shu, L. Modified DenseNet for Automatic Fabric Defect Detection With Edge Computing for Minimizing Latency. IEEE Internet Things J. 2020, 7(10), 9623–9636. [CrossRef]
  131. Gauttam, H.; Pattanaik, K.K.; Bhadauria, S.; Nain, G.; Prakash, P.B. An Efficient DNN Splitting Scheme for Edge-AI Enabled Smart Manufacturing. J. Ind. Inf. Integr. 2023, 34, 100481. [CrossRef]
  132. Fraga-Lamas, P.; Lopes, S.I.; Fernández-Caramés, T.M. Green IoT and Edge AI as Key Technological Enablers for a Sustainable Digital Transition towards a Smart Circular Economy: An Industry 5.0 Use Case. Sensors 2021, 21(17), 5745. [CrossRef]
  133. Yang, X.; Han, M.; Tang, H.; Li, Q.; Luo, X. Detecting Defects With Support Vector Machine in Logistics Packaging Boxes for Edge Computing. IEEE Access 2020, 8, 64002–64010. [CrossRef]
  134. Schmitt, J.; Bönig, J.; Borggräfe, T.; Beitinger, G.; Deuse, J. Predictive Model-Based Quality Inspection Using Machine Learning and Edge Cloud Computing. Adv. Eng. Inform. 2020, 45, 101101. [CrossRef]
  135. Ma, Q.; Niu, J.; Ouyang, Z.; Li, M.; Ren, T.; Li, Q. Edge Computing-Based 3D Pose Estimation and Calibration for Robot Arms. In Proceedings of the 2020 7th IEEE Int. Conf. on Cyber Security and Cloud Computing (CSCloud)/6th IEEE Int. Conf. on Edge Computing and Scalable Cloud (EdgeCom), New York, NY, USA, 22–24 August 2020; pp. 246–251. https://ieeexplore.ieee.org/document/9170983/ Accessed 2025-05-24. [CrossRef]
  136. Zhang, C.; Ji, W. Edge Computing Enabled Production Anomalies Detection and Energy-Efficient Production Decision Approach for Discrete Manufacturing Workshops. IEEE Access 2020, 8, 158197–158207. [CrossRef]
  137. Du, J.; Li, X.; Gao, Y.; Gao, L. Integrated Gradient-Based Continuous Wavelet Transform for Bearing Fault Diagnosis. Sensors 2022, 22(22), 8760. [CrossRef]
  138. Yang, D.; Karimi, H.R.; Gelman, L. A Fuzzy Fusion Rotating Machinery Fault Diagnosis Framework Based on the Enhancement Deep Convolutional Neural Networks. Sensors 2022, 22(2), 671. [CrossRef]
  139. Janssens, O.; Loccufier, M.; Van Hoecke, S. Thermal Imaging and Vibration-Based Multisensor Fault Detection for Rotating Machinery. IEEE Trans. Ind. Inform. 2019, 15(1), 434–444. [CrossRef]
  140. Duquesnoy, M.; Liu, C.; Dominguez, D.Z.; Kumar, V.; Ayerbe, E.; Franco, A.A. Machine Learning-Assisted Multi-Objective Optimization of Battery Manufacturing from Synthetic Data Generated by Physics-Based Simulations. arXiv 2022, abs/2205.01621. https://arxiv.org/abs/2205.01621 Accessed 2025-05-20. [CrossRef]
  141. Khosravi, H.; Farhadpour, S.; Grandhi, M.; Raihan, A.S.; Das, S.; Ahmed, I. Strategic Data Augmentation with CTGAN for Smart Manufacturing: Enhancing Machine Learning Predictions of Paper Breaks in Pulp-and-Paper Production. arXiv 2023, abs/2311.09333. https://arxiv.org/abs/2311.09333 Accessed 2025-05-20. [CrossRef]
  142. Liaskovska, S.; Tyskyi, S.; Martyn, Y.; Augousti, A.T.; Kulyk, V. Systematic Generation and Evaluation of Synthetic Production Data for Industry 5.0 Optimization. Technologies 2025, 13(2), 84. [CrossRef]
  143. Singh, R.; Gill, S.S. Edge AI: A Survey. Internet Things Cyber-Phys. Syst. 2023, 3, 71–92. [CrossRef]
  144. Lim, W.Y.B.; Luong, N.C.; Hoang, D.T.; Jiao, Y.; Liang, Y.-C.; Yang, Q.; Niyato, D.; Miao, C. Federated Learning in Mobile Edge Networks: A Comprehensive Survey. arXiv 2019, abs/1909.11875. https://arxiv.org/abs/1909.11875. Accessed 2025-05-20. [CrossRef]
  145. Texas Instruments. TIDA-010955 Arc Fault Detection Using Embedded AI Models Reference Design. Texas Instruments 2024. https://www.ti.com/tool/TIDA-010955. Accessed 2025-05-21.
  146. Texas Instruments. Motor Fault Detection Using Embedded AI Models. Texas Instruments 2024. https://www.ti.com/lit/pdf/spracw8. Accessed 2025-05-21.
  147. Xu, C.; Zhu, G. Intelligent Manufacturing Lie Group Machine Learning: Real-Time and Efficient Inspection System Based on Fog Computing. J. Intell. Manuf. 2021, 32(1), 237–249. [CrossRef]
  148. Guerra, R.H.; Quiza, R.; Villalonga, A.; Arenas, J.; Castano, F. Digital Twin-Based Optimization for Ultraprecision Motion Systems With Backlash and Friction. IEEE Access 2019, 7, 93462–93472. [CrossRef]
  149. ABB. ABB Ability™ Digital Powertrain—Condition Monitoring of Rotating Equipment Fitted with ABB Ability™ Smart Sensors (EN). https://new.abb.com/service/motion/data-and-advisory-services/condition-monitoring-for-rotating-equipment.
  150. ABB. ABB Ability Predictive Maintenance for Grinding. https://new.abb.com/mining/services/advanced-digital-services/predictive-maintenance-grinding. Accessed 2025-05-21.
  151. IBM. IBM Maximo Application Suite (2025). https://www.ibm.com/docs/en/masv-and-l/cd?topic=overview-maximo-application-suite-technical. Accessed 2025-05-21.
  152. PTC. ThingWorx: Industrial IoT Software | IIoT Platform. https://www.ptc.com/en/products/thingworx. Accessed 2025-05-21.
  153. Microsoft Azure. Azure IoT – Internet of Things Platform. https://azure.microsoft.com/en-us/solutions/iot. Accessed 2025-05-21.
  154. Uptake. Predictive Maintenance. https://uptake.com/. Accessed 2025-05-21.
  155. MathWorks. Predictive Maintenance Toolbox. https://www.mathworks.com/products/predictive-maintenance.html. Accessed 2025-05-21.
  156. Cognex. VisionPro Software. https://www.cognex.com/products/machine-vision/vision-software/visionpro-software. Accessed 2025-05-21.
  157. KEYENCE. Vision Systems | KEYENCE America. https://www.keyence.com/products/vision/vision-sys/. Accessed 2025-05-21.
  158. NI. What Is the NI Vision Development Module. https://www.ni.com/en/shop/data-acquisition-and-control/add-ons-for-data-acquisition-and-control/what-is-vision-development-module.html. Accessed 2025-05-21.
  159. Integrys. Matrox Imaging Library (MIL) Developing Machine Vision. https://integrys.com/product/matrox-imaging-library-mil/. Accessed 2025-05-21.
  160. ZEISS. ZEISS PiWeb | Quality Data Management. https://www.zeiss.com/metrology/us/software/zeiss-piweb.html. Accessed 2025-05-21.
  161. Mao, M.; Hong, M. YOLO Object Detection for Real-Time Fabric Defect Inspection in the Textile Industry: A Review of YOLOv1 to YOLOv11. Sensors 2025, 25(7), 2270. [CrossRef]
  162. Razavi, M.; Mavaddati, S.; Koohi, H. ResNet Deep Models and Transfer Learning Technique for Classification and Quality Detection of Rice Cultivars. Expert Syst. Appl. 2024, 247, 123276. [CrossRef]
  163. Yadao, G.G.; Julkipli, O.M.Y.; Manlises, C.O. Performance Analysis of EfficientNet for Rice Grain Quality Control—An Evaluation against YOLOv7 and YOLOv8. In: 2024 7th International Conference on Information and Computer Technologies (ICICT), pp. 93–98. IEEE, Honolulu, HI, USA (2024). https://ieeexplore.ieee.org/document/10541526/. Accessed 2025-05-21. [CrossRef]
  164. Amazon Web Services. Computer Vision SDK - AWS Panorama. https://aws.amazon.com/panorama/. Accessed 2025-05-21.
  165. Edge Impulse. The Leading Edge AI Platform. https://edgeimpulse.com/. Accessed 2025-05-21.
  166. Coral. https://coral.ai/. Accessed 2025-05-21.
  167. Siemens. Plant Simulation Software | Siemens Software. https://plm.sw.siemens.com/en-US/tecnomatix/plant-simulation-software/. Accessed 2025-05-21.
  168. Rockwell Automation. Arena Simulation Software. https://www.rockwellautomation.com/de-de/products/software/arena-simulation.html. Accessed 2025-05-21.
  169. AnyLogic. Simulation Modeling Software Tools & Solutions. https://www.anylogic.com/. Accessed 2025-05-21.
  170. AspenTech. Aspen Plus | Leading Process Simulation Software. https://www.aspentech.com/en/products/engineering/aspen-plus. Accessed 2025-05-21.
  171. AspenTech. Aspen HYSYS | Process Simulation Software. https://www.aspentech.com/en/products/engineering/aspen-hysys. Accessed 2025-05-21.
  172. Ansys. Ansys Twin Builder | Create and Deploy Digital Twin Models. https://www.ansys.com/products/digital-twin/ansys-twin-builder. Accessed 2025-05-21.
  173. Schmid, J.; Teichert, K.; Chioua, M.; Schindler, T.; Bortz, M. Simulation and Optimal Control of the Williams-Otto Process Using Pyomo. arXiv 2020. https://arxiv.org/abs/2004.07614. Accessed 2025-05-21. [CrossRef]
  174. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 2623–2631. ACM, Anchorage AK, USA (2019). https://dl.acm.org/doi/10.1145/3292500.3330701. Accessed 2025-05-21. [CrossRef]
  175. Eckman, D.J.; Henderson, S.G.; Shashaani, S. SimOpt: A Testbed for Simulation-Optimization Experiments. INFORMS J. Comput. 2023, 35(2), 495–508. [CrossRef]
  176. Raffin, A.; Hill, A.; Gleave, A.; Kanervisto, A.; Ernestus, M.; Dormann, N. Stable-Baselines3: Reliable Reinforcement Learning Implementations. J. Mach. Learn. Res. 2021, 22(268), 1–8.
  177. Ray Project. RLlib: Industry-Grade, Scalable Reinforcement Learning — Ray 2.46.0. https://docs.ray.io/en/latest/rllib/index.html. Accessed 2025-05-21.
  178. General Electric (GE). Predix Overview. https://www.gevernova.com/software/documentation/predix-platforms/PDFs/Predix%20Overview.pdf. Accessed 2025-05-21.
  179. Siemens. Insights Hub Capability Packages Product Sheet. (2024). https://plm.sw.siemens.com/en-US/insights-hub/resources/product-sheets/. Accessed 2025-05-21.
  180. Intel. Intel® Distribution of OpenVINO™ Toolkit. https://www.intel.com/content/www/us/en/developer/tools/openvino-toolkit/overview.html. Accessed 2025-05-21.
  181. NVIDIA. TensorRT. https://developer.nvidia.com/tensorrt. Accessed 2025-05-21.
  182. NVIDIA. Dynamo. https://developer.nvidia.com/dynamo. Accessed 2025-05-21.
  183. Amazon Web Services. IoT Edge, Open Source Edge - AWS IoT Greengrass. https://aws.amazon.com/greengrass/. Accessed 2025-05-21.
  184. Microsoft Azure. Digital Twins – Modeling and Simulations. https://azure.microsoft.com/en-us/products/digital-twins. Accessed 2025-05-21.
Figure 1. Classification of Machine Learning Algorithm.
Figure 1. Classification of Machine Learning Algorithm.
Preprints 165103 g001
Figure 2. Schematic diagram of a digital twin with an AI component.[66]
Figure 2. Schematic diagram of a digital twin with an AI component.[66]
Preprints 165103 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated