Submitted:
25 August 2025
Posted:
26 August 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
- Scalability and computational complexity in large, high-dimensional environments;
- Robustness and adaptability to uncertainty and real-world disruptions;
- Integration with digitalization (IIoT, cloud/edge platforms, and cyber-physical systems).
2. Scalability and Computational Complexity
2.1. The Combinatorial Nature of Industrial Scheduling
2.2. Recent Methodological Developments
2.2.1. Metaheuristics and Hybrid Algorithms
- Genetic algorithms (GAs) and memetic hybrids. Modern GA variants integrate local search, path relinking, or destroy-and-repair moves to accelerate convergence on very large instances and complex shop settings; hybrids tuned for industrial-scale unrelated/parallel machines and sequence-dependent setups are increasingly common. Representative examples show GA+local-search hybrids scaling to hundreds of machines/jobs while retaining solution quality (Blum & Roli, 2003; Ferreira et al., 2022).
- Simulated annealing (SA) and tabu search (TS). Classical SA/TS ideas—probabilistic uphill moves and adaptive memory—continue to underpin strong baselines. Contemporary implementations pair TS with constraint-aware neighborhoods or embed instance-specific neighborhoods learned from data to reduce cycling and improve the intensification/diversification balance. Conceptual surveys still frame best practices for hybrid design (Blum & Roli, 2003).
- Large-neighborhood search (LNS) and learning-enhanced LNS. LNS “destroy-and-repair” is particularly effective under tight timing constraints. Recent neural LNS variants use deep networks (often graph-based) to propose destroy sets or repair decisions, yielding large speed/quality gains across combinatorial problems and increasingly in scheduling (Hottung & Tierney, 2022).
- Hyper-heuristics (rule selection/generation). Instead of solving a schedule directly, hyper-heuristics learn which heuristic to deploy when. A recent line uses deep reinforcement learning (DRL) hyper-heuristics to select operators on-the-fly, improving generalizability across shop configurations (Panzer & Bender, 2022; Smit et al., 2024).
- Learning-assisted parameter control & initialization. Reviews highlight the benefit of machine-learned parameter schedules, warm-starts, and population initializers to stabilize metaheuristics on high-variance instance distributions—especially for multi-objective settings (Bengio et al., 2021).
2.2.2. Decomposition and Parallelization
- Logic-Based Benders Decomposition (LBBD). LBBD separates combinatorial assignment/sequence decisions (handled by CP/MIP/heuristics) from schedule-feasibility subproblems, iteratively exchanging powerful logic cuts. Recent papers demonstrate strong performance on flexible/distributed job-shops and highlight modeling patterns and cut design that make LBBD competitive on industrial testbeds (Naderi et al., 2022; Juvin et al., 2023).
- Hierarchical/rolling-horizon schemes. Multi-level decompositions—e.g., plan vs. schedule, coarse time windows vs. fine sequencing—remain essential when the full horizon is prohibitive. Newer work integrates domain constraints from chemical/process systems and uses decomposition to keep digital-twin/CP models responsive at runtime; learning-guided rolling horizons are emerging to adapt window sizes and priorities on the fly (Liñán & Méndez, 2024; Forbes & Kelly, 2024).
- Dantzig–Wolfe/column generation and branch-and-price. Modern implementations in open frameworks (e.g., SCIP/GCG) expose decomposition hooks, enabling practitioners to combine exact and heuristic components and to scale on shared/distributed memory (Bestuzheva et al., 2021).
- Parallel solver ecosystems. Documented advances from 2001–2020 show order-of-magnitude speedups from algorithmic and hardware progress; contemporary suites include UG, a unified framework for parallelizing branch-and-bound/price/cut across cores and clusters. These capabilities benefit both pure MIP/CP scheduling and hybrid MH+MIP workflows (Koch et al., 2022; Bestuzheva et al., 2021).
2.2.3. AI-Driven and Data-Driven Methods
-
Deep reinforcement learning (DRL) for dispatching and end-to-end scheduling.
- ○
- Learned dispatching rules. GNN-based DRL learns to choose the next operation/machine given a disjunctive-graph state, outperforming hand-crafted rules and transferring to larger instances (Zhang et al., 2020).
- ○
- Systematic evidence (2022–2024). Surveys map model choices (GNNs, attention/transformers), training regimes, robustness/generalization gaps, and industrial case studies—useful for selecting architectures and evaluation protocols (Panzer & Bender, 2022; Zhang et al., 2024; Smit et al., 2024).
- ○
- Digital-twin–in-the-loop training and deployment. Coupling DRL with twins improves sample efficiency and safety prior to shop-floor rollout (Zhang et al., 2022).
-
Learning-augmented optimization (L4CO) for exact solvers.
- ○
- Cut selection via RL/imitation. DRL policies for cutting-plane selection in MILP and successors (2020–2024) reduce nodes/time across instance families; these techniques directly accelerate large MIP/CP models of scheduling (Tang et al., 2020; Huang et al., 2022; Wang et al., 2023).
- ○
- Learned branching/diving and node selection. Neural policies guide B&B traversal and primal heuristics, improving primal-dual gaps and anytime behavior on real MIP workloads (Nair et al., 2020; Bengio et al., 2021).
- Neural Large-Neighborhood Search (Neural-LNS). Deep networks propose destroy/repair actions within LNS, maintaining metaheuristic scalability while injecting structural priors (Hottung & Tierney, 2022).
- Supervised and interpretable learning of rules/policies. Data-driven mining of dispatching rules from near-optimal schedules and interpretable learned rules (e.g., sparse/structured models) offer transparent alternatives for regulated environments—often used to warm-start DRL or guide MH neighborhoods (Ferreira et al., 2022).
- Surrogate-assisted optimization. ML surrogates approximate expensive objective/simulation evaluations (e.g., multi-objective, dynamic shops), enabling deeper search within fixed time budgets and stabilizing online rescheduling (Ferreira et al., 2022; Panzer & Bender, 2022).
- Foundation-model ideas (early stage). “LLMs as optimizers” (OPRO) and LLM-guided search/planning are being tested as meta-controllers—suggesting heuristic templates or operator sequences that a solver or metaheuristic then refines. While nascent, this strand aims at zero-/few-shot generalization across plants and products (Yang et al., 2024; Bengio et al., 2021).
2.3. Industrial Impact
3. Robustness and Adaptability to Uncertainty
3.1. The Prevalence of Uncertainty in Industrial Scheduling
3.2. Recent Methodological Developments
3.2.1. Robust Optimization
- Min–max and min–max regret formulations. These guard against worst-case or worst-regret scenarios—useful where delivery penalties or rework costs are high (Aissi et al., 2009). While conservative, recent practice tunes uncertainty budgets to balance robustness and performance, often informed by empirical variance estimates extracted from shop data (Bertsimas & Sim, 2004).
- Adjustable robust optimization (ARO). Defers part of the decision (e.g., dispatching, batching) until information is revealed, improving adaptability versus static designs (Ben-Tal et al., 2004). Rolling-horizon ARO for job shops with uncertain processing times demonstrates strong performance under continuous disturbances (Cohen et al., 2023).
- Interval/set-based uncertainty. Interval activity durations and release dates yield tractable robust counterparts and are attractive in regulated or contract-driven environments; hybrid robust approaches for projects exemplify this trend (Bruni et al., 2017).
- Learning-in-the-loop robust models. Robust parameters (e.g., uncertainty budgets, scenario weights) can be calibrated from historical trace data or forecasts and periodically retuned; neural surrogates speed robust evaluation when embedded inside metaheuristics or rolling-horizon loops (Zhang et al., 2022).
3.2.2. Stochastic and Probabilistic Modeling
- Chance-constrained scheduling. Constraints (e.g., due-date adherence) are enforced with high probability, enabling explicit trade-offs between service levels and efficiency (Birge & Louveaux, 2011). In data-rich plants, estimated distributions are kept up to date from streaming data and predictive models.
- Markov decision processes (MDP). MDP formulations capture sequential uncertainty and state transitions. For job-shop settings with stochastic processing times, MDPs provide a principled foundation and also underpin modern DRL policies (Zhang et al., 2017; Puterman, 2005).
- Simulation-based evaluation and design. Monte Carlo/discrete-event simulation (DES) remains essential when analytic tractability is limited. It supports proactive design of robust schedules, stress-tests rollout policies, and serves as a safe training ground for learning-based controllers (Vieira et al., 2003; Mönch et al., 2013).
3.2.3. Real-Time, Predictive, and Reactive Scheduling
- Rescheduling and repair algorithms. Minimal-perturbation repairs stabilize operations after disruptions, reducing shop-floor turbulence. Frameworks and taxonomies remain highly relevant (Vieira et al., 2003; Ouelhadj & Petrovic, 2009), and are increasingly combined with learned predictors of disruption impact to prioritize repairs.
- Rolling-horizon and event-driven updates. Periodic or event-triggered reoptimization integrates naturally with MES/ERP. State-of-practice implementations use hierarchical decompositions and fast heuristics/MIP models, often parallelized, to refresh plans at high cadence (Vieira et al., 2003; Weng et al., 2022).
- Predictive analytics and machine learning. Supervised models forecast delays, failures, and congestion; DRL agents learn dispatching policies that generalize across shop states. Reviews synthesize model choices (GNNs, attention/transformers), training regimes, and robustness/generalization gaps (Serrano-Ruiz et al., 2021; Zhang et al., 2024).
- Digital-twin-in-the-loop decision-making. Twins provide high-fidelity simulators for safe testing and sample-efficient training/deployment of real-time policies (Zhang et al., 2020; Zhang et al., 2022).
- Multi-agent and self-organizing control. Decentralized agent-based frameworks enhance resilience by localizing decisions while coordinating globally through negotiation/market or contract-net mechanisms—well aligned with cyber-physical production systems (Leitão et al., 2016a; Seitz et al., 2021).
- End-to-end AI stacks at scale. In practice, the strongest systems are hybrids: fast decomposed MIP/CP or robust metaheuristics at the core, augmented by DRL policies, learned repair operators, neural surrogates, and digital twins to navigate vast decision spaces under tight time limits (Lee & Lee, 2022; Zhang et al., 2024).
3.3. Industrial Impact
4. Integration with Digitalization and Industry 4.0
4.1. Industrial Scheduling in the Age of Digital Transformation
- rapidly process high-frequency streaming data from sensors and MES/ERP logs,
- interact with intelligent machines and human operators in collaborative CPSs, and
- adapt autonomously to both predicted and unforeseen disruptions.
4.2. Recent Methodological Developments
4.2.1. Data-Driven Scheduling and Real-Time Data Integration
- Sensor-Enabled, Closed-Loop Scheduling. Modern shop floors, equipped with IIoT sensors and CPSs, continuously generate streams of data on machine status, job progress, and environmental conditions. Scheduling algorithms can now operate in closed-loop mode, where feedback from the shop floor directly drives updates to production plans (Wang et al., 2016; Rauch et al., 2020). These approaches improve agility but also raise challenges in data quality assurance, latency management, and interoperability with legacy systems. Emerging solutions apply streaming analytics and lightweight deep models at the edge to process sensor inputs in milliseconds.
- Digital Twin-Based Scheduling. Digital twins (DTs)—virtual replicas of physical systems—are increasingly central to scheduling in Industry 4.0. DTs mirror the current shop state and can simulate disruptions, evaluate dispatching rules, and test repair strategies before they are deployed on the shop floor. This enables dynamic rescheduling, what-if analysis, and proactive maintenance scheduling (Uhlemann et al., 2017; Kritzinger et al., 2018). Recent work links DTs with reinforcement learning agents, providing safe training environments where policies are stress-tested virtually before live deployment (Zhang et al., 2021; Leng et al., 2020).
- Cloud and Edge Computing for Distributed Scheduling. Cloud-based scheduling platforms offer scalable cooperative optimization, supporting multi-plant and supply-chain-level scheduling tasks with heavy computation offloaded to distributed clusters (Mourtzis & Vlachou, 2018). In contrast, edge computing brings intelligence closer to the shop floor, enabling low-latency rescheduling in response to real-time events (Lu et al., 2020). Hybrid cloud–edge architectures are gaining traction, where global optimization runs in the cloud while local edge agents handle immediate decisions, balancing responsiveness and scalability.
4.2.2. Autonomous, Intelligent, and Decentralized Scheduling
- Agent-Based and Multi-Agent Scheduling Systems: Autonomous software agents (machines, cells, workpieces) negotiate job allocations and routing independently, supporting decentralized, modular scheduling architectures aligned with flexible manufacturing systems (Leitão et al., 2016a; Giret et al., 2017). Recent advances leverage digital twins (Siatras et al., 2024) and multi-agent reinforcement learning (Xu et al., 2025) to enhance negotiation, coalition formation, and adaptive learning for global performance.
- Self-Optimizing and Adaptive Control Algorithms: Self-optimizing scheduling algorithms continuously adapt parameter values, decision rules, or objectives in light of new data or predicted disturbances (Kusiak, 2017). Deep reinforcement learning methods such as multi-agent dueling DRL (Qin et al., 2023), graph-based MARL (Zhang et al., 2023), and hierarchical MARL (Wang et al., 2025) are enabling scalable and resilient scheduling in dynamic environments.
- Emerging Architectures: Knowledge-graph-enhanced MARL (Qin & Lu, 2024), attention-based coordination (Zheng et al., 2025), and decentralized training strategies (Malucelli et al., 2025) represent next-generation paradigms, further strengthening adaptability and autonomy in Industry 4.0 scheduling.
4.2.3. Interoperability, Standardization, and Security
- Interoperable Architectures. Modern scheduling stacks integrate with heterogeneous ERP/MES/SCM ecosystems via standardized information models and open APIs. OPC UA–centric service models and Asset Administration Shell (AAS)–based dataspace connectors enable plug-and-operate exposure of machine capabilities and scheduling services across sites and partners—supporting decentralized optimization and rapid reconfiguration (Beregi et al., 2021; Neubauer et al., 2023).
- Semantically Enriched, AI-Ready Data Layers. Knowledge-graph and model-driven integration (e.g., KG-backed twins, auto-generated data collection architectures) provide a common vocabulary across planning, dispatching, and control. This boosts data quality and feature consistency for deep learning and RL schedulers, shortens data engineering cycles, and improves cross-system explainability (Wan et al., 2024; Trunzer et al., 2021).
- Security and Data Provenance. As scheduling moves onto IIoT/cloud fabrics, compliance-by-design with ICS/IIoT security baselines (e.g., IEC 62443 mappings, NIST ICS guidance) is essential. End-to-end provenance and tamper-evident audit trails—sometimes blockchain-anchored and paired with ML for predictive auditing—help ensure integrity, confidentiality, and traceability of schedule decisions and event logs across organizational boundaries (Cindrić et al., 2025; NIST, 2023; Hu et al., 2020; Umer et al., 2024).
- Data Sovereignty & Federated Collaboration (added). Dataspace-oriented integration (AAS + policy-enforced connectors) supports inter-company scheduling use cases (capacity sharing, subcontracting) while retaining usage-control over shared datasets and learned models—key for privacy-preserving, multi-party optimization (Neubauer et al., 2023).
- Operational Hardening for AI-Driven Scheduling (added). As DL/RL components enter the loop, interface standards and security controls must extend to model artifacts and pipelines (versioned data/model registries, signed inference services, and policy-aware event buses), ensuring reproducibility and trustworthy deployment in time-critical rescheduling scenarios (Beregi et al., 2021; NIST, 2023).
4.3. Industrial Impact
5. Conclusions and Research Directions
- Policy learning for real-time decisions. DRL agents trained in simulation or digital twins learn dispatching, routing, and batching policies that scale to many machines and diverse job mixes, offering competitive makespan/tardiness with tight reaction times. Centralized or multi-agent variants increasingly handle disturbances and changing shop states (Kovács et al., 2022; Zhang et al., 2022; Wang et al., 2021).
- Generalization and transfer. Graph and attention models encode precedence, resource compatibilities, and machine–job relations, enabling transfer across families of instances and faster adaptation to new products or line configurations (Cappart et al., 2021; Peng et al., 2021; Wang et al., 2024).
- Perception-to-schedule loops. CNN/RNN/LSTM pipelines for predictive maintenance and anomaly detection feed early warnings to schedulers, enabling proactive repair policies and fewer bottlenecks by aligning maintenance windows with production plans (Bampoula et al., 2021).
- Interpretability and assurance. Black-box policies face scrutiny in regulated and safety-critical operations. Tooling for XAI/XRL, post-hoc rationales, counterfactuals, and certifiable robustness remains underused in scheduling, yet is increasingly feasible (Milani et al., 2024).
- Data quality and benchmarks. Many plants lack curated, labeled datasets for learning and objective comparison. Open, standardized benchmarks (including realistic simulators and DT-backed logs) are essential to measure progress and reproducibility (Parente et al., 2020; Gao et al., 2024).
- Legacy integration and lifecycle MLOps. Industrial IT/OT landscapes demand hardened interfaces (model registries, signed inference, versioned features), standardized semantics, and zero-downtime rollout/rollback for policies—especially when rescheduling is time-critical (Abdel-Aty et al., 2022; NIST, 2023).
- Robustness and safety. Policies must remain stable under distribution shift, sensor noise, or partial outages. Methods from robust and safe RL—risk-sensitive training, certified bounds, disturbance/adversary models—should be brought into the scheduling loop with plant-level validation (Moos et al., 2022).
- Human-in-the-loop. Operators and planners bring tacit knowledge and risk judgments. Practical systems will blend human guidance with learned policies—e.g., learning from interventions, preference feedback, or human-authored constraints—to ensure actionable, trusted decisions (Mourtzis, 2022).
- Interpretable and certifiable neural scheduling (XRL, policy simplification, safety monitors) with plant-ready evidence artifacts (Milani et al., 2024).
- Open datasets, simulators, and DT-based benchmarks for dynamic shop floors (events, breakdowns, product changeovers), enabling apples-to-apples evaluation and reproducibility (Parente et al., 2020; Kovács et al., 2022).
- Seamless integration of AI with IoT platforms, digital twins, and edge/cloud, using interoperable data models/ontologies and policy-aware event buses (Abdel-Aty et al., 2022; Gao et al., 2024).
- Federated and privacy-preserving learning for cross-site/cross-enterprise scheduling, with model provenance and usage controls (NIST, 2023).
- Design for robustness: training against disturbances, runtime monitors, and rollback strategies to keep service levels under shocks (Moos et al., 2022).
- Human-in-the-loop frameworks that combine optimization/learning with operator intent, safety culture, and multi-objective business constraints (Mourtzis, 2022).
Author Contributions
Funding
Abbreviations
| AAS | Asset Administration Shell |
| AGV | Automated Guided Vehicle |
| AI | Artificial Intelligence |
| ARO | Adjustable Robust Optimization |
| CNN | Convolutional Neural Network |
| CP | Constraint Programming |
| CPS | Cyber-Physical System |
| DES | Discrete-Event Simulation |
| DRL | Deep Reinforcement Learning |
| DT | Digital Twin |
| ERP | Enterprise Resource Planning |
| GA | Genetic Algorithm |
| GCG | Generic Column Generation |
| GNN | Graph Neural Network |
| HPC | High-Performance Computing |
| ICS | Industrial Control Systems |
| IEC | International Electrotechnical Commission |
| IIoT | Industrial Internet of Things |
| IoT | Internet of Things |
| KPI | Key Performance Indicator |
| KG | Knowledge Graph |
| LBBD | Logic-Based Benders Decomposition |
| LLM | Large Language Model(s) |
| LNS | Large-Neighborhood Search |
| LSTM | Long Short-Term Memory |
| MDP | Markov Decision Process |
| MES | Manufacturing Execution System |
| MILP | Mixed-Integer Linear Programming |
| MIP | Mixed-Integer Programming |
| ML | Machine Learning |
| MLOps | Machine-Learning Operations |
| NIST | National Institute of Standards and Technology |
| NP-hard | Nondeterministic Polynomial-time hard |
| OPC UA | Open Platform Communications Unified Architecture |
| OPRO | Optimizers by Prompting |
| OR | Operations Research |
| PdM | Predictive Maintenance |
| RNN | Recurrent Neural Network |
| RL | Reinforcement Learning |
| SA | Simulated Annealing |
| SCM | Supply Chain Management |
| SCIP | Solving Constraint Integer Programs (optimization framework) |
| TS | Tabu Search |
| UG | Unified parallelization framework for branch-and-bound/price/cut |
| XAI | Explainable Artificial Intelligence |
| XRL | Explainable Reinforcement Learning |
References
- Abdel-Aty, T. A. Abdel-Aty, E. Negri, and S. Galparoli. 2022. Asset Administration Shell in Manufacturing: Applications and Relationship with Digital Twin. IFAC-PapersOnLine 55, 10: 2533–2538. [Google Scholar] [CrossRef]
- Aissi, H. Aissi, C. Bazgan, and D. Vanderpooten. 2009. Min–max and min–max regret versions of combinatorial optimization problems: A survey. European Journal of Operational Research 197, 2: 427–438. [Google Scholar] [CrossRef]
- Allahverdi, A. Allahverdi, C. T. Ng, T. C. E. Cheng, and M. Y. Kovalyov. 2008. A survey of scheduling problems with setup times or costs. European Journal of Operational Research 187, 3: 985–1032. [Google Scholar] [CrossRef]
- Bampoula, X. Bampoula, G. Siaterlis, N. Nikolakis, and K. Alexopoulos. 2021. A Deep Learning Model for Predictive Maintenance in Cyber-Physical Production Systems Using LSTM Autoencoders. Sensors 21, 3: 972. [Google Scholar] [CrossRef]
- Bengio, Y. Bengio, A. Lodi, and A. Prouvost. 2021. Machine Learning for Combinatorial Optimization: a Methodological Tour d’Horizon. European Journal of Operational Research 290, 2: 405–421. [Google Scholar] [CrossRef]
- Ben-Tal, A. Ben-Tal, A. Goryashko, E. Guslitzer, and A. Nemirovski. 2004. Adjustable robust solutions of uncertain linear programs. Mathematical Programming 99, 2: 351–376. [Google Scholar] [CrossRef]
- Beregi, R. Beregi, D. Németh, P. Turek, L. Monostori, and J. Váncza. 2021. Manufacturing Execution System Integration through the Standardization of a Common Service Model for Cyber-Physical Production Systems. Applied Sciences 11, 16: 7581. [Google Scholar] [CrossRef]
- Bertsimas, D., and M. Sim. 2004. The price of robustness. Operations Research 52, 1: 35–53. [Google Scholar] [CrossRef]
- Bestuzheva, K. Bestuzheva, M. Besançon, W.-K. Chen, A. Chmiela, T. Donkiewicz, and et al. 2021. The SCIP Optimization Suite 8.0. arXiv. [Google Scholar] [CrossRef]
- Bestuzheva, K. Bestuzheva, S. Vigerske, T. Achterberg, and et al. 2021. The SCIP Optimization Suite 8.0. arXiv. [Google Scholar] [CrossRef]
- Birge, Louveaux, J. R. Birge, and F. Louveaux. 2011. Introduction to Stochastic Programming, 2nd ed. Springer. [Google Scholar] [CrossRef]
- Blazewicz. 2007. Handbook on Scheduling: From Theory to Applications. Edited by J. Blazewicz, K. H. Ecker, E. Pesch, G. Schmidt and J. Weglarz. Springer: https://link.springer.com/book/10.1007/978-3-540-32220-7.
- Blum, Roli, C. Blum, and A. Roli. 2003. Metaheuristics in combinatorial optimization: Overview and conceptual comparison. ACM Computing Surveys 35, 3: 268–308. [Google Scholar] [CrossRef]
- Bruni, M. E. Bruni, L. Di Puglia Pugliese, P. Beraldi, and F. Guerriero. 2017. An adjustable robust optimization model for the resource-constrained project scheduling problem with uncertain activity durations. Omega 71: 66–84. [Google Scholar] [CrossRef]
- Cappart, Q. Cappart, D. Chételat, E. Khalil, A. Lodi, C. Morris, and P. Veličković. 2021. Combinatorial optimization and reasoning with graph neural networks (Survey). IJCAI 2021. [Google Scholar] [CrossRef]
- Cindrić, I. Cindrić, M. Jurčević, and T. Hadjina. 2025. Mapping of Industrial IoT to IEC 62443 Standards. Sensors 25, 3: 728. [Google Scholar] [CrossRef]
- Cohen, I. Cohen, K. Postek, and S. Shtern. 2023. An adaptive robust optimization model for parallel machine scheduling. European Journal of Operational Research 306, 1: 83–104. [Google Scholar] [CrossRef]
- Fang, K. Fang, N. Uhan, F. Zhao, and J. W. Sutherland. 2011. A new approach to scheduling in manufacturing for power consumption and carbon footprint reduction. Journal of Manufacturing Systems 30, 4: 234–240. [Google Scholar] [CrossRef]
- Ferreira, C. Ferreira, G. Figueira, and P. Amorim. 2022. Effective and interpretable dispatching rules for dynamic job shops via guided empirical learning. Omega 111: 102643. [Google Scholar] [CrossRef]
- Forbes, M. A. Forbes, M. G. Harris, H. M. Jansen, F. A. van der Schoot, and T. Taimre. 2024. Combining optimisation and simulation using logic-based Benders decomposition. European Journal of Operational Research 312, 3: 840–854. [Google Scholar] [CrossRef]
- Gahm, C. Gahm, F. Denz, M. Dirr, and A. Tuma. 2016. Energy-efficient scheduling in manufacturing companies: A review and research framework. European Journal of Operational Research 248, 3: 744–757. [Google Scholar] [CrossRef]
- Gao, Q. Gao, F. Gu, L. Li, and J. Guo. 2024. A framework of cloud–edge collaborated digital twin for flexible job shop scheduling with conflict-free routing. Robotics and Computer-Integrated Manufacturing 86: 102672. [Google Scholar] [CrossRef]
- Garey, M. R., and D. S. Johnson. 1979. Computers and Intractability: A Guide to the Theory of NP-Completeness. W. H. Freeman. (open copy). https://perso.limos.fr/~palafour/PAPERS/PDF/Garey-Johnson79.pdf.
- Gasse, M. Gasse, D. Chételat, N. Ferroni, L. Charlin, and A. Lodi. 2019. Exact Combinatorial Optimization with Graph Convolutional Neural Networks. NeurIPS arXiv:1906.01629. [Google Scholar]
- Giret, A. Giret, D. Trentesaux, and V. Prabhu. 2015. Sustainability in manufacturing operations scheduling: A state of the art review. Journal of Manufacturing Systems 37: 126–140. [Google Scholar] [CrossRef]
- Giret, A. Giret, D. Trentesaux, M. A. Salido, E. Garcia, and E. Adam. 2017. A holonic multi-agent methodology to design sustainable intelligent manufacturing control systems. Journal of Cleaner Production 167: 1370–1386. [Google Scholar] [CrossRef]
- Graham, R. L. Graham, E. L. Lawler, J. K. Lenstra, and A. H. G. Rinnooy Kan. 1979. Optimization and Approximation in Deterministic Sequencing and Scheduling: A Survey. Annals of Discrete Mathematics 5: 287–326. [Google Scholar] [CrossRef]
- Gupta, Stafford, J. N. D. Gupta, and E. F. Stafford. 2006. Flowshop scheduling research after five decades. European Journal of Operational Research 169, 3: 699–711. [Google Scholar] [CrossRef]
- Herrmann, J. W., ed. 2006. Handbook of Production Scheduling. Springer. [Google Scholar] [CrossRef]
- Hottung, Tierney, A. Hottung, and K. Tierney. 2022. Neural large neighborhood search for routing problems. Artificial Intelligence 313: 103786. [Google Scholar] [CrossRef]
- Hu, R. Hu, Z. Yan, W. Ding, and L. T. Yang. 2020. A survey on data provenance in IoT. World Wide Web 23, 2: 1441–1463. [Google Scholar] [CrossRef]
- Huang, Z. Huang, K. Wang, F. Liu, H.-L. Zhen, W. Zhang, M. Yuan, J. Hao, Y. Yu, and J. Wang. 2022. Learning to select cuts for efficient mixed-integer programming. Pattern Recognition 124: 108353. [Google Scholar] [CrossRef]
- Ivanov, Dolgui, D. Ivanov, and A. Dolgui. 2020a. A digital supply chain twin for managing the disruption risks and resilience in the era of Industry 4.0. Production Planning & Control 32, 9: 775–788. [Google Scholar] [CrossRef]
- Ivanov, Dolgui, D. Ivanov, and A. Dolgui. 2020b. Viability of intertwined supply networks: Extending the supply chain resilience angles toward survivability. International Journal of Production Research 58, 10: 2904–2915. [Google Scholar] [CrossRef]
- Ivanov, Dolgui, D. Ivanov, and A. Dolgui. 2021. OR-methods for coping with the ripple effect in supply chains during COVID-19: Managerial insights and research implications. International Journal of Production Economics 232: 107921. [Google Scholar] [CrossRef]
- Juvin, C. Juvin, L. Houssin, and P. Lopez. 2023. Logic-based Benders decomposition for the preemptive flexible job-shop scheduling problem. Computers & Operations Research 154: 106156. [Google Scholar] [CrossRef]
- Khalil, E. B. Khalil, P. Le Bodic, L. Song, G. Nemhauser, and B. Dilkina. 2016. Learning to Branch in Mixed Integer Programming. AAAI 2016: https://ojs.aaai.org/index.php/AAAI/article/view/10080.
- Koch, T. Koch, T. Berthold, J. Pedersen, and C. Vanaret. 2022. Progress in mathematical programming solvers from 2001 to 2020. EURO Journal on Computational Optimization 10: 100031. [Google Scholar] [CrossRef]
- Kool, W. Kool, H. van Hoof, and M. Welling. 2018. Attention, Learn to Solve Routing Problems! arXiv:1803.08475. https://arxiv.org/abs/1803.08475.
- Kouvelis, Yu, P. Kouvelis, and G. Yu. 1997. Robust Discrete Optimization and Its Applications. Springer. [Google Scholar] [CrossRef]
- Kovács, B. Kovács, P. Tassel, M. Gebser, and G. Seidel. 2022. A Customizable Reinforcement Learning Environment for Semiconductor Fab Simulation. 2022 Winter Simulation Conference; pp. 2663–2674. [Google Scholar] [CrossRef]
- Kritzinger, W. Kritzinger, M. Karner, G. Traar, J. Henjes, and W. Sihn. 2018. Digital Twin in manufacturing: A categorical literature review and classification. IFAC-PapersOnLine 51, 11: 1016–1022. [Google Scholar] [CrossRef]
- Kusiak, and A. Kusiak. 2017. Smart manufacturing must embrace big data. Nature 544, 7648: 23–25. [Google Scholar] [CrossRef] [PubMed]
- Kusiak, and A. Kusiak. 2019. Fundamentals of smart manufacturing: A multi-thread perspective. Annual Reviews in Control 47: 214–220. [Google Scholar] [CrossRef]
- Lee, Lee, Y. H. Lee, and S. Lee. 2022. Deep reinforcement learning based scheduling within production plan in semiconductor fabrication. Expert Systems with Applications 191: 116222. [Google Scholar] [CrossRef]
- Leitão, P. Leitão, A. W. Colombo, and S. Karnouskos. 2016a. Industrial automation based on cyber-physical systems technologies: Prototype implementations and challenges. Computers in Industry 81: 11–25. [Google Scholar] [CrossRef]
- Leitão, P. Leitão, S. Karnouskos, L. Ribeiro, J. Lee, T. Strasser, and A. W. Colombo. 2016b. Smart agents in industrial cyber-physical systems. Proceedings of the IEEE 104, 5: 1086–1101. [Google Scholar] [CrossRef]
- Li, Y. Li, Z. Tao, L. Wang, B. Du, J. Guo, and S. Pang. 2023. Digital twin-based job shop anomaly detection and dynamic scheduling. Robotics and Computer-Integrated Manufacturing 79: 102443. [Google Scholar] [CrossRef]
- Liñán, Ricardez-Sandoval, D. A. Liñán, and L. A. Ricardez-Sandoval. 2024. Multicut logic-based Benders decomposition for discrete-time scheduling and dynamic optimization of network batch plants. AIChE Journal 70, 9: e18491. [Google Scholar] [CrossRef]
- Liu, R. Liu, R. Piplani, and C. Toro. 2022. Deep reinforcement learning for dynamic scheduling of a flexible job shop. International Journal of Production Research 60, 13: 4049–4069. [Google Scholar] [CrossRef]
- Lu, Y. Lu, C. Liu, K. I.-K. Wang, H. Huang, and X. Xu. 2020. Digital Twin-driven smart manufacturing: Connotation, reference model, applications and research issues. Robotics and Computer-Integrated Manufacturing 61: 101837. [Google Scholar] [CrossRef]
- Ma, J. Ma, H. Zhou, C. Liu, M. E, Z. Jiang, Q. Wang, and et al. 2020. Study on edge-cloud collaborative production scheduling based on enterprises with multi-factory. IEEE Access 8: 30069–30080. [Google Scholar] [CrossRef]
- Malucelli, N. Malucelli, D. Domini, G. Aguzzi, and M. Viroli. 2025. Neighbor-Based Decentralized Training Strategies for Multi-Agent Reinforcement Learning. Proceedings of the 40th ACM/SIGAPP Symposium on Applied Computing (SAC ’25). [Google Scholar] [CrossRef]
- Milani, S. Milani, S. Faraji, J. Wu, T. McCann, M. Ghassemi, and S. Santu. 2024. Explainable Reinforcement Learning: A Survey and Comparative Review. ACM Computing Surveys 56, 7: 168. [Google Scholar] [CrossRef]
- Mönch, L. Mönch, J. W. Fowler, and S. J. Mason. 2013. Production Planning and Control for Semiconductor Wafer Fabrication Facilities: Modeling, Analysis, and Systems. Springer. [Google Scholar] [CrossRef]
- Monostori, and L. Monostori. 2016. Cyber-physical systems in manufacturing. CIRP Annals 65, 2: 621–641. [Google Scholar] [CrossRef]
- Moos, J. Moos, K. Hansel, H. Abdulsamad, S. Stark, D. Clever, and J. Peters. 2022. Robust Reinforcement Learning: A Review of Foundations and Recent Advances. Machine Learning and Knowledge Extraction 4, 1: 276–315. [Google Scholar] [CrossRef]
- Mourtzis, and D. Mourtzis. 2022. Advances in Adaptive Scheduling in Industry 4.0. Frontiers in Manufacturing Technology 2: 937889. [Google Scholar] [CrossRef]
- Mourtzis, Vlachou, D. Mourtzis, and E. Vlachou. 2018. A cloud-based cyber-physical system for adaptive shop-floor scheduling and condition-based maintenance. Journal of Manufacturing Systems 47: 179–198. [Google Scholar] [CrossRef]
- Mourtzis, D. Mourtzis, E. Vlachou, and N. Milas. 2016. Industrial Big Data as a Result of IoT Adoption in Manufacturing. Procedia CIRP 55: 290–295. [Google Scholar] [CrossRef]
- Naderi, Roshanaei, B. Naderi, and V. Roshanaei. 2022. Critical-path-search logic-based Benders decomposition approaches for flexible job shop scheduling. INFORMS Journal on Optimization 4, 1: 1–28. [Google Scholar] [CrossRef]
- Nair, V. Nair, S. Bartunov, F. Gimeno, and et al. 2020. Solving mixed integer programs using neural networks. arXiv. [Google Scholar] [CrossRef]
- National Institute of Standards and Technology (NIST). 2023. Guide to Operational Technology (OT) Security. [CrossRef]
- Neubauer, M. Neubauer, L. Steinle, C. Reiff, S. Ajdinović, L. Klingel, A. Lechler, and A. Verl. 2023. Architecture for Manufacturing-X: Bringing Asset Administration Shell, Eclipse Dataspace Connector and OPC UA together. Manufacturing Letters 37: 1–6. [Google Scholar] [CrossRef]
- Nitsche, B. Nitsche, J. Brands, H. Treiblmaier, and J. Gebhardt. 2023. The impact of multiagent systems on autonomous production and supply chain networks: use cases, barriers and contributions to logistics network resilience. Supply Chain Management: An International Journal 28, 5: 894–908. [Google Scholar] [CrossRef]
- Ouelhadj, Petrovic, D. Ouelhadj, and S. Petrovic. 2009. A survey of dynamic scheduling in manufacturing systems. Journal of Scheduling 12, 4: 417–431. [Google Scholar] [CrossRef]
- Panzer, Bender, M. Panzer, and B. Bender. 2022. Deep Reinforcement Learning in Production Systems: A Systematic Literature Review. International Journal of Production Research 60, 13: 4316–4341. [Google Scholar] [CrossRef]
- Parente, M. Parente, G. Figueira, P. Amorim, and A. Marques. 2020. Production scheduling in the context of Industry 4.0: Review and trends. International Journal of Production Research 58, 17: 5401–5431. [Google Scholar] [CrossRef]
- Park, Park, I.-B. Park, and J. Park. 2023. Scalable Scheduling of Semiconductor Packaging Facilities Using Deep Reinforcement Learning. IEEE Transactions on Cybernetics 53, 6: 3518–3531. [Google Scholar] [CrossRef] [PubMed]
- Peng, Y. Peng, B. Choi, and J. Xu. 2021. Graph Learning for Combinatorial Optimization: A Survey of State-of-the-Art. Data Science and Engineering 6: 119–141. [Google Scholar] [CrossRef]
- Pinedo, and M. L. Pinedo. 2016. Scheduling: Theory, Algorithms, and Systems, 5th ed. Springer. [Google Scholar] [CrossRef]
- Puterman, and M. L. Puterman. 2005. Markov Decision Processes: Discrete Stochastic Dynamic Programming, 2nd ed. Wiley. [Google Scholar] [CrossRef]
- Qin, Lu, Z. Qin, and Y. Lu. 2024. Knowledge graph-enhanced multi-agent reinforcement learning for adaptive scheduling in smart manufacturing. Journal of Intelligent Manufacturing.
- Qin, Z. Qin, D. Johnson, and Y. Lu. 2023. Dynamic production scheduling towards self-organizing mass personalization: A multi-agent dueling deep reinforcement learning approach. Journal of Manufacturing Systems 68: 242–257. [Google Scholar] [CrossRef]
- Rauch, E. Rauch, C. Linder, and P. Dallasega. 2020. Anthropocentric perspective of production before and within Industry 4.0. Computers & Industrial Engineering 139: 105644. [Google Scholar] [CrossRef]
- Romero, D. Romero, J. Stahre, M. Taisch, and et al. 2020. The Operator 4.0: Towards socially sustainable factories of the future. Computers & Industrial Engineering 139: 106128. [Google Scholar] [CrossRef]
- Seitz, M. Seitz, F. Gehlhof, L. A. Cruz Salazar, A. Fay, and B. Vogel-Heuser. 2021. Automation platform independent multi-agent system for robust networks of production resources in Industry 4.0. Journal of Intelligent Manufacturing 32, 7: 2023–2041. [Google Scholar] [CrossRef]
- Serrano-Ruiz, J. C. Serrano-Ruiz, J. Mula, and R. Poler. 2021. Smart manufacturing scheduling: A literature review. Journal of Manufacturing Systems 61: 265–287. [Google Scholar] [CrossRef]
- Siatras, V. Siatras, E. Bakopoulos, P. Mavrothalassitis, N. Nikolakis, and K. Alexopoulos. 2024. Production Scheduling Based on a Multi-Agent System and Digital Twin: A Bicycle Industry Case. Information 15, 6: 337. [Google Scholar] [CrossRef]
- Smit, I. G. Smit, J. Zhou, R. Reijnen, Y. Wu, J. Chen, C. Zhang, Z. Bukhsh, Y. Zhang, and W. Nuijten. 2024/2025. Graph neural networks for job shop scheduling problems: A survey. Computers & Operations Research 176: 106914. [Google Scholar] [CrossRef]
- Song, L. Song, Y. Li, and J. Xu. 2023. Dynamic Job-Shop Scheduling Based on Transformer and Deep Reinforcement Learning. Processes 11, 12: 3434. [Google Scholar] [CrossRef]
- Tang, Y. Tang, S. Agrawal, and Y. Faenza. 2020. Reinforcement Learning for Integer Programming: Learning to Cut. Proceedings of the 37th International Conference on Machine Learning (ICML 2020), PMLR 119: 9367–9376. [Google Scholar] [CrossRef]
- Tao, Zhang, F. Tao, and M. Zhang. 2017. Digital Twin Shop-Floor: A New Shop-Floor Paradigm Towards Smart Manufacturing. IEEE Access 5: 20418–20427. [Google Scholar] [CrossRef]
- Trunzer, E. Trunzer, B. Vogel-Heuser, J.-K. Chen, and M. Kohnle. 2021. Model-Driven Approach for Realization of Data Collection Architectures for Cyber-Physical Systems of Systems to Lower Manual Implementation Efforts. Sensors 21, 3: 745. [Google Scholar] [CrossRef]
- Uhlemann, T. H.-J. Uhlemann, C. Schock, C. Lehmann, S. Freiberger, and R. Steinhilper. 2017. The Digital Twin: Demonstrating the potential of real-time data acquisition in production systems. Procedia Manufacturing 9: 113–120. [Google Scholar] [CrossRef]
- Umer, M. A. Umer, M. Umer, M. Pandey, and S. Abdulla. 2024. Leveraging Artificial Intelligence and Provenance Blockchain Framework to Mitigate Risks in Cloud Manufacturing in Industry 4.0. Electronics 13, 3: 660. [Google Scholar] [CrossRef]
- Vieira, G. E. Vieira, J. W. Herrmann, and E. Lin. 2003. Rescheduling manufacturing systems: A framework of strategies, policies, and methods. Journal of Scheduling 6, 1: 39–62. [Google Scholar] [CrossRef]
- Vinyals, O. Vinyals, M. Fortunato, and N. Jaitly. 2015. Pointer Networks. NeurIPS. https://arxiv.org/abs/1506.03134.
- Wan, Y. Wan, Y. Liu, Z. Chen, C. Chen, X. Li, F. Hu, and et al. 2024. Making knowledge graphs work for smart manufacturing: Research topics, applications and prospects. Journal of Manufacturing Systems 76: 1–22. [Google Scholar] [CrossRef]
- Wang, L. Wang, Z. Pan, and J. Wang. 2021. A review of reinforcement learning-based intelligent optimization for manufacturing scheduling. Complex System Modeling and Simulation 1, 4: 257–270. [Google Scholar] [CrossRef]
- Wang, R. Wang, G. Wang, J. Sun, F. Deng, and J. Chen. 2024. Flexible Job Shop Scheduling via Dual Attention Network-Based Reinforcement Learning. IEEE Transactions on Neural Networks and Learning Systems 35, 3: 3091–3102. [Google Scholar] [CrossRef]
- Wang, S. Wang, J. Wan, D. Li, and C. Zhang. 2016. Implementing smart factory of Industrie 4.0: An outlook. International Journal of Distributed Sensor Networks 12, 1: 3159805. [Google Scholar] [CrossRef]
- Wang, W. Wang, Y. Zhang, Y. Wang, G. Pan, and Y. Feng. 2025. Hierarchical multi-agent deep reinforcement learning for dynamic flexible job-shop scheduling with transportation. International Journal of Production Research (advance online publication), 1–28. [Google Scholar] [CrossRef]
- Wang, Z. Wang, X. Li, J. Wang, Y. Kuang, M. Yuan, J. Zeng, Y. Zhang, and F. Wu. 2023. Learning cut selection for mixed-integer linear programming via hierarchical sequence model. arXiv. [Google Scholar] [CrossRef]
- Weng, W. Weng, J. Chen, M. Zheng, and S. Fujimura. 2022. Realtime scheduling heuristics for just-in-time production in large-scale flexible job shops. Journal of Manufacturing Systems 63: 64–77. [Google Scholar] [CrossRef]
- Xia, K. Xia, C. Sacco, M. Kirkpatrick, C. Saidy, L. Nguyen, A. Kircaliali, and R. Harik. 2021. A digital twin to train deep reinforcement learning agent for smart manufacturing plants: Environment, interfaces and intelligence. Journal of Manufacturing Systems 58: 210–230. [Google Scholar] [CrossRef]
- Xiong, H. Xiong, S. Shi, D. Ren, and J. Hu. 2022. A survey of job shop scheduling problem: The types and models. Computers & Operations Research 142: 105731. [Google Scholar] [CrossRef]
- Xu, H. Xu, W. Yu, D. Griffith, and N. Golmie. 2018. A Survey on Industrial Internet of Things: A Cyber-Physical Systems Perspective. IEEE Access 6: 78238–78259. https://pmc.ncbi.nlm.nih.gov/articles/PMC9074819/. [CrossRef]
- Xu, L. D. Xu, E. L. Xu, and L. Li. 2018. Industry 4.0: State of the art and future trends. International Journal of Production Research 56, 8: 2941–2962. [Google Scholar] [CrossRef]
- Xu, Y. Xu, Z. Wang, and X. Li. 2025. Multi-agent reinforcement learning for flexible job shop scheduling: A review. Frontiers in Industrial Engineering 2: 1611512. [Google Scholar] [CrossRef]
- Yang, C. Yang, X. Wang, Y. Lu, H. Liu, Q. V. Le, D. Zhou, and X. Chen. 2024. Large Language Models as Optimizers (OPRO). ICLR 2024. https://openreview.net/forum?id=Bb4VGOWELI.
- Zhang, C. Zhang, M. Juraschek, and C. Herrmann. 2024. Deep reinforcement learning-based dynamic scheduling for resilient and sustainable manufacturing: A systematic review. Journal of Manufacturing Systems 77: 962–989. [Google Scholar] [CrossRef]
- Zhang, C. Zhang, W. Song, Z. Cao, J. Zhang, P. S. Tan, and C. Xu. 2020. Learning to dispatch for job-shop scheduling via deep reinforcement learning. arXiv. [Google Scholar] [CrossRef]
- Zhang, C. Zhang, X. Wang, and J. Li. 2023. DeepMAG: Multi-agent graph reinforcement learning for dynamic job shop scheduling. Knowledge-Based Systems 259: 110083. [Google Scholar] [CrossRef]
- Zhang, F. Zhang, J. Bai, D. Yang, and Q. Wang. 2022. Digital twin data-driven proactive job-shop scheduling strategy towards asymmetric manufacturing execution decision. Scientific Reports 12: 1546. [Google Scholar] [CrossRef]
- Zhang, L. Zhang, Y. Yan, Y. Hu, and W. Ren. 2022. Reinforcement learning and digital twin-based real-time scheduling method in intelligent manufacturing systems. IFAC-PapersOnLine 55, 10: 359–364. [Google Scholar] [CrossRef]
- Zhang, M. Zhang, F. Tao, and A. Y. C. Nee. 2020. Digital twin-enhanced dynamic job-shop scheduling. Journal of Manufacturing Systems 58: 146–156. [Google Scholar] [CrossRef]
- Zhang, T. Zhang, S. Xie, and O. Rose. 2017. Real-time job shop scheduling based on simulation and Markov decision processes. Proceedings of the 2017 Winter Simulation Conference (WSC); pp. 3357–3368. [Google Scholar] [CrossRef]
- Zhang, Y. Zhang, H. Zhu, D. Tang, T. Zhou, and Y. Gui. 2022. Dynamic job shop scheduling based on deep reinforcement learning for multi-agent manufacturing systems. Robotics and Computer-Integrated Manufacturing 78: 102412. [Google Scholar] [CrossRef]
- Zheng, J. Zheng, Y. Zhao, Y. Li, J. Li, L. Wang, and D. Yuan. 2025. Dynamic flexible flow shop scheduling via cross-attention networks and multi-agent reinforcement learning. Journal of Manufacturing Systems 80: 395–411. [Google Scholar] [CrossRef]
- Zhong, R. Y. Zhong, X. Xu, E. Klotz, and S. T. Newman. 2017. Intelligent manufacturing in the context of Industry 4.0: A review. Engineering 3, 5: 616–630. [Google Scholar] [CrossRef]
| Approach | Core strengths | Limitations | Typical application areas | Representative references |
|---|---|---|---|---|
| Genetic algorithms & memetic hybrids | Flexible; multi-objective ready; easy to hybridize with local search/repair; robust on heterogeneous constraints | Parameter tuning; stochastic variance; may plateau without strong neighborhoods | Parallel/flow/flexible job shops; sequence-dependent setups; large unrelated-machine problems | (Blum & Roli, 2003; Ferreira et al., 2022) |
| Simulated annealing / Tabu search | Simple and effective baselines; good intensification/diversification; easy to embed constraints | Cooling/tenure sensitivity; may require problem-specific neighborhoods | Job/flow shops; batching; setup-heavy sequencing | (Blum & Roli, 2003) |
| Large-neighborhood search (LNS) / Neural-LNS | Powerful destroy–repair exploration; learned destroy/repair improves speed & quality; anytime behavior | Designing repairs that preserve feasibility; training data/compute for neural variants | High-mix shops; near-real-time improvement; rolling re-optimization | (Hottung & Tierney, 2022) |
| Hyper-heuristics (selection / generation) | Generalizes across instance types; automates rule choice; compatible with DRL | Performance ceiling if candidate pool is weak; requires meta-level data | Mixed-model production; variable routing/loads | (Panzer & Bender, 2022; Smit & Van Vliet, 2024) |
| Logic-Based Benders Decomposition (LBBD) | Strong logic cuts; separates assignment/sequence from timing; integrates CP/MIP/heuristics | Modeling effort; cut engineering; potential many iterations | Flexible/distributed job shops; process/chemical scheduling | (Naderi et al., 2022; Juvin et al., 2023; Liñán & Méndez, 2024; Forbes & Kelly, 2024) |
| Hierarchical / rolling-horizon schemes | Scales long horizons; aligns with planning→scheduling tiers; supports simulation-in-the-loop | Coordination overhead; myopic decisions if horizons too short | Plant-level planning with shop-floor dispatch; digital-twin what-if analysis | (Liñán & Méndez, 2024; Forbes & Kelly, 2024) |
| Column generation / branch-and-price frameworks | Decompose by columns/routes; strong bounds; mix with heuristics | Pricing complexity; stabilization needed; parallelization non-trivial | Large machine/route generation models; transportation–production links | (Bestuzheva et al., 2021; Koch et al., 2022) |
| Parallel solver ecosystems | Multicore/cluster speedups; parallel B&B/price/cut (UG); mature tooling | Needs HPC resources; solver engineering expertise | Large MIP/CP scheduling; scenario-decomposed planning | (Bestuzheva et al., 2021; Koch et al., 2022) |
| DRL dispatching policies (GNN/attention) | Learns size-agnostic rules; reacts online; strong anytime performance | Sample efficiency; stability/robustness; policy explainability | Dynamic job/flexible job shops; real-time dispatch | (Zhang et al., 2020; Panzer & Bender, 2022; Zhang et al., 2024; Smit & Van Vliet, 2024) |
| Learning-augmented optimization (ML for OR) | Learned branching/cuts/node selection; warm-starts; improves primal-dual gaps | Generalization across distributions; integration into certified workflows | Large MIP/CP scheduling; hybrid MH+MIP stacks | (Nair et al., 2020; Tang et al., 2020; Tian et al., 2021; Wang et al., 2022; Bengio et al., 2021) |
| Surrogate-/supervised rule learning | Fast evaluations; interpretable policies; good for high-volume data | Surrogate bias; retraining under drift; limited exploration | Repetitive/flow environments; KPI-specific rule mining | (Ferreira et al., 2022; Gil-Gala et al., 2023) |
| Digital twin–in-the-loop RL | Safe policy training; proactive, state-aware rescheduling; sim-to-real transfer | Twin fidelity/sync cost; integration complexity | Smart factories; semiconductor/assembly lines | (Zhang et al., 2022) |
| Foundation-model–guided heuristics (OPRO) | Rapid heuristic design/tuning; few-shot adaptability; complements DRL/OR | Very early stage; needs feasibility guards and evaluation harness | Rapid ramp-up for new product mixes/lines | (Yang et al., 2024; Bengio et al., 2021) |
| Approach | Core strengths | Limitations | Typical application areas | Representative references |
| Min–max & Min–max Regret Robust Optimization | Strong guarantees; interpretable; protects against penalties | Conservative; scalability issues with large scenario sets | Semiconductor fabs, aerospace, contract manufacturing | (Aissi et al., 2009; Bertsimas & Sim, 2004) |
| Adjustable Robust Optimization (ARO) | Balances robustness and flexibility; realistic for dynamic shops | More complex; heavier computation | Job shops with uncertain processing times | (Ben-Tal et al., 2004; Cohen et al., 2023) |
| Interval/Set-Based Models | Tractable; practical for bounded uncertainties | Can yield conservative schedules | Project-driven and regulated industries | (Bruni et al., 2017) |
| Learning-in-the-loop Robust Models | Adaptive; efficient evaluation; improves robustness | Requires quality data; explainability issues | Flexible manufacturing, online scheduling | (Zhang et al., 2022) |
| Chance-Constrained Scheduling | Balances service levels vs efficiency; intuitive | Relies on accurate distribution estimation | Service industries, logistics, large projects | (Birge & Louveaux, 2011) |
| Markov Decision Processes (MDP) | Principled sequential control; foundation for DRL | Curse of dimensionality for large systems | Stochastic job shops, batch processes | (Zhang et al., 2017; Puterman, 2005) |
| Simulation-Based Evaluation (DES/Monte Carlo) | Flexible; captures complex interactions; supports stress-testing | Computationally expensive | Semiconductor, project-based, high-uncertainty industries | (Vieira et al., 2003; Mönch et al., 2013) |
| Rescheduling & Repair Algorithms | Stable shop floor behavior; minimal disruption | Myopic if frequent disruptions occur | MES/MRP systems, dynamic job shops | (Vieira et al., 2003; Ouelhadj & Petrovic, 2009) |
| Rolling-Horizon/Event-Driven Updates | Continuous adaptation; ERP/MES integration | Risk of nervousness with frequent updates | High-mix, volatile production | (Vieira et al., 2003; Weng et al., 2022) |
| Predictive Analytics & ML | Data-driven; real-time adaptability; generalizable policies | Data hungry; legacy integration challenges | Smart factories, flexible electronics | (Serrano-Ruiz et al., 2021; Zhang et al., 2024) |
| Digital-Twin-in-the-Loop Scheduling | Safe training/testing; improves sample efficiency | Twin fidelity/synchronization cost | Intelligent manufacturing, reconfigurable factories | (Zhang et al., 2020; Zhang et al., 2022) |
| Multi-Agent & Self-Organizing Systems | Resilient; scalable; fault-tolerant | Coordination and global optimality issues | Cyber-physical production, distributed factories | (Leitão et al., 2016a; Seitz et al., 2021) |
| End-to-End AI Stacks at Scale | Hybrid performance; scalable and adaptive under real-time constraints | Engineering complexity; integration & MLOps challenges | Large-scale Industry 4.0, smart factories | (Lee & Lee, 2022; Zhang et al., 2024) |
| Approach | Core Strengths | Limitations | Typical Application Areas | Representative References |
| Sensor-Enabled, Closed-Loop Scheduling | Real-time responsiveness; immediate adaptation to shop-floor events; integration of IIoT/CPS data streams | Data quality and latency challenges; integration with legacy systems; requires robust edge analytics | High-variability shop floors; condition-based rescheduling; flow-shop monitoring | Wang et al. (2016); Rauch et al. (2020) |
| Digital Twin-Based Scheduling | Virtual experimentation; safe training/testbed for RL agents; proactive rescheduling and predictive maintenance | High development and synchronization costs; computationally intensive | Job-shop/flexible shop scheduling; disruption management; predictive control | Uhlemann et al. (2017); Kritzinger et al. (2018); Zhang et al. (2021); Leng et al. (2020) |
| Cloud and Edge Computing for Distributed Scheduling | Scalable optimization (cloud); low-latency local response (edge); hybrid setups balance global and local | Security and data-transfer overhead; partitioning optimization tasks is complex | Multi-plant coordination; distributed supply chains; real-time edge rescheduling | Mourtzis & Vlachou (2018); Lu et al. (2020) |
| Agent-Based and Multi-Agent Scheduling Systems | Decentralization, modularity, and negotiation capabilities; well-suited to flexible manufacturing | Coordination overhead; global optimality hard to guarantee | Flexible job-shop systems; distributed resource allocation | Leitão et al. (2016a); Giret et al. (2017); Siatras et al. (2024); Xu et al. (2025) |
| Self-Optimizing and Adaptive Control Algorithms | Continuous adaptation to data and disturbances; reinforcement learning and heuristic evolution enable resilience | Sample inefficiency in RL; difficulty in explainability; requires large/high-quality datasets | Dynamic job-shop scheduling; mass personalization; adaptive planning | Kusiak (2017); Qin et al. (2023); Zhang et al. (2023); Wang et al. (2025) |
| Emerging Architectures (KG-MARL, attention-based, decentralized training) | Enhanced context-awareness; improved coordination; scalable decentralized learning | Complexity of design; limited industrial deployments; integration with legacy IT/OT | Smart manufacturing scheduling; dynamic flow/assembly shops | Qin & Lu (2024); Zheng et al. (2025); Malucelli et al. (2025) |
| Interoperable Architectures (OPC UA, AAS, open APIs) | Seamless integration across ERP/MES/SCM; supports plug-and-operate scheduling services | Requires ecosystem-wide standard adoption; potential vendor lock-in | Multi-system integration; cross-site scheduling; Manufacturing-X initiatives | Beregi et al. (2021); Neubauer et al. (2023) |
| Semantically Enriched, AI-Ready Data Layers | Standard vocabulary for heterogeneous data; improves explainability and feature quality for DL/RL | Knowledge graph development overhead; ontology alignment challenges | Digital twins; predictive scheduling; cross-enterprise scheduling | Wan et al. (2024); Trunzer et al. (2021) |
| Security and Data Provenance | Ensures integrity, confidentiality, and traceability of scheduling data; supports compliance (IEC 62443, NIST) | Added overhead in performance; blockchain solutions not yet fully scalable | Regulated supply chains; critical infrastructures; cloud manufacturing | Cindrić et al. (2025); NIST (2023); Hu et al. (2020); Umer et al. (2024) |
| Data Sovereignty & Federated Collaboration | Policy-enforced data sharing across organizations; supports privacy-preserving optimization | Governance complexity; interoperability still evolving | Inter-company scheduling; collaborative supply chains; subcontracting | Neubauer et al. (2023) |
| Operational Hardening for AI-Driven Scheduling | Secure and reproducible ML pipelines; signed model artifacts; trustworthy rescheduling | Requires ML lifecycle governance; raises infrastructure complexity | AI-driven job-shop scheduling; cloud-edge rescheduling services | Beregi et al. (2021); NIST (2023) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).