2. Related Works
Publications [
1,
2,
3,
4] are review articles. In [
1], a review on route planning approaches in multi-agent systems is presented, including the classification of
A*-based methods, heuristic approaches, biologically inspired techniques, and artificial intelligence (AI)-based algorithms. It is shown that many existing approaches do not account for battery constraints or time-varying moving obstacles. Article [
2] focuses on the route planning and control of mobile robots, including AGVs and Autonomous Mobile Robots (AMRs) in logistics environments (warehouses and intralogistics) based on a control model combined with layered planning (task scheduling and motion planning) while considering constraints in complex intralogistics settings. In [
3], a review of modern methods of path planning in dynamic environments for various robotic systems is conducted. Sensor modalities, obstacle types, planning strategies (avoid, plan, replan), and key challenges (adaptation time, inter-agent communication) are analyzed. A classification of planning approaches is provided, and weaknesses are identified, particularly in multi-step planning for dynamic agents. In [
4], the authors present a comprehensive survey on path planning methods for dynamic environments. They categorize existing approaches into constraint-based, dynamic, and multi-agent and highlight open research gaps. Particular attention is drawn to the necessity of integrating battery limitations, dynamic obstacles, and inter-agent coordination.
In [
5], the authors enhanced the
A* algorithm in combination with the Dynamic Window Approach (DWA) for AGVs operating in factory environments with both dynamic and static obstacles. The method dynamically adjusts the heuristic weight in
A* based on obstacle density and integrates it with DWA for local adaptation. As a result, the planning time is reduced, and the generated route is smoother than that produced by classical
A*. Research [
6] proposes a path planning system for multi-agent manufacturing environments based on an improved
A* algorithm. The results show that the approach enhances agent coordination and reduces conflicts in shared spaces with cooperating AGVs. In [
7], a dynamic path planning method is proposed for an AGV that follows a moving target using integrated sensory and visual information. The results demonstrate that the AGV successfully tracks a moving target in changing environments using the proposed method.
In [
8],
D* Lite was used in a conflict-based search for multi-task routing. This enabled the integration of dynamic replanning with the detection of potential conflicts among agents. A comparison with the standard approach shows a reduction in the number of disputes and an improvement in planning success under conditions of obstacles. An algorithm for planning a route for an AGV operating in a dynamic environment, based on
A* and DWA, is proposed in [
9]. A feature of this study is the inclusion of route adaptation to dynamic obstacles, accounting for speed and safety constraints. Experimental results show that the proposed hybrid approach exhibits better adaptability and shorter paths than the standalone
A* and DWA algorithms. The authors of [
10] developed a hybrid algorithm that utilizes kinematic-constraint
A* with DWA for an AGV in a dynamic environment. The proposed algorithm reduces the path length by 3.6% and the local planning time by 50% compared with traditional methods. In [
11],
A* under AGV kinematic constraints was combined with local DWA planning in dynamic environments. As a result, the path generated by the global plan became smoother in terms of the number of turns and time compared to that of classical
A*. In [
12], a multi-criteria path planning algorithm was proposed, combining global Ant Colony Optimization (ACO) search to reduce the path cost and local optimization through a modified DWA. In simulations, the algorithm exhibited significant reductions in path length and the number of turns compared with the baseline. In [
13], the authors improved
A* with adaptive heuristics and a combination with DWA analysis in environments with different types of obstacles. That is, adaptive heuristics were applied with a search approach to increase efficiency and reduce redundant nodes. The results demonstrate an increase in planning efficiency, achieved by reducing the number of nodes and search time compared to basic
A*.
In [
14], an improved version of
D* Lite for handling moving objects in dynamic settings is applied to an underwater vehicle for multi-goal planning and collision avoidance in an unknown dynamic environment. The results show that the modified
D* Lite plans efficient routes under non-static conditions and, compared to the basic versions, achieves a higher success rate. Work [
15] is dedicated to a hybrid framework that combines
D* Lite as a global planner and Multi-Agent Reinforcement Learning (MARL) for local decisions in an environment with dynamic obstacles and changing states using a "switching" mechanism between the global planner and the local learning agent while maintaining operability when the environment changes. As a result, multi-agent planning success improves, the number of conflicts decreases, and path efficiency increases relative to a purely centralized approach. In [
16], the authors propose a path planning scheme for AGV trajectory control in terminals with both static and dynamic obstacles. It is shown that the proposed scheme effectively bypasses obstacles and increases trajectory stability. Paper [
17] presents a multi-agent system with centralized decision-making that uses
D* Lite to account for autonomous behavior and coordination in the presence of conflicts and interactions between agents. It is shown that the scheme enables the coordination of agents’ routes and replanning in a changing environment. In [
18], the authors presented an improved potential field method for multiple AGVs in a dynamic port environment that accounts for the minimum safe distance between AGVs and the dynamic repulsive potential. The algorithm adapts to environmental changes in real time, reducing the probability of collisions.
Methods using AI are presented in publications [
19,
20,
21,
22,
23,
24,
25]. In [
19], an optimized shortest-path model for AGVs based on a support vector machine (SVM) was developed, resulting in increased planning efficiency under complex production conditions. In [
20], a review of path-planning approaches for multiple mobile robots was conducted, covering classical, heuristic, and AI-based methods. The authors emphasized the trend toward decentralized planning in multi-robot systems and identified key challenges, including real-time operation, resource constraints, and inter-robot coordination. In [
21], a dynamic multi-criteria path-planning model based on multi-agent deep reinforcement learning (MADRL) was proposed. It has been shown that agents can reach their targets while simultaneously optimizing multiple criteria, such as time, distance, and the number of turns, in complex environments. In [
22], a hybrid AGV path-optimization algorithm was introduced, integrating the evolutionary Grey Wolf Optimizer (GWO) algorithm with local partially matched crossover (PMX) mutation to improve route smoothness and length. Experiments confirmed that the evolved routes are smoother and shorter than those of classical heuristic approaches. The novelty of the approach in [
23] lies in combining a learning-based method (a K-L network enhanced with a genetic algorithm) with global planning to achieve smoother navigation paths. The results demonstrate effective avoidance of dynamic obstacles and improved coordination among multiple AGVs. In [
24], deep reinforcement learning methods—specifically Deep Deterministic Policy Gradient (DDPG) and Multi-Agent Deep Deterministic Policy Gradient (MADDPG)—were applied for cooperative AGV navigation in dynamic environments with obstacle avoidance. The novelty of the method lies in centralized training with decentralized execution and optimized experience buffers. The study shows that agents successfully coordinate and avoid collisions, thereby reducing travel time and conflicts. Paper [
25] proposes an intelligent motion-control system for a wheeled mobile robotic platform that integrates navigation, fuzzy logic, and advanced sensor-data processing to ensure adaptive behavior in dynamic environments. The novelty lies in the development of a rule-based fuzzy controller and a remote software implementation, with simulation results demonstrating effective platform motion under varying sensor conditions.
Publications [
26,
27,
28,
29] consider the remaining SoC. In [
26], a priority-based scheduling approach for AGVs using the ACO algorithm for collision avoidance is proposed, with the AGVs’ battery level determining the collision avoidance priority. Experimental results confirm that battery-based prioritization reduces conflicts and improves AGV traffic efficiency in a factory. The authors of [
27] consider the problem of real-time dispatching of an AGV fleet under battery constraints. The proposed solution based on deep reinforcement learning demonstrated improvements in both battery-utilization efficiency and overall system performance. The authors of [
28] developed a task-scheduling method for AGVs under battery constraints that combines task redistribution, charging management via a Gurobi/LP module, and optimization using a battery-threshold policy. The results demonstrate industry-applicable solutions that minimize delays and traffic-related costs. In [
29], optimization of AGV routing and loading schedules in a container terminal, accounting for load-flow dynamics, is considered. The novelty of the approach lies in a dual-threshold charging strategy adapted to load-flow conditions, integrated with routing and scheduling. The results are verified through simulations, and it is validated that the flexible charging strategy increases AGV availability and reduces downtime.
Based on the conducted literature review, most approaches for AGV route planning can be grouped into the following categories:
- 1.
Classical heuristic and graph-search methods, including A*, Dijkstra, D*, D* Lite, Theta*, Lifelong Planning A*, etc. They guarantee the existence of an optimal or approximate route. However, these methods suffer from high computational complexity when scaled to large or dynamic environments and generally do not account for the remaining AGV battery charge.
- 2.
Local optimization methods, including the DWA, Artificial Potential Field (APF), Velocity Obstacle (VO), and their improved variants. They may implicitly account for energy consumption through constraints on maximum speed or acceleration.
- 3.
Evolutionary and swarm-intelligence methods, such as Genetic Algorithms (GA), ACO, Particle Swarm Optimization (PSO), GWO, etc. These approaches perform well for multi-criteria tasks but often require many iterations and are unstable in dynamic environments. Therefore, they are effective for real-time tasks but are prone to local minima and do not guarantee a global optimum. They can incorporate energy as an optimization criterion, but in a simplified manner, without dynamic discharge modeling or constraints on returning to the charging station.
- 4.
Machine Learning and Reinforcement Learning methods, such as Deep Q-Learning, DDPG, Proximal Policy Optimization (PPO), MADDPG, and Graph Neural Networks (GNNs), for multi-agent planning. These methods enable AGVs to adapt their behavior to environmental changes but require large amounts of training data and substantial computational resources. They can incorporate SoC into the state space. However, literature analysis shows that these methods typically model only a minimum SoC constraint after which the AGV must return to charging and do not address battery degradation or inter-cell charge variation.
- 5.
Hybrid methods integrate global planners (e.g., A*, D*) with local obstacle-avoidance techniques (e.g., DWA, APF) or combine heuristic methods with reinforcement learning or evolutionary algorithms. They strive to balance speed, adaptability, and accuracy, yet most studies do not jointly consider energy constraints, time efficiency, and the presence of live dynamic obstacles. In such methods, battery considerations are typically limited to pre-planned charging-point routing or are completely ignored in agent-coordination scenarios.
Thus, key research gaps include the inadequate integration of energy management, adaptive dynamic obstacle avoidance, and stable route execution in multi-agent environments.