1. Introduction
Equipment maintenance knowledge graphs have emerged as fundamental infrastructures for modern intelligent manufacturing and maintenance systems, providing structured representations of complex relationships among faults, components, maintenance activities, diagnostic procedures, and specialized tools [
1,
2]. By organizing maintenance domain knowledge into graph-structured formats where entities represent real-world objects and edges capture semantic relationships, these knowledge graphs enable intelligent systems to perform sophisticated reasoning for diagnostic support, predictive maintenance, and maintenance solution recommendation [
3,
4]. However, despite their increasing adoption in industrial applications, knowledge graphs inherently suffer from incompleteness—a significant portion of valid relationships remain unrecorded due to the limitations of manual curation, the dynamic nature of equipment systems, and the sparsity of maintenance records for rare fault scenarios [
5]. This incompleteness severely constrains the effectiveness of downstream applications such as augmented reality (AR)-based maintenance guidance systems [
6], intelligent fault diagnosis [
7], and proactive maintenance scheduling [
8,
9].
Knowledge graph completion through link prediction—the task of automatically inferring missing relationships between entities—has thus become a critical challenge in equipment maintenance informatics. The objective is to develop computational models that can predict whether a relationship exists between a given entity pair, thereby enriching the knowledge graph with previously unknown but semantically valid connections. For instance, given a specific fault symptom and a maintenance procedure, a link prediction model should determine whether the procedure is applicable to that fault, even if such an association has not been explicitly documented in historical maintenance records. Successful link prediction not only enhances knowledge graph completeness but also enables more accurate knowledge recommendation, more comprehensive fault correlation analysis, and more reliable maintenance decision support.
However, knowledge associations in the equipment maintenance domain exhibit unique characteristics that pose significant challenges to existing link prediction approaches. First, maintenance knowledge encompasses both tight local neighborhood relationships and complex multi-hop path dependencies. Local relationships, such as the direct causation between a specific component failure and its observable symptoms, require models to effectively capture fine-grained structural patterns within immediate neighborhoods. Conversely, many critical maintenance associations involve indirect, multi-hop reasoning paths—for example, the transitive logic chain “fault symptom → affected subsystem → responsible component → required maintenance activity → specialized tool” requires traversing multiple intermediate entities. Second, the importance and reliability of different paths vary significantly: some paths represent strong causal relationships grounded in engineering principles, while others may reflect spurious correlations arising from data artifacts. Third, the maintenance domain exhibits high relation heterogeneity, with diverse relation types (causal, compositional, temporal, functional) requiring differentiated modeling strategies.
Traditional knowledge graph link prediction methods struggle to adequately address these domain-specific challenges. Embedding-based models map entities and relations into low-dimensional vector spaces and define scoring functions to measure triple plausibility. While these methods excel at capturing direct binary relationships and demonstrate high computational efficiency, their representational capacity is fundamentally limited to local patterns. They lack explicit mechanisms for modeling multi-hop reasoning paths, making them inadequate for capturing the transitive dependencies prevalent in maintenance knowledge. Graph neural network (GNN)-based approaches address this limitation by aggregating information from multi-hop neighborhoods through iterative message passing. These methods generate entity representations that are contextualized by their local topological structures, thereby improving their ability to capture complex relational patterns. However, traditional GNN architectures are constrained by limited receptive fields—even with multiple layers, they struggle to effectively propagate information across long paths, and deeper architectures often suffer from over-smoothing and gradient vanishing issues.
To explicitly incorporate multi-hop semantic information, recent research has explored path-based and subgraph-based reasoning methods that encode sequences of relations connecting entity pairs and integrate path semantics into scoring functions. These methods enhance interpretability and generalization to unseen entities. Despite these advances, existing path-based methods face critical limitations in path sampling efficiency and quality. Random walk strategies, while computationally tractable, suffer from high blindness—they sample paths uniformly without considering their semantic relevance to the prediction task, leading to the inclusion of many uninformative paths. Meta-path methods, which constrain walks to predefined path templates, require extensive domain expertise and manual specification of path schemas, limiting their flexibility and generalization across diverse maintenance scenarios.
More recent work has attempted to integrate advanced architectures from deep learning and policy learning, employing learnable message passing and global self-attention mechanisms to overcome the locality constraints of traditional methods. While these advanced methods demonstrate state-of-the-art performance on benchmark datasets such as FB15k-237 [
10] and WN18RR [
11], they typically require substantial computational resources and large-scale training data, which may not be readily available in specialized domains like equipment maintenance. Furthermore, they often lack explicit mechanisms for leveraging domain-specific structural priors, such as the hierarchical organization of maintenance procedures or the causal nature of fault propagation.
To address the aforementioned limitations and unlock effective multi-hop reasoning for equipment maintenance knowledge graphs while maintaining computational tractability, this paper proposes an attention-aware path reasoning model that synergistically integrates local structure encoding via graph attention networks with global semantic reasoning via path-based inference. From a graph symmetry perspective, traditional random walk and path reasoning methods implicitly assume path equivalence, treating all reasoning chains connecting entity pairs as equally important regardless of their semantic relevance or structural significance. This symmetry assumption fails to capture the inherent heterogeneity in knowledge graphs where different paths vary substantially in their informativeness for specific prediction tasks. The key innovation of this work lies in breaking this symmetry through a principled attention-guided mechanism: a target-guided biased path sampling strategy that transforms attention weights learned by GAT into probabilistic transition biases, enabling asymmetric path selection that adaptively prioritizes high-quality, task-relevant semantic paths. Unlike blind random walks that maintain path-level symmetry or manually defined meta-paths that require extensive domain expertise, the proposed approach automatically learns to distinguish and emphasize edges that are both locally important (as indicated by GAT attention) and globally relevant (as measured by similarity to target entities). This learned asymmetric weighting reflects the true structural importance and task-specific utility of different reasoning trajectories. The sampled paths are then encoded using bidirectional recurrent neural networks and aggregated via an attention mechanism, generating rich path context representations that capture the most salient reasoning chains. These global path representations are fused with local node embeddings learned by GAT through a gating mechanism, enabling the model to balance complementary information sources. Finally, a DistMult-based scoring function produces triple plausibility scores for link prediction. This symmetry-breaking design not only improves prediction accuracy by focusing computational resources on informative paths, but also enhances model interpretability by revealing which asymmetric structural patterns contribute most strongly to each prediction.
The main contributions of this work can be summarized as follows:
A novel dual-branch architecture is proposed that coherently integrates graph attention networks for local neighborhood encoding with path encoders for global multi-hop reasoning. This unified framework enables the model to synergistically leverage both fine-grained structural patterns and long-range semantic dependencies, effectively addressing the unique challenges of equipment maintenance knowledge graph completion.
An innovative target-guided biased random walk strategy is introduced that, for the first time, explicitly transforms GAT-learned attention weights into probabilistic transition biases for path sampling. Unlike conventional applications of graph attention networks that only employ attention for node representation learning, this work establishes a novel cross-layer mapping from local structural saliency (captured by attention weights) to global reasoning control (manifested in path sampling probabilities). This design enables the model to adaptively focus computational resources on high-quality semantic paths that are both locally salient and globally relevant to prediction targets, significantly improving path sampling efficiency and quality without requiring manual path schema specification or domain expertise.
An attention-based path aggregation mechanism is designed that dynamically weights the importance of different sampled paths and provides interpretability through path attention distributions. By revealing which reasoning chains contribute most strongly to each prediction, the model offers transparent decision support crucial for safety-critical maintenance scenarios where trustworthiness and explainability are paramount.
Comprehensive experiments on benchmark datasets (FB15k-237, WN18RR) and a domain-specific equipment maintenance knowledge graph demonstrate consistent and substantial improvements over state-of-the-art baselines, with the model achieving 0.614 MRR on the maintenance dataset (13.9% gain over CompGCN) and 0.485 MRR on WN18RR (3.6% gain over PTransE). Furthermore, practical deployment in an AR-based intelligent maintenance assistance system validates the approach’s utility for proactive knowledge delivery, demonstrating its value for enhancing operational efficiency and maintenance decision support in real-world industrial applications.
The remainder of this paper is organized as follows.
Section 2 reviews related work on knowledge graph link prediction methods and equipment maintenance applications.
Section 3 presents the technical details of the proposed attention-aware path reasoning model.
Section 4 describes comprehensive experimental evaluation on benchmark datasets and domain-specific knowledge graphs.
Section 5 showcases the practical deployment in an intelligent knowledge push system for AR-assisted equipment maintenance. Finally,
Section 6 concludes the paper and outlines future research directions.
2. Related Work
2.1. Knowledge Graph Link Prediction Methods
The research paradigm of knowledge graph link prediction has undergone significant evolution from embedding-based representation learning to structure-aware models. Early representative works such as TransE [
12], RESCAL [
13], and DistMult [
14] map entities and relations into low-dimensional vector spaces and measure the plausibility of triples through designed scoring functions. ComplEx [
15] extends these approaches by introducing complex-valued embeddings to model asymmetric and inverse relations. Although these methods have advantages of computational efficiency and parameter simplicity, their modeling capability is mainly limited to local binary relations, making it difficult to effectively capture complex multi-hop semantic associations between entities.
To overcome these limitations, graph neural networks [
16] have been introduced into knowledge graph representation learning. Relational graph convolutional networks aggregate multi-hop neighborhood information through message passing mechanisms, generating node representations that are aware of local topological structures. The study of graph connectivity and structural properties [
17,
18,
19] provides theoretical foundations for understanding information propagation in graph-based models. Building on this foundation, graph attention networks [
20] further introduce attention mechanisms to dynamically characterize the importance differences of different neighbor nodes, thereby improving the modeling capability for heterogeneous relations. Advanced GNN architectures such as SACN [
21] integrate weighted graph convolutional networks with Conv-TransE for end-to-end structure-aware modeling, while CompGCN [
22] employs composition-based operations to jointly embed entities and relations. Tensor factorization methods like TuckER [
23] leverage Tucker decomposition to model three-way interactions for knowledge graph completion. Research on graph fault tolerance and diagnostic properties [
24,
25,
26,
27] provides insights into the robustness and reliability of graph-based structures, which is particularly relevant for critical applications such as equipment maintenance systems. However, traditional graph neural network models, limited by their restricted receptive fields, still show insufficient capability in modeling long-range dependencies.
To explicitly model multi-hop semantic associations between entities, reasoning methods based on paths and subgraph structures have gradually attracted attention. PTransE [
28] and similar models encode path sequences connecting entity pairs and integrate path semantics into the link prediction scoring process. SEAL [
29] and GraIL [
30] transform the link prediction problem into a local subgraph classification task centered on target node pairs, enhancing model interpretability while also improving generalization ability to unseen entities. Convolutional approaches such as ConvE [
11] and InteractE [
31] leverage 2D convolutions on entity-relation embeddings to capture expressive interaction patterns. However, these methods still face challenges in path sampling efficiency and quality: random walk strategies often lack targeting, while meta-path methods such as those explored in [
32] rely heavily on domain prior knowledge.
A critical limitation of traditional path-based methods lies in their implicit symmetry assumption regarding path importance. Most existing approaches treat all paths of similar length or structure as equally informative, either through uniform random sampling or predefined path templates. This path-level symmetry fails to capture the inherent heterogeneity in knowledge graphs where different reasoning chains vary significantly in their semantic relevance and reliability. Breaking this symmetry through adaptive, task-dependent path weighting remains an open challenge that requires principled mechanisms to distinguish high-quality semantic paths from spurious correlations without extensive manual schema engineering.
In recent years, researchers have attempted to integrate policy learning and Transformer architectures into knowledge graph reasoning. For example, NBFNet [
33] borrows from the generalized Bellman-Ford algorithm to explicitly model information propagation on all possible paths in a learnable manner; graph structure Transformers such as Graphormer [
34] and TGformer [
35] overcome the locality limitation of traditional GNNs through global self-attention mechanisms, significantly enhancing the model’s perception capability for long-range dependencies and complex structural patterns. These advances provide new ideas and methodological support for deep semantic reasoning in knowledge graphs.
2.2. Application Background in Equipment Maintenance
Modern equipment maintenance is rapidly developing toward intelligence and precision. Augmented reality (AR)-based maintenance assistance systems provide intuitive operation guidance for technicians by overlaying virtual information onto real-world scenes [
6]. However, the effectiveness of such systems highly depends on precise knowledge supply from the backend. Most current systems still adopt keyword or rule-based knowledge push methods, making it difficult to perform deep semantic reasoning and associated knowledge recommendation based on dynamic maintenance contexts (such as fault phenomena and operational steps) [
1].
Recent advances in equipment maintenance have demonstrated the growing importance of intelligent technologies. AI-driven predictive maintenance frameworks [
7,
36] enable early detection of equipment failures through real-time sensor data analysis. IoT-based reliability analysis approaches [
8] enhance system reliability through interconnected monitoring. Maintenance optimization strategies [
37,
38,
39] integrate degradation modeling, spare parts management, and scheduling to balance cost efficiency with equipment reliability. Temporal knowledge graph methods [
40] capture dynamic evolution of maintenance events for predictive decision-making. Advanced optimization frameworks [
41,
42] employ multi-objective approaches and data-driven priority scoring for maintenance resource allocation. Intelligent maintenance strategies also encompass specialized domains including medical equipment [
43], remote maintenance equipment [
44], industrial process equipment [
45], data center cooling systems [
9], distribution grid equipment [
46], unsupervised fault classification [
47], warranty optimization for second-hand equipment [
48], condition-based maintenance with queueing systems [
49], and sustainable port maintenance [
50]. Furthermore, digital twin-embedded approaches [
51] enable self-evolving operational paradigms through cognitive agents and virtual-real co-evolution, representing the frontier of intelligent maintenance methodologies.
To address the above challenges, this research relies on an intelligent maintenance platform construction project for certain equipment types, aiming to build an auxiliary system with knowledge graph-AR dual-end collaboration [
3]. Among them, the knowledge graph end undertakes the core reasoning task, responsible for analyzing the current maintenance process and proactively pushing highly relevant knowledge (such as related fault cases, specialized tools, safety tips, etc.) to the AR terminal for visual presentation [
2], as illustrated in
Figure 1.
This paper focuses on a key technology in this system—link prediction in equipment maintenance knowledge graphs. Its goal is to break through the inherent incompleteness limitations of knowledge graphs, automatically discover potential semantic associations between entities, thereby providing powerful reasoning capability support for upper-layer knowledge push services [
4,
5], achieving an intelligent transformation from "people finding knowledge" to "knowledge finding people".
3. Methodology
This section presents the technical details of the proposed attention-aware path reasoning model for equipment maintenance knowledge graphs. The problem is first defined, followed by an overview of the overall framework and detailed descriptions of each key module.
3.1. Problem Definition
This paper defines a knowledge graph as , where E represents the entity set, R represents the relation set, and is the set of known facts (triples). The task of link prediction is defined as: given a query or , the model needs to predict the missing entity, i.e., find the most likely t or h from all candidate entities. This paper mainly focuses on the prediction scenario. The model’s objective is to learn a scoring function that assigns high scores to true triples and low scores to false triples.
In the equipment maintenance domain, entities E can be instantiated as {fault phenomenon , core component , maintenance activity , maintenance tool , ...}, and relations R can be instantiated as {causes, requires, uses, located_at} and other maintenance-specific relations.
3.2. Framework Overview
The overall workflow of the proposed approach is illustrated in
Figure 2, which depicts the end-to-end pipeline from knowledge graph input through attention-guided path reasoning to knowledge delivery in AR-based maintenance systems.
The proposed model framework, as shown in
Figure 3, adopts a unified architecture that synergistically integrates local structural encoding with global semantic path reasoning. At its foundation, the Graph Attention Encoder learns vector representations that fuse local neighborhood information for all entities and relations, capturing fine-grained relational patterns within the knowledge graph structure. Building upon these attention-aware embeddings, the Target-Guided Path Sampler leverages GAT attention weights as sampling bias to discover meaningful semantic paths between entity pairs, thereby focusing computational resources on high-value reasoning trajectories. The sampled variable-length paths are then processed by the Path Encoding and Aggregator, which encodes them into fixed-length vectors and synthesizes a unified path context representation through attention-weighted aggregation. Finally, the Fusion and Prediction Module integrates the local neighborhood representations with global path context through a gating mechanism and produces final link prediction scores via a scoring function, enabling the model to make informed predictions by considering both immediate structural cues and long-range semantic dependencies.
The workflow of the model can be briefly described as follows: First, the entire knowledge graph is input to the GAT encoder to obtain embeddings for all nodes. For each training triple , using the attention weights provided by the trained GAT, target-guided random walks are performed from h to t to sample a path set . Subsequently, the path encoder maps each path to a vector, and the path aggregator aggregates these path vectors into a global path context vector through attention weights. Finally, the GAT embedding of head entity h is fused with , and together with the relation r embedding and tail entity t GAT embedding, they are input to the scoring function to calculate the score .
3.3. Graph Attention Encoder
This paper employs multi-head graph attention networks to encode local neighborhood information of nodes, whose structure is shown in
Figure 4. For each entity
, its input feature is the initial feature vector
. After passing through one GAT layer, its output feature
is obtained by weighted summation of neighbor node features:
where
is the attention coefficient of node
i to neighbor node
j,
is the neighbor set of node
i,
W is a shared weight matrix,
a is the parameter vector of the attention mechanism, ∥ denotes concatenation operation, and
is a nonlinear activation function. To stabilize the learning process, a multi-head attention mechanism is adopted, concatenating or averaging the outputs of
K heads as the final GAT embedding
of nodes. These learned attention weights
are used for subsequent path sampling strategies. The architecture is illustrated in
Figure 4.
3.4. Target-Guided Path Sampling Strategy
Traditional random walk sampling strategies have high blindness. To solve this problem, this paper designs a target-guided biased random walk strategy, whose core is to use the attention weights learned by GAT as the bias for walk transitions.
For a given head-tail entity pair
, the path sampling algorithm is as follows: Starting from
h, at each step, let the current entity be
. First, obtain all outgoing edge neighbors
of
. Then, calculate the unnormalized probability
of transitioning to each neighbor
:
where
is the attention weight provided by GAT.
is a function measuring the similarity between
and target
t. This paper adopts their cosine similarity in the GAT embedding space:
is a hyperparameter controlling the strength of target guidance. Finally, is normalized to obtain the transition probability distribution, and the next entity is sampled according to this distribution.
The advantage of this strategy is that GAT attention weights ensure that walks tend toward locally important edges, while the target similarity term guides the walk direction toward target entity t, thereby efficiently sampling high-quality, highly relevant semantic paths.
3.5. Path Encoding and Aggregator
For a sampled path
(where
,
), it is encoded into a fixed-length vector
. This paper employs bidirectional GRU as the path encoder:
Here, each node on the path uses its GAT embedding as input, and the final path vector is the concatenation of the bidirectional GRU’s final hidden states.
For an entity pair
,
M paths
are sampled and encoded as
. Not all paths are equally important; therefore, attention mechanisms are introduced for weighted aggregation to obtain the global path context vector
:
where
,
b,
q are learnable parameters. The attention weight
reflects the importance of path
for the current prediction, enhancing model interpretability.
3.6. Fusion and Prediction Module
To synergistically leverage local neighborhood information and global path information, the GAT embedding
of head entity
h is fused with the path context vector
. This paper adopts a gating fusion mechanism to dynamically control information flow:
where
g is the gating vector,
is the sigmoid function, and ⊙ denotes element-wise multiplication. Finally, the fused head entity representation
, relation
r embedding
, and tail entity
t GAT embedding
are input to the scoring function
. This paper selects the DistMult scoring function:
where
denotes trilinear dot product. Model training adopts negative sampling strategy, using margin ranking loss function:
where
T is the positive sample set,
is the corresponding generated negative sample set, and
is the margin hyperparameter.
This chapter provides a detailed introduction to the link prediction model based on GAT attention-guided random walks. The model captures local structure through the GAT encoder, obtains high-quality semantic paths through the innovative target-guided path sampling strategy, obtains global context through the path encoding aggregator, and finally fuses two types of information through a gating mechanism for prediction. This design effectively combines the local and global information of knowledge graphs, providing a powerful tool for knowledge graph completion in equipment maintenance and other fields.
4. Experimental Results and Analysis
4.1. Datasets
This paper selects two widely used public datasets in the field of knowledge graph link prediction: FB15k-237 and WN18RR. Both datasets avoid test leakage problems by removing reversible relations from the original datasets (FB15k and WN18), making the evaluation more challenging and realistic.
Table 1 lists detailed statistics for both datasets.
FB15k-237 is derived from the large-scale general knowledge base Freebase, containing rich real-world facts with diverse relation types and complex structure. WN18RR is derived from the lexical database WordNet, mainly describing semantic relations between words (such as hypernymy, synonym relations), with obvious hierarchical structure characteristics.
4.2. Evaluation Metrics
Following the general evaluation protocol for link prediction tasks, this paper adopts the following ranking metrics:
Mean Reciprocal Rank (MRR) computes the average of reciprocals of ranks assigned to all correct entities, defined as:
where
is the number of test queries and
is the rank of the correct entity for the
i-th query. This metric is particularly sensitive to overall ranking performance.
Hits@N (N=1, 3, 10) measures the proportion of test triples where the correct entity ranks within the top N positions:
where
is the indicator function. Hits@1 measures precision, while Hits@10 measures recall capability.
All metrics adopt filtered settings, i.e., when calculating rankings, all known true triples in the knowledge graph (all true triples in training, validation, and test sets except the currently predicted triple) are filtered out to avoid negative impacts of known facts on ranking results, making evaluation more accurate.
4.3. Ablation Study
To systematically evaluate the contribution of each core component in the proposed framework, comprehensive ablation experiments are conducted on both FB15k-237 and WN18RR datasets. The ablation study is designed to assess four key architectural choices: the necessity of the entire path reasoning module, the effectiveness of GAT attention-guided sampling versus uniform random walk, the benefit of attention-based path aggregation over simple averaging, and the advantage of gating fusion compared to direct concatenation. For each variant, the corresponding component is removed or replaced while keeping all other settings identical to ensure fair comparison.
The results presented in
Table 2 demonstrate the effectiveness of each proposed component. Removing the entire path module leads to the most substantial performance degradation, with MRR dropping by 0.032 (9.1%) on FB15k-237 and 0.044 (9.1%) on WN18RR. This significant decline reveals that global semantic path information provides crucial complementary knowledge beyond the local neighborhood structures captured by GAT, particularly for reasoning about complex multi-hop relations. Without path reasoning, the model degenerates to a pure GAT-based approach, confirming that explicit path information is indispensable for capturing long-range dependencies in knowledge graphs.
Building upon this foundation, the attention-guided sampling mechanism proves equally important. When replaced with uniform random walk, the model experiences notable performance drops across both datasets (0.018 MRR on FB15k-237 and 0.023 on WN18RR). This validates the core hypothesis that leveraging GAT attention weights as sampling bias enables the model to efficiently discover high-quality semantic paths that are most relevant to the prediction task. While uniform sampling distributes computational resources uniformly across all possible paths, the guided strategy adaptively concentrates on promising reasoning trajectories, thereby improving both efficiency and effectiveness.
The attention-based path aggregation mechanism further refines the model’s reasoning capability. Compared to simple average pooling, the attention mechanism yields moderate but consistent improvements (0.009 MRR on FB15k-237 and 0.012 on WN18RR), indicating that sampled paths contribute unequally to the final prediction. The attention weights successfully identify and emphasize the most informative reasoning paths while downweighting less relevant ones. Similarly, the gating fusion mechanism demonstrates its value through marginal yet consistent gains over direct concatenation, adaptively balancing the contributions of local neighborhood information and global path context based on input characteristics.
The ablation study collectively confirms that all proposed components contribute positively to the model’s performance, with the path reasoning module and attention-guided sampling strategy being the most critical innovations. From a symmetry perspective, the performance degradation observed when removing key components reflects not merely the loss of individual modules, but rather the destruction of the learned asymmetric reasoning structure. Specifically, replacing attention-guided sampling with uniform random walk restores path-level symmetry, forcing the model to treat all reasoning chains equally regardless of their structural importance. This symmetry constraint significantly limits the model’s ability to focus on task-relevant paths, explaining the substantial performance decline. Similarly, removing attention-based path aggregation eliminates the learned asymmetric weighting of different reasoning chains, preventing the model from distinguishing high-quality semantic paths from noisy ones. The consistent improvements across both datasets demonstrate the robustness and generalizability of the architectural design choices, and validate the effectiveness of breaking path-level symmetry through principled attention mechanisms.
4.4. Comparative Experiments
To comprehensively evaluate the performance of the proposed model, extensive comparative experiments are conducted against various baseline methods representing different technical paradigms. Representative models are selected including translation-based methods (TransE), bilinear models (DistMult), convolutional approaches (ConvE), graph neural networks (R-GCN), and path-based methods (PTransE). All models are evaluated on both FB15k-237 and WN18RR datasets under identical experimental settings to ensure fair comparison.
Table 3 presents the comparative results, revealing several important observations. The proposed model achieves the best performance across all metrics on both datasets, demonstrating its superiority over various baseline approaches. On FB15k-237, which contains diverse relation types and complex structural patterns from the Freebase knowledge base, the proposed model attains an MRR of 0.350 and Hits@10 of 0.535, outperforming the strongest baseline PTransE by 4.2% and 1.7% respectively. This improvement is particularly significant given that PTransE also leverages path information, suggesting that the attention-guided sampling and neural encoding mechanisms provide more effective path reasoning capabilities than traditional path-based approaches.
The performance gains are even more pronounced on WN18RR, where the proposed model achieves an MRR of 0.485 and Hits@10 of 0.580, surpassing PTransE by 3.6% and 4.1% respectively. The stronger improvements on this dataset can be attributed to its hierarchical semantic structure, which naturally exhibits long-range dependencies through chains of hypernymy and meronymy relations. The target-guided path sampling strategy excels in discovering such multi-hop reasoning paths, while the attention-weighted aggregation mechanism effectively identifies and emphasizes the most semantically coherent paths among multiple candidates.
Compared to pure embedding-based methods (TransE, DistMult), the proposed model demonstrates substantial advantages, with relative improvements exceeding 36% in MRR on FB15k-237 and 167% on WN18RR over TransE. This validates the importance of explicitly modeling structural and path information beyond simple vector space representations. Similarly, while R-GCN leverages graph neural networks to aggregate neighborhood information, it still underperforms the proposed model by 10.4% in MRR on FB15k-237, confirming that local message passing alone is insufficient for complex multi-hop reasoning scenarios. The consistent improvements across both datasets demonstrate the robustness and generalizability of the proposed approach.
Beyond standard benchmarks, the proposed method is evaluated on a domain-specific equipment maintenance knowledge graph to assess its practical applicability. This evaluation is particularly important as domain knowledge graphs often exhibit distinct characteristics compared to general-purpose benchmarks, including sparser connectivity, more specialized relation types, and stronger emphasis on procedural dependencies. Comparisons are made against a comprehensive set of baseline methods spanning different paradigms: translation distance models (TransE), tensor factorization approaches (DistMult, ComplEx, TuckER), convolutional neural networks (ConvE, InteractE), and graph neural networks (SACN, CompGCN). The results are presented in
Table 4.
The results on the equipment maintenance knowledge graph demonstrate the practical effectiveness of the proposed approach in domain-specific scenarios. The proposed method achieves substantial improvements across all metrics, with an MRR of 0.614 and Hits@10 of 0.795, representing 13.9% and 9.1% improvements over the strongest baseline CompGCN respectively. More notably, the Mean Rank (MR) metric shows a dramatic reduction from 338 to 265, indicating that the proposed model ranks correct entities significantly higher on average. This is particularly valuable in practical applications where maintenance personnel need quick access to the most relevant knowledge.
The performance gaps between the proposed method and baselines are generally larger on the domain dataset compared to general benchmarks. For instance, the improvement over CompGCN on the maintenance dataset (13.9% in MRR) exceeds that over PTransE on WN18RR (3.6%). This suggests that domain-specific knowledge graphs, with their characteristic focus on procedural and causal relationships, particularly benefit from the attention-guided path reasoning approach. Traditional methods struggle to capture the complex maintenance workflows and multi-step dependencies inherent in equipment repair scenarios, whereas the proposed model’s ability to discover and reason over semantically coherent paths aligns naturally with the procedural nature of maintenance knowledge.
4.5. Case Analysis
Beyond quantitative performance metrics, specific prediction cases are examined to illustrate the model’s reasoning process and interpretability. Interpretability is crucial for knowledge graph applications, particularly in critical domains like equipment maintenance where users need to understand and trust the system’s recommendations. The proposed model provides interpretability through two mechanisms: the discovered semantic paths and their associated attention weights.
Consider a representative example from WN18RR involving the query (dog, hypernym, ?), which seeks to identify broader categories for the concept "dog". The model successfully predicts correct hypernyms such as "animal" and "entity" by leveraging both local and global reasoning. At the local level, the GAT encoder captures direct relations such as dog→[is_a]→canine, which provides immediate taxonomic information. More interestingly, the path sampler discovers a longer, semantically coherent reasoning chain: dog→[member_meronym]→pack→[hypernym]→group→[hypernym]→abstraction. This four-hop path reveals an alternative reasoning strategy: recognizing that dogs are members of packs, packs are types of groups, and groups are abstractions, the model infers the hierarchical category structure, as illustrated in
Figure 5.
The attention mechanism in the path aggregator assigns this longer path a weight of 0.34, compared to 0.52 for the direct canine path and lower weights (0.08-0.14) for other sampled paths. These attention distributions reveal that while the model primarily relies on the most direct evidence, it also incorporates complementary reasoning paths to strengthen prediction confidence. Without the explicit path reasoning module, a pure embedding-based model would need to compress all this hierarchical information into fixed-dimensional vectors, potentially losing the transparent reasoning structure that the proposed approach preserves.
This case illustrates two key advantages of the proposed model. First, the attention-guided sampling successfully identifies semantically meaningful paths rather than arbitrary random walks—the discovered path follows coherent semantic transitions through membership and taxonomic relations. Second, the attention weights provide actionable explanations that domain experts can verify and trust, making the model suitable for deployment in practical applications where prediction rationale is as important as prediction accuracy.
From a symmetry perspective, the learned path attention weights (0.52 for direct path vs. 0.34 for multi-hop path vs. 0.08-0.14 for others) constitute interpretable asymmetric reasoning evidence. This asymmetric distribution demonstrates that the model does not treat all reasoning chains equally, but rather learns to distinguish between high-confidence direct evidence and supporting contextual paths. The asymmetry is task-dependent and data-driven: for different queries, the model adaptively adjusts the relative importance of various reasoning strategies. This breaks the implicit symmetry assumption in traditional approaches where all paths of similar length would be weighted equally. The learned asymmetric attention mechanism not only improves prediction accuracy but also provides transparent decision support by revealing which structural patterns and reasoning chains contribute most strongly to each specific prediction, thereby enhancing the model’s trustworthiness for safety-critical applications such as equipment maintenance systems.
5. Intelligent Knowledge Push System
The link prediction model constructed in this paper is primarily used for the knowledge push function in equipment intelligent maintenance assistance systems. This section further explains how this model serves as the "core reasoning engine" in practical AR maintenance scenarios, enabling the system to more accurately and timely push relevant knowledge to maintenance personnel.
The basic approach for knowledge push is as follows: The front-end AR device sends the current maintenance context to the knowledge graph backend, such as the component being handled, fault phenomena appearing, or operational steps. The system then maps this contextual information to the knowledge graph and initiates reasoning queries accordingly. Through the path reasoning capability provided by link prediction, the system can not only find explicitly existing associations in the graph but also discover knowledge content that is not directly recorded but logically relevant, thereby achieving more intelligent and practical knowledge recommendation.
To verify the application capability of the proposed link prediction model in real scenarios, this section demonstrates its role in driving intelligent knowledge push services through a typical equipment maintenance process case. This case is based on the constructed equipment maintenance knowledge graph and simulates part of the workflow from inspection to knowledge push, as illustrated in
Figure 6.
As shown in
Figure 6, the task workflow consists of dust cover removal, wheel hub mounting on rack, and end cap removal. Each step of the process matches corresponding nodes in the graph and performs relation reasoning, ultimately returning knowledge nodes with high similarity to the process. The task first performs dust cover removal, at which time knowledge highly similar to the current task will be returned. When the process proceeds to the next step, knowledge highly similar to the next step will be pushed, and so on, completing knowledge push for the entire process. The specific workflow of the push process is shown in
Figure 7.
The knowledge push interface on the AR end demonstrates the system’s practical application. When the maintenance process reaches the "wheel hub mounting on rack" stage, the system first identifies and matches corresponding nodes in the knowledge graph based on the current process context. Subsequently, the proposed model performs link prediction reasoning on the "required tools" relation based on this node: on one hand, it recalls tools already existing in the graph, and on the other hand, it predicts candidate tool nodes that are highly relevant but not explicitly recorded in the graph. Finally, the system returns TOP-K highly relevant tools pushed to the AR interface, effectively assisting maintenance personnel in preparing tools in advance and improving operational efficiency.
6. Conclusions
This paper addresses the challenge that traditional link prediction methods struggle to effectively capture both local graph structural information and global multi-hop semantic path information in knowledge graph link prediction tasks. A knowledge graph link prediction model based on GAT attention-guided random walks is proposed. The core of this work lies in designing a unified framework consisting of a graph attention encoder, target-guided path sampler, path encoding and aggregator, and fusion prediction module. The most important innovation of this research is proposing a target-oriented path sampling strategy guided by GAT attention weights. This strategy creatively transforms the local neighborhood importance weights learned by GAT into transition probability bias in the path sampling process, enabling the sampling process to adaptively focus on high-quality semantic paths highly relevant to prediction targets, thereby significantly improving the utilization rate and quality of path information and the model’s reasoning capability.
From the perspective of graph symmetry modeling, this work makes a fundamental contribution by breaking the implicit symmetry assumption in traditional path-based reasoning approaches. While conventional methods treat paths uniformly or rely on manually defined schemas, the proposed attention-guided mechanism introduces asymmetric structural weighting that reflects both local topological importance and global task relevance. This asymmetry is not arbitrary but learned in a principled manner through graph attention networks, enabling the model to automatically discover and prioritize semantically meaningful reasoning chains. The learned path attention distributions provide interpretable asymmetric reasoning evidence, revealing which structural patterns contribute most strongly to each prediction. This symmetry-breaking perspective elevates the model’s innovation from a performance improvement to a principled approach for modeling structural heterogeneity in graph reasoning tasks.
Experiments on the two standard datasets FB15k-237 and WN18RR, as well as domain-specific datasets, demonstrate that the proposed model outperforms various mainstream baseline models in key metrics such as MRR, Hits@1, Hits@3, and Hits@10. Ablation experiments also verify the effectiveness and necessity of each core component of the model. Furthermore, the case analysis demonstrates the model’s capability in multi-hop reasoning using complex paths and its ability to provide decision-making basis through path attention, offering core support for knowledge push and other applications.
The proposed model effectively combines local and global information of knowledge graphs, providing powerful tools for knowledge graph completion in equipment maintenance and other fields. Future research directions include: (1) exploring more sophisticated path representation learning methods; (2) extending the model to temporal knowledge graphs; (3) investigating the model’s transferability across different domains; (4) integrating external knowledge sources to further enhance reasoning capabilities.
Author Contributions
Conceptualization, M.C. and Y.B.; methodology, M.C.; software, M.C.; validation, M.C. and Y.B.; formal analysis, M.C.; investigation, M.C.; resources, Y.B.; data curation, M.C.; writing—original draft preparation, M.C.; writing—review and editing, M.C. and Y.B.; visualization, M.C.; supervision, Y.B.; project administration, Y.B.; funding acquisition, Y.B. All authors have read and agreed to the published version of the manuscript.
Funding
Please add funding information here.
Data Availability Statement
Conflicts of Interest
The authors declare no conflict of interest.
References
- Zhang, Z.; Yu, J.; Yang, B.; Du, K.; Wang, S.; Qi, X. A knowledge graphs construction method enhanced by multimodal large language model for industrial equipment operation and maintenance. Advanced Engineering Informatics 2025, 68, 103705. [Google Scholar] [CrossRef]
- Qi, X.; Yang, B.; Wang, S.; Zhang, Z.; Zhang, Y.; Du, K. Few-shot and chain-of-thought prompting for equipment maintenance knowledge graph construction via large language models. Knowledge-Based Systems 2026, 335, 115266. [Google Scholar] [CrossRef]
- Li, X.; Li, G.; Wang, T.; Dong, Z.; Qin, Z.; Chu, F. Knowledge extraction and retrieval-augmented generation for intelligent maintenance of wind power equipment based on graph attention networks. Chinese Journal of Mechanical Engineering 2025, 100141. [Google Scholar] [CrossRef]
- Wang, G.; Li, Y.; Cheng, C.X.; Li, R.; Ding, C.X.; Zhang, Y.; Liu, Y. Integrated optimization of equipment degradation modeling and spare parts inventory for predictive maintenance in power systems. Sustainable Energy Technologies and Assessments 2025, 83, 104626. [Google Scholar] [CrossRef]
- Yuan, J.; Lei, Y.; Li, N.; Yang, B.; Li, X.; Chen, Z.; Han, W. A framework for modeling and optimization of mechanical equipment considering maintenance cost and dynamic reliability via deep reinforcement learning. Reliability Engineering & System Safety 2025, 264, 111424. [Google Scholar] [CrossRef]
- xin Wang, Z.; cheng Zeng, Q.; Haralambides, H. Shift of emphasis toward intelligent equipment maintenance in port operations: A critical review of emerging trends and challenges. Ocean & Coastal Management 2024, 259, 107408. [Google Scholar] [CrossRef]
- Tang, Y.; Zhou, Y.; Wu, T.; Wang, C.; Li, Z.; Li, K. AI-driven predictive maintenance for medical imaging equipment: A deep learning framework based on the IoMT data. Reliability Engineering & System Safety 2026, 270, 112152. [Google Scholar] [CrossRef]
- Dui, H.; Zhang, A.; Zhou, Y.; Tang, C. IoT-based reliability analysis and maintenance strategy of unmanned equipment communication systems. Reliability Engineering & System Safety 2026, 266, 111628. [Google Scholar] [CrossRef]
- Huang, J.; Zou, S.; Li, Y.; Xiao, T.; Zhang, J. Performance and maintenance strategy of a data center cooling system under equipment aging. Applied Thermal Engineering 2025, 281, 128623. [Google Scholar] [CrossRef]
- Toutanova, K.; Chen, D. Observed versus latent features for knowledge base and text inference. In Proceedings of the Proceedings of the 3rd workshop on continuous vector space models and their compositionality, 2015; pp. 57–66. [Google Scholar]
- Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2d knowledge graph embeddings. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2018; Vol. 32. [Google Scholar]
- Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Proceedings of the Advances in neural information processing systems, 2013; Vol. 26. [Google Scholar]
- Nickel, M.; Tresp, V.; Kriegel, H.P. A three-way model for collective learning on multi-relational data. Proceedings of the ICML 2011, Vol. 11, 3104482–3104584. [Google Scholar]
- Yang, B.; Yih, W.T.; He, X.; Gao, J.; Deng, L. Embedding entities and relations for learning and inference in knowledge bases. arXiv 2014, arXiv:1412.6575. [Google Scholar]
- Martínez, V.; Berzal, F.; Cubero, J.C. A survey of link prediction in complex networks. ACM computing surveys 2016, 49, 1–33. [Google Scholar] [CrossRef]
- Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the European semantic web conference, Cham, 2018; pp. 593–607. [Google Scholar]
- Wang, S.; Wang, Y.; Wang, M. Connectivity and matching preclusion for leaf-sort graphs. Journal of Interconnection Networks 2019, 19, 1940007. [Google Scholar] [CrossRef]
- Wang, M.; Lin, Y.; Wang, S. The connectivity and nature diagnosability of expanded k-ary n-cubes. RAIRO-Theoretical Informatics and Applications-Informatique Théorique et Applications 2017, 51, 71–89. [Google Scholar] [CrossRef]
- Wang, S.; Wang, M. A Note on the Connectivity of m-Ary n-Dimensional Hypercubes. Parallel Processing Letters 2019, 29, 1950017. [Google Scholar] [CrossRef]
- Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar] [PubMed]
- Shang, C.; Tang, Y.; Huang, J.; Bi, J.; He, X.; Zhou, B. End-to-end structure-aware convolutional networks for knowledge base completion. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence 2019, Vol. 33, 3060–3067. [Google Scholar] [CrossRef]
- Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P. Composition-based multi-relational graph convolutional networks. arXiv 2019, arXiv:1911.03082. [Google Scholar]
- Balažević, I.; Allen, C.; Hospedales, T. Tucker: Tensor factorization for knowledge graph completion. In Proceedings of the Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), 2019; pp. 5185–5194. [Google Scholar]
- Wang, M.; Yang, W.; Guo, Y.; Wang, S. Conditional fault tolerance in a class of Cayley graphs. International Journal of Computer Mathematics 2016, 93, 67–82. [Google Scholar] [CrossRef]
- Wang, M.; Ren, Y.; Lin, Y.; Wang, S. The tightly super 3-extra connectivity and diagnosability of locally twisted cubes. American Journal of Computational Mathematics 2017, 7, 127–144. [Google Scholar] [CrossRef]
- Wang, M.; Xu, S.; Jiang, J.; Xiang, D.; Hsieh, S.Y. Global reliable diagnosis of networks based on Self-Comparative Diagnosis Model and g-good-neighbor property. Journal of Computer and System Sciences 2025, 103698. [Google Scholar] [CrossRef]
- Wang, M.; Yang, W.; Wang, S. Conditional matching preclusion number for the Cayley graph on the symmetric group. Acta Math. Appl. Sin. (Chinese Series) 2013, 36, 813–820. [Google Scholar]
- Lin, Y.; Liu, Z.; Luan, H.; Sun, M.; Rao, S.; Liu, S. Modeling relation paths for representation learning of knowledge bases. arXiv 2015, arXiv:1506.00379. [Google Scholar] [CrossRef]
- Zhang, M.; Chen, Y. Link prediction based on graph neural networks. In Proceedings of the Advances in neural information processing systems, 2018; Vol. 31. [Google Scholar]
- Teru, K.; Denis, E.; Hamilton, W. Inductive relation prediction by subgraph reasoning. In Proceedings of the International conference on machine learning. PMLR, 2020; pp. 9448–9457. [Google Scholar]
- Vashishth, S.; Sanyal, S.; Nitin, V.; Agrawal, N.; Talukdar, P. InteractE: Improving convolution-based knowledge graph embeddings by increasing feature interactions. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2020; Vol. 34, 3009–3016. [Google Scholar] [CrossRef]
- Research on Knowledge Graph Construction Technology for Equipment Maintenance and Diagnosis. Master’s thesis, 2023. [CrossRef]
- Zhu, Z.; Zhang, Z.; Xhonneux, L.P.; Tang, J. Neural bellman-ford networks: A general graph neural network framework for link prediction. Advances in neural information processing systems 2021, 34, 29476–29490. [Google Scholar]
- Ying, C.; Cai, T.; Luo, S.; Zheng, S.; Ke, G.; He, D.; Liu, T.Y. Do transformers really perform badly for graph representation? Advances in neural information processing systems 2021, 34, 28877–28888. [Google Scholar]
- Shi, F.; Li, D.; Wang, X.; Li, B.; Wu, X. Tgformer: A graph transformer framework for knowledge graph embedding. IEEE Transactions on Knowledge and Data Engineering, 2024. [Google Scholar]
- Bu Sinnah, Z.A. Explainable AI-driven predictive maintenance for mitigating process safety risks in safety-critical industrial equipment. Journal of Loss Prevention in the Process Industries 2026, 100, 105907. [Google Scholar] [CrossRef]
- Yuan, J.; Lei, Y.; Yang, B.; Li, N.; Li, X.; Chen, Z.; Han, W. A hybrid maintenance strategy for equipment with competitive failure modes: Sudden failure and multi-stage degradation failure. Mechanical Systems and Signal Processing 2025, 234, 112846. [Google Scholar] [CrossRef]
- Wang, J.; Zhang, Z.; Huang, Y.; Hu, H. Proactive dispatch of rail maintenance equipment under weather-induced disruptions: an optimization framework integrating machine learning-driven risk assessment. Advanced Engineering Informatics 2026, 69, 103990. [Google Scholar] [CrossRef]
- Lu, L.; Deng, Q.; Luo, Q.; Hu, Y.; Zhang, J.; Deng, K. Integrated optimization of dynamic flexible job shop scheduling with AGV breakdowns and multi-site equipment maintenance on the demand side. Computers & Industrial Engineering 2025, 208, 111361. [Google Scholar] [CrossRef]
- Lu, J.; Chen, H.; Chen, J.; Xiao, Z.; Li, R.; Xiao, G.; Wang, Q. Temporal knowledge graph fusion with neural ordinary differential equations for the predictive maintenance of electromechanical equipment. Knowledge-Based Systems 2025, 317, 113450. [Google Scholar] [CrossRef]
- Yang, Y.; Song, W.; Wang, J. A high-dimensional many-objective co-optimization method on equipment utilization and maintenance. Array 2026, 29, 100672. [Google Scholar] [CrossRef]
- Wehbi, L.; Bhulai, S.; Slik, J. Optimizing maintenance of heavy equipment: A data-driven approach. Computers & Industrial Engineering 2026, 211, 111645. [Google Scholar] [CrossRef]
- Gandhare, S.; Kumar, P.; Madankar, T.; Singh, D.; Bhamu, J. Development and validation of a FMEA-based medical equipment maintenance framework using Industry 4.0. International Journal of Quality & Reliability Management 2025, 42, 2250–2290. [Google Scholar] [CrossRef]
- Benzoni, G.; Truong, V.D.; Brace, W.; Cammi, A.; Introini, C.; Miccichè, G.; Mikko, S.; Tripodo, C. Petri net modelling – A remote maintenance simulation approach for fusion remote maintenance equipment concepts assessment. Fusion Engineering and Design 2025, 215, 114996. [Google Scholar] [CrossRef]
- Nadim, K.; Ragab, A.; Ghezzaz, H.; Ouali, M.S. Reliable Dynamic Causality Analysis for Efficient Prescriptive Maintenance of Degraded Industrial Equipment. Journal of Loss Prevention in the Process Industries 2026, 105922. [Google Scholar] [CrossRef]
- Wang, T.; Hu, G.; Cho, S.; Tan, S.W.; Chiang, K.L. A reliability-centered maintenance process for equipment in distribution grids based on extended virtual age model. Energy 2025, 328, 135998. [Google Scholar] [CrossRef]
- Bhardwaj, A.S.; Veeramani, D. An unsupervised multi-class ensemble classifier for identifying equipment failure mechanisms from maintenance records. Reliability Engineering & System Safety 2025, 264, 111410. [Google Scholar] [CrossRef]
- Darghouth, M.N. An integrated two-dimensional warranty framework for second-hand equipment considering condition-based maintenance, upgrades, and past life. Journal of Modelling in Management 2025, 20, 2136–2166. [Google Scholar] [CrossRef]
- Luo, Z.; Lian, Z.; Li, J.; Yang, Z. Maintenance strategies for a queueing system with a standby equipment: An integrated approach using condition-based monitoring. Reliability Engineering & System Safety 2025, 263, 111219. [Google Scholar] [CrossRef]
- Sepehri, A.; Kirichek, A.; van den Heuvel, M.; de Geus, M.; van Koningsveld, M. Sustainable port maintenance: dredging equipment selection in time-emission trade-offs. Results in Engineering 2026, 109247. [Google Scholar] [CrossRef]
- Cong, W.; Tao, W.; Jinsong, B. Agentic digital twin-embedded maintenance methodology for energy equipment: A self-evolving operational paradigm. Journal of Manufacturing Systems 2026, 84, 100–116. [Google Scholar] [CrossRef]
Figure 1.
Knowledge graph and AR-based maintenance assistance system showing the integration of process recognition, entity embedding, graph reasoning, and knowledge delivery to AR terminals.
Figure 1.
Knowledge graph and AR-based maintenance assistance system showing the integration of process recognition, entity embedding, graph reasoning, and knowledge delivery to AR terminals.
Figure 2.
Overall workflow diagram showing the knowledge graph-based maintenance knowledge push system architecture. The system begins with equipment maintenance knowledge graph construction, proceeds through entity recognition and graph embedding, performs reasoning via the proposed attention-guided path model to complete missing links, and finally delivers actionable knowledge to AR terminals for maintenance technicians.
Figure 2.
Overall workflow diagram showing the knowledge graph-based maintenance knowledge push system architecture. The system begins with equipment maintenance knowledge graph construction, proceeds through entity recognition and graph embedding, performs reasoning via the proposed attention-guided path model to complete missing links, and finally delivers actionable knowledge to AR terminals for maintenance technicians.
Figure 3.
Proposed model framework. The model consists of four main components: Graph Attention Encoder for local structure encoding, Target-Guided Path Sampler for semantic path discovery, Path Encoder and Aggregator for path representation learning, and Fusion and Prediction Module for final scoring. The path information and context vector flow through the system to generate predictions .
Figure 3.
Proposed model framework. The model consists of four main components: Graph Attention Encoder for local structure encoding, Target-Guided Path Sampler for semantic path discovery, Path Encoder and Aggregator for path representation learning, and Fusion and Prediction Module for final scoring. The path information and context vector flow through the system to generate predictions .
Figure 4.
Graph Attention Network (GAT) architecture for equipment maintenance knowledge graph. The GAT processes the input knowledge graph through transformation layers and multi-head attention mechanisms to generate node embeddings.
Figure 4.
Graph Attention Network (GAT) architecture for equipment maintenance knowledge graph. The GAT processes the input knowledge graph through transformation layers and multi-head attention mechanisms to generate node embeddings.
Figure 5.
Case analysis example showing multi-hop reasoning for the query (dog, hypernym, ?). The model discovers both a direct path through "canine" (shown with dashed line) and a longer semantic path through "pack", "group", and "abstraction", ultimately predicting "animal" and "entity" as correct hypernyms. Solid arrows represent the discovered multi-hop reasoning chain.
Figure 5.
Case analysis example showing multi-hop reasoning for the query (dog, hypernym, ?). The model discovers both a direct path through "canine" (shown with dashed line) and a longer semantic path through "pack", "group", and "abstraction", ultimately predicting "animal" and "entity" as correct hypernyms. Solid arrows represent the discovered multi-hop reasoning chain.
Figure 6.
Knowledge push case diagram showing the three-stage maintenance process: dust cover disassembly, mounting the hub onto the disassembly bench, and end cover disassembly. For each task process step, the system identifies relevant KG nodes and pushes corresponding knowledge such as required tools.
Figure 6.
Knowledge push case diagram showing the three-stage maintenance process: dust cover disassembly, mounting the hub onto the disassembly bench, and end cover disassembly. For each task process step, the system identifies relevant KG nodes and pushes corresponding knowledge such as required tools.
Figure 7.
Detailed knowledge push workflow. The system processes task steps through entity recognition, matches them with KG nodes, performs graph embedding and reasoning, generates candidate knowledge, and finally pushes relevant information to the AR interface.
Figure 7.
Detailed knowledge push workflow. The system processes task steps through entity recognition, matches them with KG nodes, performs graph embedding and reasoning, generates candidate knowledge, and finally pushes relevant information to the AR interface.
Table 1.
Dataset statistics for FB15k-237 and WN18RR benchmark datasets. All columns show counts (no directional preference).
Table 1.
Dataset statistics for FB15k-237 and WN18RR benchmark datasets. All columns show counts (no directional preference).
| Dataset |
Entities |
Relations |
Training |
Validation |
Test |
| FB15k-237 |
14,541 |
237 |
272,115 |
17,535 |
20,466 |
| WN18RR |
40,943 |
11 |
86,835 |
3,034 |
3,134 |
Table 2.
Ablation study results on FB15k-237 and WN18RR datasets. Metrics are reported as MRR↑ / Hits@10↑ (higher is better). Bold indicates the complete model with all components.
Table 2.
Ablation study results on FB15k-237 and WN18RR datasets. Metrics are reported as MRR↑ / Hits@10↑ (higher is better). Bold indicates the complete model with all components.
| Model Variant |
FB15k-237 |
WN18RR |
| Complete Model |
0.350 / 0.535 |
0.485 / 0.580 |
| w/o Path Module |
0.318 / 0.492 |
0.441 / 0.528 |
| w/o GAT Attention Guidance (uniform random walk) |
0.332 / 0.508 |
0.462 / 0.549 |
| w/o Path Attention Aggregation (average pooling) |
0.341 / 0.521 |
0.473 / 0.562 |
| w/o Gating Fusion (direct concatenation) |
0.345 / 0.526 |
0.478 / 0.570 |
Table 3.
Link prediction performance comparison on FB15k-237 and WN18RR test sets. All metrics: higher is better (↑). Bold indicates the best performance.
Table 3.
Link prediction performance comparison on FB15k-237 and WN18RR test sets. All metrics: higher is better (↑). Bold indicates the best performance.
| Model |
FB15k-237 |
WN18RR |
| MRR↑ |
H@1↑ |
H@3↑ |
H@10↑ |
MRR↑ |
H@1↑ |
H@3↑ |
H@10↑ |
| TransE |
0.257 |
0.174 |
0.284 |
0.441 |
0.182 |
0.027 |
0.295 |
0.501 |
| DistMult |
0.281 |
0.198 |
0.310 |
0.477 |
0.425 |
0.390 |
0.440 |
0.504 |
| ConvE |
0.325 |
0.237 |
0.356 |
0.521 |
0.430 |
0.400 |
0.440 |
0.520 |
| R-GCN |
0.317 |
0.231 |
0.348 |
0.507 |
0.455 |
0.421 |
0.470 |
0.542 |
| PTransE |
0.336 |
0.245 |
0.369 |
0.526 |
0.468 |
0.435 |
0.482 |
0.557 |
| Proposed Model |
0.350 |
0.260 |
0.382 |
0.535 |
0.485 |
0.452 |
0.497 |
0.580 |
Table 4.
Comparative results on equipment maintenance domain knowledge graph dataset. MR↓: Mean Rank (lower is better), other metrics↑: higher is better. Bold indicates the best performance.
Table 4.
Comparative results on equipment maintenance domain knowledge graph dataset. MR↓: Mean Rank (lower is better), other metrics↑: higher is better. Bold indicates the best performance.
| Model Type |
Model Name |
MR↓ |
MRR↑ |
H@1↑ |
H@3↑ |
H@10↑ |
| Translation Distance |
TransE |
624 |
0.326 |
0.215 |
0.387 |
0.512 |
| Tensor Factorization |
DistMult |
535 |
0.401 |
0.294 |
0.456 |
0.593 |
| Tensor Factorization |
ComplEx |
578 |
0.369 |
0.258 |
0.423 |
0.559 |
| Tensor Factorization |
TuckER |
419 |
0.478 |
0.372 |
0.534 |
0.676 |
| CNN |
ConvE |
492 |
0.427 |
0.321 |
0.482 |
0.625 |
| CNN |
InteractE |
456 |
0.453 |
0.348 |
0.507 |
0.651 |
| GNN |
SACN |
375 |
0.506 |
0.405 |
0.568 |
0.702 |
| GNN |
CompGCN |
338 |
0.539 |
0.437 |
0.596 |
0.729 |
| Proposed Method |
Proposed Method |
265 |
0.614 |
0.516 |
0.673 |
0.795 |
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).