Preprint
Article

This version is not peer-reviewed.

Unifying Ontology Construction and Semantic Alignment for Deterministic Enterprise Reasoning at Scale

Submitted:

12 March 2026

Posted:

13 March 2026

You are already at the latest version

Abstract
While enterprises amass vast quantities of data, much of it remains chaotic and effectively dormant, preventing decision-making based on comprehensive information. Existing neuro-symbolic approaches rely on disjoint pipelines and struggle with error propagation. We introduce the large ontology model (LOM), a unified framework that seamlessly integrates ontology construction, semantic alignment, and logical reasoning into a single end-to-end architecture. LOM employs a construct-align-reason (CAR) pipeline, leveraging its unified architecture across all three stages: it first autonomously constructs a domain-specific ontological universe from raw data, then aligns neural generation with this structural reality using a graph-aware encoder and reinforcement learning, and finally executes deterministic reasoning over the constructed topology, node attributes and relation types. We evaluate LOM on a comprehensive benchmark constructed from diverse real-world enterprise datasets. Experimental results demonstrate that LOM-4B achieves 88.8% accuracy in ontology completion and 94% in complex graph reasoning tasks, significantly outperforming state-of-the-art LLMs. These findings validate that autonomous logical construction is essential for achieving deterministic, enterprise-grade intelligence.
Keywords: 
;  ;  ;  ;  ;  

1. Introduction

Contemporary enterprise AI faces a critical bottleneck in the prevalence of data silos that hinder the effective deployment of ontology intelligence at scale. Bridging the gap between academic advances and large-scale enterprise application remains a significant challenge, as real-world data is often fragmented and noisy. We address this by proposing an end-to-end framework that unifies ontology construction, semantic alignment, and deterministic reasoning within a single coherent architecture, paired with robust engineering practices for reliable deployment at scale.
Existing approaches often employ disparate model architectures to handle the distinct phases of construction, alignment, and reasoning, leading to significant engineering complexity and fragmentation. Furthermore, large language models exhibit impressive creativity, but often fail in deterministic reasoning tasks, particularly when applied to complex, noisy enterprise environments. We identify this limitation as a probabilistic wall, where scaling parameters alone yield diminishing returns for logical consistency. Standard LLMs frequently suffer from information loss during context graph transfer and remain constrained by a probabilistic wall that limits deterministic reasoning. To address these challenges, we unify these processes under a single cohesive model architecture. Critically, our approach to alignment extends beyond mere multimodal understanding to emphasize the dynamic capability of ontology updates, ensuring the model remains synchronized with evolving enterprise data.
A critical bottleneck lies in the dynamic evolution of the ontology—specifically, the non-trivial task of predicting latent relationships within sparse topological structures populated by entities with dense, unstructured textual attributes. In complex enterprise environments, the probabilistic nature of standard LLMs often fails to capture the strict deterministic dependencies required for valid ontology updates, leading to hallucinated links or missed connections. To overcome this, we leverage a graph-aware encoder [1] that explicitly models the topological context of each node, enabling the inference of missing relationships through a synergy of semantic content and structural patterns. This approach necessitates a rigorous text-ontology alignment training phase, where the model is optimized via a link prediction objective. By training to predict the existence and type of edges between nodes based on their textual descriptions and local subgraph structure, we force the model to ground its linguistic fluency in the logical reality of the graph. This alignment ensures that the inferred updates are not only semantically coherent but also topologically sound.
The second challenge involves adapting the LOM [2] architecture for high-fidelity reasoning within the inherently noisy and ambiguous landscape of real-world enterprise data. In such environments, nodes are often laden with dense, heterogeneous textual attributes, and queries require a delicate balance between content-based filtering and structure-based deduction. A standard supervised learning approach often falls short in generalizing across this spectrum of tasks, struggling to distinguish subtle query nuances or to execute multi-hop reasoning chains without drifting. To this end, we incorporate reinforcement learning into our training pipeline. By defining a reward function that penalizes logical fallacies and rewards correct final answers and validated intermediate steps, we enable the model to autonomously discover optimal reasoning procedures (solution-step trajectories). This allows LOM to dynamically adapt its strategy—deeply integrating topological structure, node semantic content, and edge weights or types—to generate comprehensive solutions. Whether focusing on semantic matching for attribute-heavy queries or topological traversal for structural inquiries, LOM achieves robust performance across diverse tasks ranging from simple information retrieval to complex weighted graph algorithms like PageRank and minimum spanning tree.
Finally, the relentless velocity of AI advancement poses a strategic dilemma: engineering investments risk obsolescence almost as soon as they are deployed. To navigate this volatility and forecast the trajectory of intelligence, we present a theoretical analysis in section 4 that maps the evolution of AI onto a 10-Dimensional Cognitive Framework. This framework transcends linear scaling laws, proposing instead that breakthroughs occur through dimensional ascension—fundamental shifts in how information is structured and processed. We posit that current LLMs are hitting a 6D probabilistic wall, characterized by high-dimensional statistical correlation without grounding. By formalizing the criteria for ascending to the 7th dimension—where autonomous ontological construction precedes reasoning—we offer not just a retrospective taxonomy, but a predictive roadmap. This analysis provides a falsifiable theoretical basis for LOM’s architecture and serves as a strategic guide for future research, distinguishing between transient engineering optimizations and enduring dimensional leaps.
We validate the effectiveness of our approach through comprehensive experiments on a benchmark constructed from real-world enterprise data, demonstrating significant improvements over existing baselines. The main contributions are as follows:
1. Real-world enterprise CAR methodology: Using authentic enterprise data at scale, we constructed an enterprise benchmark, trained LOM models, and conducted comprehensive evaluations—completing the full pipeline from benchmark construction to model training and assessment for deployment.
2. Full-stack ontology intelligence at scale: We complete the construct-align-reason pipeline over real enterprise data and surpass much larger models at every stage—88.8% link prediction accuracy, high-fidelity semantic and structural alignment, and 95% accuracy in deterministic reasoning across 19 diverse graph tasks.
3. 10-dimensional cognitive mapping: We locate contemporary techniques along a 1D–10D spectrum, characterize the 6D probabilistic wall, and formalize 7D logic autonomy where ontology construction precedes reasoning; this offers falsifiable criteria for dimensional ascension and practical guidance for building production systems beyond leaderboard metrics.
The remainder of this paper is organized as follows. Section 2 reviews related work in ontology construction, neuro-symbolic AI, and the development trends of agentic systems. Section 3 details the LOM architecture and the construct-align-reason pipeline. Section 4 presents the 10-Dimensional Cognitive Framework, providing a theoretical mapping for predicting technological evolutionary trends. Section 5 describes the experimental setup and results, demonstrating LOM’s performance on enterprise benchmarks. Finally, Section 6 concludes the paper with a discussion of future directions.

3. Approach

We introduce the LOM [2] which bridges the gap between probabilistic generation and deterministic reasoning through a unified construct-align-reason pipeline. Unlike traditional approaches that treat these as separate engineering tasks, LOM integrates them into a cohesive cognitive process. This section details how LOM autonomously builds its own logical universe in the construction phase, grounds abstract structures in semantic reality in the alignment phase, and executes rigorous algorithms within this self-created framework in the reasoning phase. Specifically, section 3.1 elaborates on the generative mechanisms for ontology construction from both unstructured and structured data. section 3.2 discusses the dual-encoder architecture for semantic-structural alignment and dynamic updates. section 3.3 demonstrates how LOM performs high-fidelity deterministic reasoning. Finally, section 3.4 outlines the two-stage training pipeline, including structural pre-training and reinforcement learning fine-tuning.

3.1. Construct: Autonomous Ontology Synthesis

The foundation of LOM is the ability to autonomously construct a high-fidelity ontology from structured and unstructured enterprise data. Unlike traditional methods that rely on rule-based extraction or manual curation, LOM employs a generative mechanism.
Ontology construction from unstructured data: This pipeline leverages a multi-stage LLM-driven architecture to transform unstructured enterprise text into a formal ontology. After initial element extraction using LOM with pydantic-enforced JSON schemas, the framework pivots to a logic-intensive hierarchy construction phase. Here, LOM is employed to perform semantic entailment, identifying latent subclass-superclass relationships. To ensure the structural integrity of the ontology, an iterative validation mechanism is applied, where the model systematically reasons through inheritance chains to verify their logical consistency and soundness. Finally, the validated conceptual model is serialized into the OWL turtle format, providing a robust, machine-readable knowledge foundation.
Ontology construction from structured data: This pipeline implements an iterative framework to systematically transform relational database schemas into formal OWL turtle ontologies. The process begins with a structured schema traversal, processing tables in an order determined by foreign key constraints to maintain relational integrity. For each element, a RAG module fetches semantic context from existing core ontologies and external documentation. LOM then synthesizes these inputs into delta ontology fragments, which undergo rigorous verification and optimization to ensure semantic consistency and syntactic compliance. This cycle repeats, incrementally merging validated fragments until all tables, columns, and relationships are exhaustively mapped into a logically sound global ontology.
The final phase of our approach employs a systematic four-stage process to merge disparate ontological structures into a unified OWL 2 DL model. First, concept alignment utilizes embedding similarity to match equivalent classes and properties, identifying semantic overlaps between the two source ontologies. Next, conflict resolution addresses naming collisions and hierarchical contradictions, leveraging LLMs to determine priority and merging strategies. During Hierarchy Integration, we establish cross-ontology relationships using rdfs:subClassOf and owl:equivalentClass axioms to create a cohesive taxonomic structure. Finally, the process concludes with Validation and Export, where reasoners ensure logical consistency and competency questions verify domain coverage before the final ontology is exported.

3.2. Align: Semantic-Structural Fusion

Our alignment mechanism fulfills two critical functions. First, it enables ontology understanding by statically grounding topological structures in real-world business semantics, ensuring that abstract graph nodes reflect concrete operational entities. Second, it facilitates ontology update, allowing for the dynamic adaptation of the graph structure in response to real-time data influxes, thereby maintaining synchronization with the evolving enterprise environment.
This capability for dynamic adaptation is particularly critical in real-time business environments, where the ontology must evolve instantaneously with the arrival of new data. Traditional batch-processing methods often fail to meet this requirement due to inherent latency. By extending the alignment mechanism to support incremental state changes, LOM allows the graph structure to be updated continuously based on the current ontology state. This ensures that the physical laws governing the information universe adapt in real-time to new empirical evidence, preserving the ontology’s fidelity and relevance.
LOM conceptualizes the enterprise environment as a sparse graph, where nodes symbolize entities enriched with dense textual attributes and edges denote latent relationships. Through training on link prediction—specifically, identifying missing edges—LOM learns to deduce the underlying logical structure of the business domain. This mechanism effectively collapses the probabilistic noise inherent in raw data into a deterministic structural representation. In the context of our 10-Dimensional Cognitive Framework, this process signifies the transition from the 5th to the 6th dimension, as the model advances from perceiving a multiverse of potential connections to selecting the most logically consistent configuration.
Formally, we frame link prediction as a semantic-structural alignment task. Distinct from methods that serialize graphs into textual prompts, we employ a dedicated graph encoder E ϕ to extract dense topological embeddings from the observed subgraph G o b s . These embeddings are projected into the LOM latent space. The LOM then synthesizes this structural context with the semantic attributes of the target nodes u and v to predict the existence of a relationship:
y = LOM θ ( E ϕ ( G o b s ) , u , v )
where y denotes the predicted relationship status, E ϕ represents the graph encoder parameterizing the structural information, and LOM θ serves as the reasoning core that integrates topological features with semantic understanding. Consequently, the predicted adjacency entry is defined by the indicator function:
A ^ u v = I ( y = yes )
where A ^ u v represents the predicted adjacency matrix entry for the edge between nodes u and v, and I ( · ) is the indicator function which equals 1 if the model output y corresponds to a positive confirmation ("yes") and 0 otherwise.
This generative formulation distinguishes LOM from traditional embedding models that rely on static scalar scoring. By treating link prediction as context-aware reasoning, the model effectively decides on structural validity based on semantic evidence. Crucially, this inference capability supports the dynamic evolution of the ontology. As new data increments Δ D t arrive, the graph state is recursively updated:
G t + 1 = Update ( G t , Δ D t ; θ a l i g n )
where G t + 1 and G t denote the graph states at time steps t + 1 and t respectively. Δ D t represents the incremental data influx, and Update ( · ) is the state transition function parameterized by the alignment model weights θ a l i g n . This mechanism ensures that LOM operates as a living system, continuously synchronizing its internal logical structure with the shifting reality of the enterprise environment.

3.3. Reason: Logic Execution Within Self-Constructed Axioms

While LLMs demonstrate remarkable capability in simulating reasoning through probabilistic token prediction, they fundamentally lack a binding logical framework, often leading to hallucinations in complex scenarios. Despite Claude’s remarkable heuristic reasoning [26] on this particular directed Hamiltonian cycle problem, its ability to tackle diverse open questions and highly complex graph structures remains to be fully proven. The definitive 7D capability of LOM lies not in the reasoning process itself, but in its execution within a self-constructed axiomatic system. By treating the autonomously built ontology as a set of immutable physical laws, LOM transforms reasoning from a stochastic approximation into a rigorous, deterministic traversal. This ensures that every conclusion—whether derived via PageRank, minimum spanning tree, or shortest path algorithms—is mathematically guaranteed by the model’s internal worldview, achieving a level of consistency unattainable by purely probabilistic architectures.
The reasoning process can be formalized as finding the optimal path or subgraph G * G that satisfies a logical query Q. For a deterministic algorithm A (e.g., Dijkstra), the output is:
Result = A ( G a l i g n e d , Q )
Unlike LLMs where Result P ( tokens | Q ) , LOM ensures that the execution of A is bound by the constraints of the constructed ontology G a l i g n e d . This guarantees that if the ontology is correct, the reasoning result is mathematically certain.
Beyond Retrieval: The Distinction from GraphRAG It is crucial to distinguish LOM from GraphRAG. GraphRAG utilizes graphs as a retrieval index—organizing information topologically to help LLMs locate relevant text chunks. Its function is to enhance information access as external knowledge, but the final reasoning is still performed by the LLM’s probabilistic next-token prediction engine. In contrast, LOM treats the ontology as logic laws, intrinsic intelligence. It does not merely read the graph; it executes deterministic algorithms on the graph structure itself. While GraphRAG allows an LLM to see more context, LOM provides the AI with a rigorous logical framework that strictly constrains its reasoning process.
Convergent Evolution with AlphaGeometry LOM shares a fundamental neuro-symbolic DNA with DeepMind’s AlphaGeometry and AlphaProof. Both paradigms recognize that pure neural networks (LLMs) provide intuition but lack rigor, while pure symbolic systems provide rigor but lack scalability. AlphaGeometry combines a neural intuition engine with a symbolic deduction engine to solve mathematical proofs within a pre-existing axiomatic system. LOM applies this same fusion to the enterprise domain. However, LOM addresses a challenge of higher complexity: unlike mathematics, where axioms are pre-defined, enterprise axioms (business logic) are hidden within chaotic data. LOM must first autonomously construct the axiomatic universe (via the construct phase) before it can reason within it. Thus, LOM represents the same species of intrinsic intelligence evolution, adapted for the unstructured reality of the business world.

3.4. Model Training

Our training pipeline orchestrates a progressive evolution, beginning with semantic-structural alignment to ground the graph encoder within a deterministic ontology, and culminating in reinforcement learning optimization to fine-tune the logical reasoning capabilities of LOM-R1. In the foundational phase of semantic alignment and link prediction, we bridge the semantic-structural gap by aligning a pre-trained graph encoder [27] with the large language model. This alignment mechanism grounds abstract semantic representations within the concrete topological reality of the enterprise graph. Subsequently, we deploy a link prediction objective, optimized via cross-entropy, compelling the model to infer latent connections based on the synthesis of node attributes and existing graph structure. This process refines the graph encoder to capture deterministic dependencies, maximizing the likelihood of valid edges within the constructed ontology:
L c o n s t r u c t = L C E ( A ^ , A ) = ( u , v ) E log P ( A u v = 1 | G )
where L c o n s t r u c t denotes the construction loss, A ^ and A represent the predicted and ground-truth adjacency matrices respectively, E is the set of edges in the training graph, and P ( A u v = 1 | G ) signifies the probability of an edge existing between nodes u and v given the current graph state G . This optimization ensures that the model first grounds semantic meaning before learning complex topological dependencies.
Following the structural grounding, we advance to logical reasoning optimization via reinforcement learning. To address the complexity of real-world queries—which often involve subtle attribute filtering and multi-hop weighted reasoning—we introduce reinforcement learning to fine-tune the LOM-R1 model. We employ the GSPO algorithm [28] for model training. Specifically, we utilize vLLM to generate n samples for each input query Q. These samples are evaluated using a reward reshaping function that scores both the final answer accuracy and the logical validity of the reasoning path.
The training objective is formulated as the maximization of the expected reward J ( θ ) across the distribution of generated reasoning trajectories:
J ( θ ) = E y π θ ( · | Q ) [ R ( y ) ]
where the reward function R ( y ) is designed to balance result correctness with procedural soundness. It integrates a binary accuracy indicator with a continuous logical validity score, ensuring that the model is incentivized not only to retrieve the correct answer but to derive it through a rigorous inferential chain:
R ( y ) = α · I ( y a n s = y g o l d ) + β · Score l o g i c ( y p a t h )
where I ( · ) represents the indicator function, which equals 1 if the predicted answer y a n s matches the gold standard y g o l d , and 0 otherwise. Score l o g i c ( y p a t h ) quantifies the validity of the reasoning path, derived from the coherence of the graph traversal steps. The hyperparameters α and β control the trade-off between answer accuracy and reasoning quality.
Furthermore, our optimization strategy incorporates a comparative analysis of logit distributions across diverse inference trajectories [29]. This enables the model to discern and converge upon robust, deterministic logical patterns, effectively transcending the "probabilistic wall" inherent in traditional stochastic prediction.

4. The 10-Dimensional Map of AI Evolution

The theoretical foundation of our analysis is adapted from physicist Rob Bryanton’s "Imagining the tenth dimension" [30], which describes the abstract structure of the universe across hierarchical levels. When this framework is mapped onto the evolutionary history of artificial intelligence, it reveals a startlingly accurate and predictive landscape of cognitive development.
In the current era of rapid AI proliferation, traditional metrics—such as parameter count or benchmark accuracy—often fail to capture qualitative leaps in intelligence. They measure linear growth within a single paradigm, for example "better next-token prediction", but cannot distinguish between a model that effectively retrieves information in 4D and one that autonomously structures knowledge in 7D. We posit that a dimensional framework is essential to:
1. Map the cognitive territory: Precisely locate current technologies such as LLMs and agents on an evolutionary spectrum rather than a flat leaderboard.
2. Identify the probabilistic wall: Reveal why scaling laws are hitting a ceiling at the 6th dimension, where models can navigate possibilities but cannot alter the underlying logical rules.
3. Predict the trajectory: Clarify that the path to AGI requires a dimensional ascension—mastering the ability to autonomously construct the logical "physics" of a problem—rather than merely accelerating within existing dimensions.
We contextualize existing AI paradigms within this framework below.
Table 1. A 10-Dimensional Map of AI Evolution.
Table 1. A 10-Dimensional Map of AI Evolution.
Dimension Spatial Concept AI Evolutionary Stage Representative Form
1D-3D Line/Plane/Volume Statistical Learning Rule-based Systems, Machine Learning, Deep Learning
4D Time, Causal Sequences LLMs Transformers
5D All Possible Timelines Basic Agents Planning, Hypothesis Reasoning
6D Jumping Between Timelines Meta-Cognitive Agents Self-Reflection, Tool/Skill Building
7D Different Physical Constants LOM Autonomous Logic Reconstruction
8D-10D Multiverse/Omniscience ASI Paradigm Discovery, Singularity

4.1. From 1D to 4D: The Evolution of Knowledge Representation

The journey of AI can be mapped as a progressive conquest of spatial dimensions, representing increasing degrees of freedom in data representation.
1D, line: rule-based systems. Early AI operated on linear if-then logic. Like a point moving along a single line, these systems could only follow pre-defined paths. They were deterministic but rigid, unable to handle any deviation from the hard-coded rules.
2D, plane: traditional machine learning. With the advent of statistical learning such as SVMs and random forests, AI gained the ability to optimize decision boundaries across a two-dimensional plane of features. This allowed for generalization beyond exact matches, but representation remained flat and feature-dependent.
3D, volume: deep learning. Neural networks introduced depth through layers, allowing models to learn hierarchical representations of data. This three-dimensional understanding enabled the capture of complex, non-linear patterns in static inputs like images or fixed vectors.
4D, time: large language models. The transformer architecture introduced the fourth dimension, time as sequence. LLMs do not just process static data; they understand the causal flow of tokens, predicting the future t + 1 based on the history t n . This mastery of sequential probability allows them to generate fluent, context-aware text. However, their reasoning is fundamentally temporal—a probabilistic continuation of a sequence—rather than spatial or structural. When strictly causal chains are broken or when circular logic is required, as in many graph problems, they often hallucinate because they lack a stable, time-independent logical structure.

4.2. From 5D to 6D: The Ceiling of Meta-Cognition

The concept of metacognition [31], originally proposed in 1976 by developmental psychologist John Flavell, describes the capability of thinking about thinking—monitoring and regulating one’s own cognitive processes. In our framework, this distinguishes 6D intelligence from 5D.
5D, all possible timelines: basic agents. Systems like AutoGPT or BabyAGI operate in the fifth dimension. They can generate multiple plans, often called timelines, to solve a task and select the optimal decision path based on predictive outcomes. However, they lack the ability to step outside their current execution path to evaluate its validity against a higher-order framework.
6D, jumping between timelines: meta-cognitive agents. Advanced agents exhibit 6D behavior by not just executing tasks, but creating the means to execute them. This includes: (1) Scaffolding: decomposing complex goals into structured sub-processes to guide reasoning, such as OpenClaw [25]. (2) Cockpits: interfaces for real-time monitoring and intervention, such as HULA [21], allowing for dynamic strategy adjustment. (3) Agentic code reasoning: instead of directly answering a query, systems like Devin, Glean, SWE-agent [23] write and execute code as tools to derive the answer.
This shift—from directly solving requirements to creating tools that solve requirements, and ultimately to building systems that mass-produce such tools—marks the peak of current AI capabilities. Yet, it hits the "probabilistic wall." Even a 6D agent is ultimately a probabilistic engine wrapping another probabilistic engine. It can optimize how it navigates the search space, but it cannot guarantee the correctness of the underlying logical laws; for example, it might write code that is syntactically correct but logically flawed due to a misunderstanding of the domain ontology. It cannot "rewrite the physics" of the problem; it can only build better vehicles to traverse it.

4.3. The 7D Breakthrough: LOM’s Logic Autonomy

In the 7th dimension, we deal with different physical constants or, in the AI context, different logical frameworks. A 6D agent can explore all possibilities within a given world, for example a specific game or business process, but it cannot rewrite the rules of that world.
LOM operates in the 7th dimension because it autonomously constructs the universe, the ontology, in which it reasons. By defining the entities [32], relations [33], and constraints, the physical constants of the enterprise data, LOM creates a deterministic environment where logic is not just probable, but structural and binding. Just as our physical universe follows laws like Newton’s First Law, every enterprise operates on its own set of intrinsic axioms—business rules, regulatory constraints, and operational workflows. However, unlike physics, these enterprise axioms are rarely summarized into clear mathematical formulas; they are hidden within chaotic data and tacit knowledge. The capability of autonomous ontology construction is essentially the process of exploring and formalizing these hidden axioms, transforming an opaque enterprise into a computable entity. This capability allows LOM to solve problems that are impossible for 4D, 5D, and 6D models, such as finding the exact shortest path in a graph or strictly adhering to a complex supply chain policy.
Fundamentally, LOM’s CAR pipeline is not merely a sequence of steps but a unified cognitive mechanism that embodies 7D intelligence.
Integrated understanding and reasoning: Enterprise data is a duality of unstructured text, which requires semantic understanding, and structural relationships, which require logical reasoning. Traditional pipeline systems treat these separately—calling an NLP model for text and a graph model for structure—resulting in information loss and low efficiency. LOM integrates these via the Align phase, ensuring that every logical deduction in the Reason phase is grounded in the full semantic context of the node attributes.
End-to-end internalization: By autonomously constructing the ontology in the construct phase and grounding it semantically in the Align phase, LOM internalizes the entire problem space. It does not need to repeatedly query external modules or databases; it thinks within a self-contained, consistent universe. This internalization is what allows for deep, multi-hop reasoning that considers all available information simultaneously, rather than fragmenting the problem into isolated sub-tasks.
Efficiency over assembly: Unlike agentic workflows that rely on assembling disparate tools, which is slow and error-prone, LOM’s end-to-end capability enables it to execute complex reasoning with the speed and coherence of a single model, overcoming the efficiency bottlenecks of piecemeal system integration.

4.4. Reflection: Acceleration vs. Ascension

This mapping compels us to reflect on the current trajectory of AI development from a 10-dimensional perspective.
Acceleration within 4D: Most efforts in scaling LLMs, more parameters and more data, are essentially accelerating within the fourth dimension. They are building a faster, more fluent predictor of the next token, mastering the probability of sequences.
Navigation within 5D and 6D: Agentic workflows such as CoT and ReAct and tool-use frameworks are navigating the fifth and sixth dimensions more efficiently. They allow systems to explore multiple plans and create tools to traverse the problem space. However, they remain trapped within the probabilistic nature of the base model, the probabilistic wall, unable to guarantee the logical validity of the universe they operate in.
Ascension to 7D: LOM represents a dimensional ascension. By autonomously constructing the ontology and internalizing the problem space through the CAR pipeline, LOM builds the internal combustion engine, a 7D logic engine, rather than a faster horse, that is, better 4D and 5D models. It defines the physical constants of the enterprise, transforming chaotic data into a deterministic, computable entity.
This analysis suggests a critical distinction for future research: optimizing performance within current paradigms versus enabling structural shifts to higher-dimensional reasoning. We argue that achieving robust general intelligence necessitates moving beyond data scaling alone, requiring architectures capable of autonomous logical construction.

5. Experiments

5.1. Dataset

Our benchmark dataset is collected directly from internal production systems. These datasets encompass comprehensive business scenarios including human resources, finance, assets, production, supply chain, and sales—HR, finance, supply chain, and manufacturing—integrating both structured databases and unstructured documentation. This shift from synthetic to real-world data was pivotal.
Our benchmark dataset consists of 19 diverse graph reasoning tasks, designed to test both structural and semantic understanding. For each core reasoning task, we constructed a balanced dataset comprising 500 training samples and 100 testing samples. Additionally, for the link prediction task in ontology construction, we utilized a significantly larger dataset consisting of 8,500 real-world business graphs for training and 1,000 graphs for testing. These samples were derived from real-world enterprise scenarios, ensuring the model’s performance reflects practical applicability rather than theoretical benchmarks.

5.2. Experimental Setup

We trained the LOM-4B, LOM-32B, and LOM-Construct models on a distributed computing cluster with a global batch size of 64. The training infrastructure comprised four servers, each equipped with 8 × NVIDIA A800 80GB GPUs, 2 × Intel Xeon CPUs, and 1TB RAM, totaling 32 GPUs. All models were trained with full precision, FP32, to ensure maximum fidelity in capturing the subtle nuances of the text-attributed graphs.

5.3. Evaluation Metrics

The primary metric for all tasks is accuracy, defined as the percentage of test samples where the model’s output exactly matches the ground truth.

5.4. Ontology Completion Performance

We first validated the model’s capacity for autonomous ontology completion. On this challenging test set, LOM achieved an accuracy of 88.8% in link prediction. It is important to note that unlike clean academic benchmarks, these graphs are derived from raw enterprise data containing significant noise and unstructured text attributes. Achieving such high fidelity in this context is crucial; it ensures that the "universe" in which the model reasons is a faithful representation of the underlying business logic, rather than a hallucinated structure. This step effectively filters out the noise inherent in the raw data, providing a solid 6D foundation for the subsequent 7D reasoning tasks.

5.5. Graph Reasoning Performance

We conducted a comprehensive evaluation across 19 graph reasoning tasks, comparing LOM-4B and LOM-32B against large-scale LLMs, including Doubao-1.8, Qwen3-Max, and DeepSeek-V3.2. As shown in Figure 1, LOM delivers consistently strong performance across topology, path/structure, algorithmic, and semantic tasks, achieving average accuracies of 0.93 (LOM-4B) and 0.94 (LOM-32B). Scaling from 4B to 32B improves several challenging tasks while preserving structural reliability; for example, Shortest Path increases from 0.84 to 0.88, MST from 0.89 to 0.92, Cycle Detection from 0.96 to 0.98, and Summary from 0.77 to 0.88, as shown in Figure 2.
In contrast, general LLMs exhibit a pronounced bifurcation: they can remain competitive on surface-level semantic tasks—for example, Doubao-1.8 achieves 0.99 on description—yet collapse on tasks requiring deterministic structural constraints. On shortest_path, Qwen3-Max scores 0.08 and DeepSeek-V3.2 scores 0.09, and on MST, Qwen2.5-32B scores 0. This gap highlights the necessity of 7D-style logic execution within an internal ontology, rather than purely probabilistic next-token prediction.

5.6. Analysis and Discussion

  • Unified Semantics and Structure. The results validate the CAR architecture. Unlike pipeline systems that fragment semantic parsing and graph reasoning, LOM internalizes both end-to-end. This integration allows LOM to maintain state-of-the-art performance on semantic tasks (e.g., description 1.00) while excelling at structure-sensitive tasks (e.g., relation path 0.98), where traditional models often falter. The align phase ensures semantic outputs are grounded in the same rigorous ontology used for reasoning, preventing the trade-off between fluency and structural accuracy.
  • Overcoming the Probabilistic Wall. A clear boundary emerges in algorithmic tasks requiring strict constraint satisfaction. LOM achieves high accuracy on MST (0.92) and shortest path (0.88), whereas general LLMs collapse (e.g., Qwen2.5-32B scores 0 on MST). This confirms the 7D claim: probabilistic scaling alone cannot master deterministic logic. Without an internal binding structure, even large models fail to preserve multi-step invariants, illustrating a probabilistic wall that parameter scaling does not breach.
  • Logic Density over Parameter Scale. LOM demonstrates that logic density outweighs parameter count for complex reasoning. LOM-4B (93% average accuracy) and LOM-32B (94%) significantly outperform models 10x their size (e.g., Qwen2.5-32B, 70B+ models). By constraining the model with strict ontological rules (logic laws), LOM achieves higher cognitive density, proving that neuro-symbolic integration is a more efficient path to reliable reasoning than unsupervised pre-training alone.

6. Conclusions

In this paper, we introduced the LOM, a neuro-symbolic framework that unifies ontology construction, semantic alignment, and logical reasoning to address the inherent limitations of probabilistic models in deterministic domains. Through the proposed CAR pipeline, LOM autonomously formalizes latent knowledge into explicit ontological structures, effectively overcoming the probabilistic wall characterized in our 10-Dimensional cognitive framework. Empirical evaluations on large-scale enterprise benchmarks demonstrate that LOM not only achieves superior performance on complex reasoning tasks compared to models with significantly larger parameter counts but also ensures logical consistency and interpretability. Our findings indicate that equipping language models with the capability for autonomous logic construction—7D logic autonomy—is a critical direction for advancing the reliability and utility of AI in high-stakes enterprise applications.
Building upon this 7D logic foundation, our future work will extend LOM’s capabilities into autonomous action execution, strategic decision-making, and dynamic skill acquisition. This evolution will transition LOM from a passive reasoning engine to an active agent capable of intervening in real-world enterprise processes, effectively closing the loop between insight and impact.

References

  1. Zhu, H. Node classification via semantic-structural attention-enhanced graph convolutional networks. arXiv 2024, arXiv:2403.16033. [Google Scholar]
  2. Zhang, Y.; Zhu, H. Construct, Align, and Reason: Large Ontology Models for Enterprise Knowledge Management. arXiv 2026, arXiv:2602.00029. [Google Scholar]
  3. Oyewale, A.; Soru, T. LLM-Driven Ontology Construction for Enterprise Knowledge Graphs. arXiv 2026, arXiv:2602.01276. [Google Scholar] [CrossRef]
  4. Nayyeri, M.; Yogi, A.A.; Fathallah, N.; Thapa, R.B.; Tautenhahn, H.M.; Schnurpel, A.; Staab, S. Retrieval-Augmented Generation of Ontologies from Relational Databases. arXiv 2025, arXiv:2506.01232. [Google Scholar] [CrossRef]
  5. Luyen, L.N.; Abel, M.H.; Gouspillou, P. Development of Ontological Knowledge Bases by Leveraging Large Language Models. arXiv 2026, arXiv:2601.10436. [Google Scholar] [CrossRef]
  6. Cheung, B. Generative Ontology: When Structured Knowledge Learns to Create. arXiv 2026, arXiv:2602.05636. [Google Scholar] [CrossRef]
  7. Zhu, H.; Peng, H.; Lyu, Z.; Hou, L.; Li, J.; Xiao, J. Pre-training language model incorporating domain-specific heterogeneous knowledge into a unified representation. Expert Systems with Applications 2023, 215, 119369. [Google Scholar] [CrossRef]
  8. West, P.; Bhagavatula, C.; Hessel, J.; Hwang, J.; Jiang, L.; Le Bras, R.; Lu, X.; Welleck, S.; Choi, Y. Symbolic knowledge distillation: from general language models to commonsense models. In Proceedings of the Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022, pp. 4602–4625.
  9. Acharya, K.; Velasquez, A.; Song, H.H. A survey on symbolic knowledge distillation of large language models. IEEE Transactions on Artificial Intelligence 2024, 5, 5928–5948. [Google Scholar] [CrossRef]
  10. Liao, H.; He, S.; Xu, Y.; Zhang, Y.; Liu, K.; Zhao, J. Neural-symbolic collaborative distillation: Advancing small language models for complex reasoning tasks. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2025, Vol. 39, pp. 24567–24575. [CrossRef]
  11. Marra, G.; Dumančić, S.; Manhaeve, R.; De Raedt, L. From statistical relational to neurosymbolic artificial intelligence: A survey. Artificial Intelligence 2024, 328, 104062. [Google Scholar] [CrossRef]
  12. Wolfram, S. An elementary introduction to the Wolfram Language. No Title 2015. [Google Scholar]
  13. Sarker, M.K.; Zhou, L.; Eberhart, A.; Hitzler, P. Neuro-symbolic artificial intelligence: Current trends. Ai Communications 2022, 34, 197–209. [Google Scholar] [CrossRef]
  14. Kautz, H. The third AI summer: AAAI Robert S. Engelmore memorial lecture. Ai magazine 2022, 43, 105–125. [Google Scholar] [CrossRef]
  15. Trinh, T.; Luong, T. Alphageometry: An olympiad-level ai system for geometry. Google DeepMind 2024, 17, 1. [Google Scholar]
  16. Hubert, T.; Mehta, R.; Sartran, L.; Horváth, M.Z.; Žužić, G.; Wieser, E.; Huang, A.; Schrittwieser, J.; Schroecker, Y.; Masoom, H.; et al. Olympiad-level formal mathematical reasoning with reinforcement learning. Nature 2025, 1–3. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, L.; Ma, C.; Feng, X.; Zhang, Z.; Yang, H.; Zhang, J.; Chen, Z.; Tang, J.; Chen, X.; Lin, Y.; et al. A survey on large language model based autonomous agents. Frontiers of Computer Science 2024, 18, 186345. [Google Scholar] [CrossRef]
  18. Hong, S.; Zhuge, M.; Chen, J.; Zheng, X.; Cheng, Y.; Wang, J.; Zhang, C.; Wang, Z.; Yau, S.K.S.; Lin, Z.; et al. MetaGPT: Meta programming for a multi-agent collaborative framework. In Proceedings of the The twelfth international conference on learning representations, 2023. [Google Scholar]
  19. Yao, S.; Zhao, J.; Yu, D.; Du, N.; Shafran, I.; Narasimhan, K.R.; Cao, Y. React: Synergizing reasoning and acting in language models. In Proceedings of the The eleventh international conference on learning representations, 2022. [Google Scholar]
  20. Shinn, N.; Cassano, F.; Gopinath, A.; Narasimhan, K.; Yao, S. Reflexion: Language agents with verbal reinforcement learning. Advances in neural information processing systems 2023, 36, 8634–8652. [Google Scholar]
  21. Takerngsaksiri, W.; Pasuksmit, J.; Thongtanunam, P.; Tantithamthavorn, C.; Zhang, R.; Jiang, F.; Li, J.; Cook, E.; Chen, K.; Wu, M. Human-in-the-loop software development agents. In Proceedings of the 2025 IEEE/ACM 47th International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP). IEEE, 2025, pp. 342–352.
  22. Zhang, J.; Xiang, J.; Yu, Z.; Teng, F.; Chen, X.; Chen, J.; Zhuge, M.; Cheng, X.; Hong, S.; Wang, J.; et al. Aflow: Automating agentic workflow generation. arXiv 2024, arXiv:2410.10762. [Google Scholar] [CrossRef]
  23. Yang, J.; Jimenez, C.E.; Wettig, A.; Lieret, K.; Yao, S.; Narasimhan, K.; Press, O. Swe-agent: Agent-computer interfaces enable automated software engineering. Advances in Neural Information Processing Systems 2024, 37, 50528–50652. [Google Scholar]
  24. Wang, X.; Li, B.; Song, Y.; Xu, F.F.; Tang, X.; Zhuge, M.; Pan, J.; Song, Y.; Li, B.; Singh, J.; et al. Openhands: An open platform for ai software developers as generalist agents. arXiv 2024, arXiv:2407.16741. [Google Scholar]
  25. Steinberger, P.; OpenClaw Contributors. OpenClaw: Your own personal AI assistant. Any OS. Any Platform. The lobster way. GitHub repository. https://github.com/openclaw/openclaw, 2025. Accessed: 2026-03-07.
  26. Knuth, D.E. Claude’s Cycles. Stanford University Computer Science Department Paper, 2026. Revised March 6, 2026.
  27. Zhu, H.; Li, Y.; Liu, L.; Tong, H.; Lin, Q.; Zhang, C. Pre-training graph autoencoder incorporating hierarchical topology knowledge, 2025.
  28. Zheng, C.; Liu, S.; Li, M.; Chen, X.H.; Yu, B.; Gao, C.; Dang, K.; Liu, Y.; Men, R.; Yang, A.; et al. Group sequence policy optimization. arXiv 2025, arXiv:2507.18071. [Google Scholar] [PubMed]
  29. Liu, C.Y.; Zeng, L.; Liu, J.; Yan, R.; He, J.; Wang, C.; Yan, S.; Liu, Y.; Zhou, Y. Skywork-reward: Bag of tricks for reward modeling in llms. arXiv 2024, arXiv:2410.18451. [Google Scholar] [CrossRef]
  30. Bryanton, R. Imagining the tenth dimension: a new way of thinking about time, space, and string theory; Talking Dog Studios, 2006.
  31. Flavell, J.H. Metacognitive aspects of problem solving. In The nature of intelligence; Routledge, 2024; pp. 231–236.
  32. Zhu, H.; Hu, W.; Zeng, Y. Flexner: A flexible lstm-cnn stack framework for named entity recognition. In Proceedings of the CCF International Conference on Natural Language Processing and Chinese Computing. Springer, 2019, pp. 168–178.
  33. Zhu, H.; Tiwari, P.; Zhang, Y.; Gupta, D.; Alharbi, M.; Nguyen, T.G.; Dehdashti, S. SwitchNet: A modular neural network for adaptive relation extraction. Computers and Electrical Engineering 2022, 104, 108445. [Google Scholar] [CrossRef]
Figure 1. Comparative Accuracy on Enterprise Ontology Tasks.
Figure 1. Comparative Accuracy on Enterprise Ontology Tasks.
Preprints 202762 g001
Figure 2. Heatmap of Per-Task Accuracy Across Models.
Figure 2. Heatmap of Per-Task Accuracy Across Models.
Preprints 202762 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated