Preprint
Concept Paper

This version is not peer-reviewed.

ESDM–SMTJ: An Entropic Semantic Dynamics Model for Classical Probabilistic Hardware with Superparamagnetic Tunnel Junctions

Submitted:

01 December 2025

Posted:

02 December 2025

You are already at the latest version

Abstract
We introduce ESDM–SMTJ, an Entropic Semantic Dynamics Model implemented on clas- sical probabilistic hardware based on superparamagnetic tunnel junctions (SMTJs). The model represents the internal state of a symbolic or cognitive system as a trajectory Σ(τ ) in a layered state space, with τ ∈ [0, 1] interpreted as an internal computation time from initial query to final answer. Each expression e (for example 2 + 2 =?) induces a program-specific dynamics U τ e that iteratively updates Σ(τ ). Ambiguous operators such as “+” are treated as multi-modal : every occurrence admits a finite family of semantic modes i, and an entropic gate scores each mode by the predicted reduction ∆H(k) i of the output entropy if that mode is selected at position k. These scores are mapped to effective energy levels E(k) i = E0 − κ∆H(k) i in a local SMTJ p-bit block, whose Boltzmann statistics implement a softmax distribution over modes at the hardware level. The resulting dynamics exhibits rumination (high-entropy plateaus), insight-like transi- tions (sharp entropy drops) and stabilization in low-entropy attractors, together with a natural notion of semantic commit at an internal time τc < 1 and a blind reveal of the output token via SMTJ readout at τf ≈ 1. We illustrate how simple arithmetical judgements—including rare anomalies such as 2 + 2 → 5 under mis-tuned parameters—can be expressed in this frame- work, and we outline a quantum extension in which semantic modes become basis states of a Hamiltonian with complex amplitudes instead of classical probabilities.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

Many models of symbolic reasoning and cognition treat operators such as “+” as having a fixed meaning that is applied deterministically to well-defined operands. However, both human judgement and contemporary large language models exhibit a richer behaviour: operators are polysemic, internal processing involves uncertainty and revision, and intermediate states can be far from Boolean truth values. At the same time, advances in probabilistic spintronic hardware (superparamagnetic tunnel junctions, SMTJs) have made it possible to implement physically the kind of stochastic dynamics traditionally simulated in Boltzmann machines and related models.
On the hardware side, SMTJ-based probabilistic computers have been framed as a concrete realization of early proposals by Feynman and Hinton for stochastic, physics-inspired computation [1]. Energy-efficient stochastic computing primitives based on SMTJs have been analysed in detail, highlighting their suitability as low-power p-bit sources and stochastic logic elements [2], while broader surveys of unconventional and neuromorphic computing with magnetic tunnel junctions review their use in probabilistic networks, Ising/Boltzmann machines and related architectures [3]. In parallel, SMTJs operated in the superparamagnetic regime have been shown to provide true random bit streams at nanosecond time scales that pass standard NIST statistical tests, with explicit applications to Monte Carlo simulation and neuromorphic learning circuits [4]. Taken together, these works establish SMTJs as compact, CMOS-compatible entropy sources and stochastic computing primitives.
Figure 1. High-level architecture of the ESDM–SMTJ pipeline. An input expression (e.g. 2 + 2 = ? ) is mapped to a semantic trajectory Σ ( τ ) under a program-specific dynamics U e τ . At each occurrence of an ambiguous operator such as “+”, an entropic gate evaluates the expected entropy drop Δ H i ( k ) for each semantic mode i and sets corresponding energy levels E i ( k ) in an SMTJ p-bit block. The Boltzmann statistics of the SMTJs implement a softmax over modes, feeding back a concrete mode choice that drives Σ ( τ ) towards a low-entropy attractor (semantic commit). A final readout and decoding of SMTJ states realizes a blind reveal of the output token.
Figure 1. High-level architecture of the ESDM–SMTJ pipeline. An input expression (e.g. 2 + 2 = ? ) is mapped to a semantic trajectory Σ ( τ ) under a program-specific dynamics U e τ . At each occurrence of an ambiguous operator such as “+”, an entropic gate evaluates the expected entropy drop Δ H i ( k ) for each semantic mode i and sets corresponding energy levels E i ( k ) in an SMTJ p-bit block. The Boltzmann statistics of the SMTJs implement a softmax over modes, feeding back a concrete mode choice that drives Σ ( τ ) towards a low-entropy attractor (semantic commit). A final readout and decoding of SMTJ states realizes a blind reveal of the output token.
Preprints 187713 g001
In this work we propose ESDM–SMTJ, an Entropic Semantic Dynamics Model that couples a high-level semantic state space with low-level probabilistic hardware.
The internal state of a system is represented as a trajectory Σ ( τ ) in a layered state space, where τ [ 0 , 1 ] parametrizes an internal computational time. Expressions such as 2 + 2 = ? define program-specific dynamics U e τ that update Σ ( τ ) while ambiguous operators (e.g. “+”) are treated as multi-modal and resolved via an entropic gating mechanism.
At the hardware level, we assume an SMTJ-based p-bit substrate whose Boltzmann statistics implement softmax sampling over a finite set of modes. For each occurrence of an ambiguous operator, the semantic layer computes expected entropy drops Δ H i ( k ) associated with choosing mode i at position k. These values are converted into energy levels E i ( k ) for a local SMTJ block, which in turn feeds back a sampled mode i * . This loop drives the trajectory Σ ( τ ) through phases of rumination, insight and stabilization, and leads naturally to notions of semantic commit and blind reveal of the final token.

Relation to SMTJ-based TRNG and Monte Carlo simulators.

Conceptually, ESDM–SMTJ sits one level above existing SMTJ-based true random number generators and Monte Carlo accelerators. In a conventional TRNG pipeline, SMTJs are wired as entropy sources that produce (approximately) independent random bits which are then consumed by an external software or digital model. Here, by contrast, the same thermal telegraph noise is interpreted as a physically grounded entropic gate: SMTJ p-bits implement a softmax over semantic modes, with energies E i ( k ) directly tied to predicted entropy drops Δ H i ( k ) in the output distribution. Rather than driving an unrelated Monte Carlo sampler, SMTJ randomness modulates the internal trajectory Σ ( τ ) itself, shaping when and how the system commits to an attractor (e.g. a Peano-consistent mapping 2 + 2 4 ) or, in near-degenerate regimes, occasionally falls into an anomalous attractor (e.g. 2 + 2 5 ). In this sense, ESDM–SMTJ can be viewed as a semantic and cognitive layer on top of the same SMTJ-based TRNG and stochastic-computing primitives developed in Refs. [1,2,3,4].
The remainder of the paper is structured as follows. Section 2 formalizes the semantic state space, entropic gate and SMTJ coupling. Section 3 presents toy examples, including arithmetic judgements such as 2 + 2 = ? 4 and controlled anomalies. Section 4 discusses implementation aspects on SMTJ hardware, and Section 5 sketches a quantum extension where semantic modes become basis states of a Hamiltonian. Section 6.1 situates ESDM–SMTJ with respect to prior SMTJ-based p-bit Boltzmann machines and TRNG architectures. We conclude in Section 6 with open questions and possible applications to BCI/EEG-based authentication.

2. Formal Model

In this section we formalize the core components of the ESDM–SMTJ framework. We start from an abstract semantic state space and a notion of expression-induced dynamics, then introduce the entropic gate, its coupling to SMTJ p-bit blocks and the definitions of semantic commit and blind reveal.

2.1. Semantic State Space

Let T denote a finite vocabulary of tokens (e.g. digits, operators such as “+” and “=”, and special query symbols such as “?”). We write expressions as finite sequences
e = ( t 1 , , t n ) T * .
The internal semantic state of the system lives in a compact metric space S , which we will refer to as the semantic state space. We write Σ S for a generic state and
τ [ 0 , 1 ]
for an internal computational time parameter that runs from the onset of processing an expression ( τ = 0 ) to the deployment of a final response ( τ = 1 ).
To reflect the existence of qualitatively different regimes (e.g. low-level sensory encodings, higher-level symbolic representations), we assume that S is layered:
S = m = 0 M S m ,
where each S m is a submanifold or subspace of states associated with a “layer” or “plane” of representation. In many applications one may take S m R d m and view the global state as Σ = ( x , m ) with x S m , but the construction does not depend on a specific parametrization.
In later sections we will sometimes distinguish between a slow fingerprint component B (e.g. subject identity) and a fast content component C ( τ ) , writing Σ ( τ ) = ( B , C ( τ ) , m ( τ ) ) , but for the formal development it suffices to treat Σ ( τ ) as a single state variable.

2.2. Output Distribution and Entropy

We assume a finite set Y of possible output tokens (e.g. digits, “yes/no”, special symbols). An observation family is a collection of maps
E x : S [ 0 , 1 ] , x Y ,
such that for every Σ S ,
x Y E x ( Σ ) = 1 .
We interpret E x ( Σ ) as the acceptability or probability assigned to output x when the internal state is Σ . For a fixed expression e the induced output distribution at time τ is
p e ( x Σ ( τ ) ) : = E x ( Σ ( τ ) ) , x Y .
Given p e ( · Σ ( τ ) ) we define the semantic output entropy at time τ as the Shannon entropy
H e ( τ ) : = x Y p e ( x Σ ( τ ) ) log p e ( x Σ ( τ ) ) .
Low values of H e ( τ ) correspond to near-deterministic judgements (high confidence), while high values correspond to states of indecision or strong ambiguity.

2.3. Expression-Induced Dynamics

Each expression e = ( t 1 , , t n ) induces a program-specific dynamics on the semantic state space. Formally, we assume the existence of a family of maps
U e τ : S S , τ [ 0 , 1 ] ,
such that
U e 0 = id S , U e 1 = U e ,
and τ U e τ ( Σ ) is continuous for each Σ S . Given an initial state Σ 0 S , the trajectory under e is
Σ ( τ ) : = U e τ ( Σ 0 ) , τ [ 0 , 1 ] .
In many settings it is convenient to work with a discrete time index x = 0 , 1 , , n aligned with the token sequence, and to set
τ x : = x n , Σ x : = Σ ( τ x ) .
We then regard the local update induced by token t x as the transition
Σ x = F t x ( Σ x 1 ) ,
for some family of token-specific maps F t : S S . The continuous-time U e τ can be obtained, for example, by interpolating between these discrete updates.

2.4. Ambiguous Operators and Semantic Modes

We distinguish a subset O T of ambiguous operators, such as “+”, whose meaning is not fixed but can be instantiated by one of several semantic modes depending on context. For each o O we define a finite set of modes
M o = { 1 , , M o }
(e.g. standard arithmetic addition, a rare erroneous mode, a metaphorical or domain-specific reinterpretation).
Let k be a position in the expression e such that t k = o O . We denote by τ k (or equivalently by the discrete index k) the internal time where the system processes this occurrence of o, and write
Σ k : = Σ ( τ k )
for the state just before applying the operator at position k.
For each mode i M o we assume the existence of a mode-conditioned continuation of the dynamics: starting from Σ k and forcing mode i at position k, we obtain a hypothetical trajectory
Σ ( i ) ( τ ) , τ [ τ k , 1 ] ,
and in particular a terminal entropy
H e , fin ( i , k ) : = H e ( τ = 1 ; mode i chosen at position k ) .
We do not require that this continuation be implemented explicitly in the physical hardware; it can be understood as an effective predictive model used by the entropic gate.

2.5. Entropic Gate and Expected Entropy Drop

The entropic gate at position k evaluates each mode i M o by its predicted reduction of output entropy. Let
H e , k : = H e ( τ k )
be the current output entropy at the moment the operator is about to be applied. We define the expected entropy drop for mode i at position k as
Δ H i ( k ) : = H e , k H e , fin ( i , k ) .
Modes that are expected to lead to clearer, more decisive judgements (smaller final entropy) have larger Δ H i ( k ) .
The entropic gate then defines a probability distribution over modes via a softmax:
G i ( k ) ( Σ k , e ) : = exp β Δ H i ( k ) j M o exp β Δ H j ( k ) , i M o ,
where β > 0 is an inverse-temperature parameter controlling how strongly the gate favours high Δ H . The choice of semantic mode at position k is then modelled as a random variable I k with
P ( I k = i Σ k , e ) = G i ( k ) ( Σ k , e ) .
Conditioned on a realized mode i * at position k, the state update becomes
Σ k + = F o , i * ( Σ k ) , Σ k + 1 = F t k + 1 ( Σ k + ) ,
where F o , i denotes the token update map specialized to mode i of operator o.

2.6. SMTJ Implementation of the Gate

The ESDM–SMTJ model assumes that the softmax distribution (10) is realized at the hardware level by a local block of superparamagnetic tunnel junctions.
For each ambiguous occurrence ( o , k ) we associate a set of effective energy levels
E i ( k ) : = E 0 κ Δ H i ( k ) , i M o ,
for some constants E 0 R and κ > 0 . These energy levels parametrise a classical Boltzmann distribution over modes,
P i ( k ) exp E i ( k ) k B T ,
where k B is Boltzmann’s constant and T is the effective temperature of the SMTJ block.
Substituting the definition of E i ( k ) yields
P i ( k ) exp κ k B T Δ H i ( k ) .
Up to normalization, this coincides with the entropic gate distribution (10) with β = κ / ( k B T ) . Thus the SMTJ block physically implements the entropic gate by sampling from the Boltzmann distribution defined by mode-dependent energies E i ( k ) .
We denote by I k the random mode index generated by this hardware sampling process and identify
P ( I k = i Σ k , e ) = G i ( k ) ( Σ k , e ) .

2.7. Semantic Commit and Blind Reveal

The entropy curve τ H e ( τ ) induced by the dynamics U e τ and the entropic gate often exhibits distinct phases. We are particularly interested in intervals where H e ( τ ) remains high and relatively flat (rumination), followed by sharp drops (insight-like transitions) and stabilization near a low-entropy value.
Definition 1 
(Semantic commit). Let H commit > 0 and δ > 0 be chosen thresholds. We say that a trajectory Σ ( τ ) undergoessemantic commitfor expression e at internal time τ c ( 0 , 1 ) if
  • H e ( τ c ) H commit , and
  • | H e ( τ ) H e ( τ c ) | δ for all τ [ τ c , 1 ] .
In this regime the output distribution is nearly time-invariant and highly concentrated on a subset of Y .
Semantic commit captures the idea that, internally, the system has effectively “made up its mind” before producing any observable output.
The final observable response is produced at some τ f 1 by reading out a separate SMTJ block dedicated to output tokens.
Let H denote the space of microstates of this output block (e.g. binary magnetization configurations of a small array of junctions) and let
R : H Y
be a fixed decoding map from microstates to tokens (e.g. by nearest-codeword in Hamming distance to a set of reference patterns). At τ f the SMTJ output block is sampled according to a distribution concentrated near a configuration consistent with the committed semantic state, and the observed token is
Y out : = R ( H out ) ,
where H out H is the sampled microstate.
Definition 2 
(Blind reveal). We call the map
Σ ( · ) Y out
theblind revealof the expression e under the ESDM–SMTJ dynamics. Internally, semantic commit may occur at τ c < 1 , binding the response; externally, an observer only gains access to a single token Y out revealed at τ f through SMTJ readout and decoding.
In this sense, the ESDM–SMTJ model naturally separates the hidden formation of a judgement (semantic commit) from its eventual exposure via a discrete observable output (blind reveal). Subsequent sections instantiate these abstractions in toy arithmetical examples and discuss how mis-tuned parameters can lead to rare anomalous judgements such as 2 + 2 5 .

3. Toy Examples and Case Studies

In this section we instantiate the abstract ESDM–SMTJ model on a minimal arithmetic judgement. We use the expression e = 2 + 2 = ? as a running example and contrast a “normal” regime, where a Peano-consistent attractor 2 + 2 4 dominates, with an “anomalous” regime where mis-tuned parameters lead to rare 2 + 2 5 events.

3.1. Setup: Expression and Output Space

We take the token sequence
e = ( t 1 , , t 5 ) = ( 2 , + , 2 , = , ? ) ,
and restrict attention to two candidate outputs,
Y = { 4 , 5 } .
The remaining formalism applies equally to larger output alphabets.
Given a trajectory Σ ( τ ) in the semantic state space S , the observation family { E x } x Y induces a time-dependent output distribution
p e ( x Σ ( τ ) ) = E x ( Σ ( τ ) ) , x { 4 , 5 } ,
and a semantic output entropy
H e ( τ ) = x { 4 , 5 } p e ( x Σ ( τ ) ) log p e ( x Σ ( τ ) ) .
We discretize the internal time as
τ x : = x 5 , Σ x : = Σ ( τ x ) , x = 0 , 1 , , 5 .
At x = 0 the system has not yet processed the expression; at x = 5 it is ready to produce an answer.

3.2. Entropic Gate at the “+” Position

The operator “+” at position k = 2 is treated as ambiguous, with two semantic modes:
M + = { N , R } ,
where N denotes the standard arithmetic mode and R denotes a rare, context-dependent or erroneous mode.
Let Σ 2 : = Σ ( τ 2 ) be the state just before processing the “+” at position k = 2 , and let
H e , 2 : = H e ( τ 2 )
be the corresponding output entropy. For each mode i { N , R } we consider the hypothetical continuation of the dynamics where mode i is forced at position 2, resulting in a terminal entropy H e , fin ( i , 2 ) at τ = 1 . The entropic gate computes
Δ H i ( 2 ) : = H e , 2 H e , fin ( i , 2 ) , i { N , R } .
These values are converted into mode probabilities via the softmax
G i ( 2 ) ( Σ 2 , e ) = exp β Δ H i ( 2 ) exp β Δ H N ( 2 ) + exp β Δ H R ( 2 ) , i { N , R } ,
and implemented physically in an SMTJ block with energies
E i ( 2 ) = E 0 κ Δ H i ( 2 ) , i { N , R } .

3.3. Normal Regime: SMTJ–Peano Attractor

We first analyse a regime where the parameters and predictive model are tuned so that the standard arithmetic mode N is strongly favoured. For illustration, suppose that at τ 2 the output distribution over { 4 , 5 } is nearly uniform:
p e ( 4 Σ 2 ) 0.5 , p e ( 5 Σ 2 ) 0.5 ,
so that
H e , 2 log 2 .
If the system anticipates that choosing mode N will lead to a highly concentrated distribution at the end,
p e ( 4 Σ ( N ) ( 1 ) ) 0.95 , p e ( 5 Σ ( N ) ( 1 ) ) 0.05 ,
then
H e , fin ( N , 2 ) [ 0.95 log 0.95 + 0.05 log 0.05 ] log 2 ,
and the expected entropy drop for the normal mode is large,
Δ H N ( 2 ) log 2 H e , fin ( N , 2 ) .
In contrast, under the rare mode R the system may predict a much less decisive final state,
p e ( 4 Σ ( R ) ( 1 ) ) 0.6 , p e ( 5 Σ ( R ) ( 1 ) ) 0.4 ,
giving
H e , fin ( R , 2 ) [ 0.6 log 0.6 + 0.4 log 0.4 ] log 2 ,
and a small entropy drop
Δ H R ( 2 ) H e , 2 H e , fin ( R , 2 ) .
With β = κ / ( k B T ) sufficiently large, the softmax (16) then satisfies
G N ( 2 ) ( Σ 2 , e ) 1 , G R ( 2 ) ( Σ 2 , e ) 0 ,
so that the SMTJ block overwhelmingly selects mode N (standard addition). Conditioned on this choice, the subsequent trajectory Σ ( τ ) is driven towards a low-entropy attractor where
p e ( 4 Σ ( τ ) ) 1 , p e ( 5 Σ ( τ ) ) 0 ,
and the semantic output entropy H e ( τ ) drops sharply after τ 2 before stabilizing near zero.
We say that this attractor realizes an SMTJ–Peano regime: the combination of SMTJ energy landscape and semantic dynamics enforces Peano-consistent behaviour 2 + 2 4 with high probability. Semantic commit occurs at some τ c ( τ 2 , 1 ) , when H e ( τ ) falls below the commit threshold and remains low thereafter. At τ f 1 , the output SMTJ block is read out and decoded to produce a blind reveal Y out = 4 .

3.4. Anomalous Regime: Rare 2 + 2 5 Events

We next consider how the same framework can generate rare but systematic anomalies in arithmetic judgements, such as 2 + 2 5 , under mis-tuned parameters or atypical internal states.
One simple route is through a distorted prediction of entropy drops. Suppose that due to noise, mis-estimation of future dynamics or an atypical context, the system assigns similar or even higher expected entropy reduction to the rare mode:
Δ H N ( 2 ) Δ H R ( 2 ) or Δ H R ( 2 ) > Δ H N ( 2 ) .
In the first case, the entropic gate becomes nearly indifferent:
G N ( 2 ) ( Σ 2 , e ) G R ( 2 ) ( Σ 2 , e ) 1 2 ,
while in the second case it may even favour the rare mode R. At the hardware level this corresponds to an SMTJ energy landscape where the basin associated with mode R is not sufficiently penalized and can be sampled with non-negligible probability.
Conditioned on I 2 = R , the mode-conditioned continuation Σ ( R ) ( τ ) is now realized in the actual dynamics. The subsequent updates F t 3 , F t 4 , F t 5 push the state into a different basin of attraction where the output distribution at late times may satisfy, for instance,
p e ( 4 Σ ( R ) ( τ ) ) 0.2 , p e ( 5 Σ ( R ) ( τ ) ) 0.8 ,
and the entropy H e ( τ ) again drops and stabilizes, but around a different mode. Semantic commit still occurs at some τ c ( R ) ( τ 2 , 1 ) , yet the committed response is now biased towards 5 rather than 4. At τ f the output SMTJ block, having equilibrated near a microstate decoded as 5, produces a blind reveal Y out = 5 .
From the point of view of the formal model, this behaviour is not a violation of the underlying axioms but an excursion into a different attractor of the coupled semantic–SMTJ system. The same entropic gate that normally stabilizes arithmetic truth can, under mis-tuned Δ H i ( k ) or altered energy parameters ( E 0 , κ ) , stabilize a qualitatively different mapping between input expressions and output tokens.

3.5. C++ CLI Demo and AR(1) Thermal Noise Model

To make the entropic gate more concrete, we implemented a small C++ command–line demo that mirrors the structure of the ESDM–SMTJ model. The program performs three steps:
  • It builds an EEG–based fingerprint B ref from several synthetic windows using the imaginary part of coherency (as in the original BCI prototype).
  • It authenticates a new window by computing the distance r to B ref and mapping this to an entropic AUTH_GATE, producing energies E ACCEPT and E REJECT and corresponding softmax probabilities.
  • It runs a PLUS_GATE for the toy expression 2 + 2 = ? , with two modes for “+” (normal Peano mode N and rare anomalous mode R), computes the entropy drops Δ H N , Δ H R , the induced energies E N , E R and the corresponding softmax probabilities P N , P R , and finally samples a mode and emits the token 4 or 5 as a blind reveal.
In the numerical example used throughout this paper, the CLI yields
H before 0.6931 , H fin ( N ) 0.1985 , H fin ( R ) 0.3251 ,
so that
Δ H N 0.4946 , Δ H R 0.3681 ,
and, with a softmax over entropy drops,
P N 0.53 , P R 0.47 .
In the particular run shown below, the gate samples the rare mode R and produces the anomalous judgement 2 + 2 5 :
=== PLUS_GATE for expression ’2+2=?’ ===
H_before  = 0.693147
H_after_N = 0.198515
H_after_R = 0.325083
DeltaH_N      = 0.494632
DeltaH_R      = 0.368064
E_N       = -0.494632
E_R       = -0.368064
P_N (normal 2+2=4) = 0.5316
P_R (rare   2+2->5) = 0.4684
Gate-chosen mode: RARE (2+2->5)
Final distribution p_after = (0.1, 0.9)
Emitted token (blind reveal): 5
For the entropic–margin demo, we additionally model the thermal fluctuations at the PLUS_GATE as a discrete–time Ornstein–Uhlenbeck process, i.e. an AR(1) noise model
ξ m + 1 = ρ ξ m + σ η m ,
where η m N ( 0 , 1 ) , ρ = exp ( Δ t / τ corr ) encodes the correlation time τ corr of the underlying SMTJ magnetization dynamics, and σ sets the effective noise level. In the CLI, we used, for example, τ corr = 0.01 s and Δ t = 2 × 10 4 s, corresponding to “Modo ruido: AR(1) correlacionado, tau_corr = 0.01, dt = 0.0002” in the output. The entropic margin process is then written as
δ ( 2 , m ) = Δ H N Δ H R + ξ m ,
and the thermal–shock index Θ shock is computed directly from the realised trajectory { δ ( 2 , m ) } m = 1 M .

3.6. Interpretation

The toy example 2 + 2 = ? 4 illustrates several aspects of the ESDM–SMTJ model:
  • Internal processing unfolds as a trajectory Σ ( τ ) with a non-trivial entropy profile H e ( τ ) , rather than as an instantaneous jump from premise to conclusion.
  • Ambiguous operators such as “+” are resolved via an entropic gate that predicts future clarity (entropy reduction) and delegates the actual choice of mode to physical sampling in an SMTJ block.
  • Peano-consistent judgements (e.g. 2 + 2 4 ) correspond to attractors in the joint semantic–SMTJ dynamics, whereas rare anomalies (e.g. 2 + 2 5 ) arise when the entropic gate operates in a nearindifferent regime with small entropic margin and elevated thermal-shock index Θ shock , so that correlated SMTJ noise can drive the softmax to sample the anomalous mode even though the normal mode has a slight average advantage.

4. Hardware Considerations for SMTJ Implementations

In this section we outline how the abstract ESDM–SMTJ model could be mapped onto realistic superparamagnetic tunnel junction (SMTJ) hardware and how brain–computer interface (BCI) / EEG signals can be incorporated into the semantic state and the entropic gate.

4.1. SMTJ p-Bit Blocks and Mode Selection

We assume that each ambiguous operator occurrence ( o , k ) is implemented by a local SMTJ-based “mode selection block” responsible for sampling the mode index I k M o .
At a high level, an SMTJ p-bit can be regarded as a noisy binary element whose magnetization stochastically switches between two states (e.g. parallel and anti-parallel) under thermal fluctuations, with a probability that depends on an effective energy barrier and external bias (current or voltage). In the simplest case of two modes, a single p-bit with an appropriate readout and bias suffices to realize the softmax (16); for multiple modes | M o | > 2 , a small network of coupled SMTJs can be configured such that each attractor of the network corresponds to one mode.
Concretely, for each occurrence ( o , k ) and mode i M o we associate an effective energy
E i ( k ) = E 0 κ Δ H i ( k ) ,
computed in the semantic layer as described in Section 2. This value is then translated into device-level control parameters (e.g. bias currents) for the SMTJ block, shaping the stationary distribution over its microstates so that the induced distribution over mode indices i matches the entropic gate probabilities G i ( k ) . A small digital or mixed-signal controller can mediate between the semantic processor (which computes Δ H i ( k ) ) and the analog SMTJ array (which implements the Boltzmann sampling).

4.2. Time Scales and Coupling to Σ ( τ )

The ESDM–SMTJ model separates two time scales:
  • A semantic time scale, parameterized by τ [ 0 , 1 ] or its discrete counterpart x = 0 , , n , on which the state Σ ( τ ) evolves under U e τ as tokens are processed and entropic gates are evaluated.
  • A device time scale, on which SMTJs switch between microstates and the local p-bit networks relax to their stationary distributions given fixed energy parameters E i ( k ) .
In a practical implementation one can exploit the fact that SMTJ switching and relaxation typically occur much faster than high-level semantic updates. For each entropic gate evaluation at position k, the semantic processor computes the set { Δ H i ( k ) } i M o , programs the corresponding energies E i ( k ) into the SMTJ block, and waits a short dwell time for the p-bits to sample from the resulting Boltzmann distribution. The sampled mode index I k is then fed back as a discrete event that updates Σ ( τ ) via the mode-specific map F o , I k .
This design preserves the flexibility of the semantic layer while leveraging the SMTJ hardware as a physical sampler that implements the entropic gate with potentially high energy-efficiency and inherent stochasticity.

4.3. Output SMTJ Block and Code-Based Decoding

The blind reveal mechanism described in Section 2 can be realized by a dedicated SMTJ output block whose attractors are associated with codewords for each output token x Y . Let H out denote the configuration space of this block (e.g. binary patterns of a small SMTJ array), and let
C = { c x H out : x Y }
be a set of reference codewords, one per token. The decoding map R : H out Y can then be defined by nearest-codeword in Hamming or Euclidean distance:
R ( h ) = arg min x Y d ( h , c x ) .
During the semantic commit phase, the semantic processor gradually biases the output SMTJ block (through control currents or equivalent parameters) so that its energy landscape favours the basin associated with the codeword c x * matching the internally committed token x * . At τ f the block is sampled, producing a microstate H out that, with high probability, lies close to c x * ; applying R yields the blind reveal Y out = x * .

4.4. Embedding BCI/EEG Features into the Semantic State

One of the motivations for the ESDM–SMTJ framework is to capture both a relatively stable “fingerprint” component and a fast content component of neural activity in a unified formalism. We briefly outline how BCI/EEG signals can be embedded into the semantic state and influence the entropic gate.
Let z R d denote a feature vector extracted from multichannel EEG during a given trial. The components of z may include, for example: band-limited power in μ / β ranges, phase-locking values (PLV) or imaginary coherence (ImCoh) between electrode pairs, and common spatial pattern (CSP) projections. We assume a learned mapping
f BCI : R d S ,
which embeds EEG features into the semantic state space. Given an incoming EEG sample, we set
Σ 0 = f BCI ( z ) ,
thereby defining the initial condition for the trajectory Σ ( τ ) when processing a symbolic expression e.
In a more structured variant, S can be factored as
S B × C × { 0 , , M } ,
where B represents a slow-varying identity subspace (subject fingerprint), C a fast content subspace, and { 0 , , M } the layer index. The mapping f BCI can then be decomposed as
f BCI ( z ) = ( B ( z ) , C 0 ( z ) , m 0 ) ,
with B ( z ) B encoding subject-specific connectivity or spectral patterns and C 0 ( z ) C encoding task-related content at the onset of processing.

4.5. BCI-Modulated Entropic Gate and Authentication

EEG-derived information can also modulate the entropic gate itself. For instance, in an authentication scenario one may wish to favour trajectories consistent with a stored identity fingerprint B ref and penalize those that deviate from it.
Let B ( z ) be the identity component inferred from the current EEG features and define a dissimilarity measure d B ( B ( z ) , B ref ) in the identity subspace. This quantity can be incorporated into the entropic gate either by adjusting the effective inverse temperature β ,
β eff = β 0 g d B ( B ( z ) , B ref ) ,
with g a decreasing function (e.g. g ( r ) = e λ r ), or by explicitly penalizing modes whose predicted trajectories are incompatible with the reference identity.
Intuitively, when the EEG fingerprint closely matches the stored identity ( d B small), the effective β eff is high, making the entropic gate more decisive and stabilizing the usual attractors (e.g. MSTJ–Peano for 2 + 2 4 ). When the fingerprint deviates strongly from the reference, β eff is reduced, the gate becomes more diffuse or biased towards a special “reject” mode, and the system is less likely to produce confident judgements in the authorized attractors.
From a BCI perspective, this realizes a form of EEG-based authentication within the ESDM–SMTJ framework: the identity component B extracted from EEG not only defines the initial state Σ 0 , but also shapes the entropic gate and thus the accessibility of certain semantic attractors. In particular, one can design SMTJ energy landscapes and decoding schemes such that successful authentication corresponds to the system reliably entering and revealing a specific attractor, while impostor fingerprints produce either high-entropy, unstable trajectories or blind reveals mapped to an explicit “reject” token.

5. Quantum Extension (Q-ESDM)

The ESDM–SMTJ model described so far is fully classical: semantic modes are weighted by real probabilities, and SMTJ p-bit hardware implements a Boltzmann softmax over these modes. In this section we sketch a quantum extension, Q-ESDM, in which semantic modes are represented as basis states of a Hilbert space and the entropic gate is implemented by a Hamiltonian whose spectrum is tied to the expected entropy drop.
Our goal is not to propose a specific physical platform, but to outline how the key ideas of multi-modal semantics, entropic gate and commit/reveal can be lifted to a quantum setting.

5.1. Hilbert Space of Semantic Modes

For each ambiguous operator o O with mode set M o = { 1 , , M o } we introduce a finite-dimensional Hilbert space
H o = span { | i : i M o } ,
with orthonormal basis vectors | i labelled by semantic modes. In the classical ESDM–SMTJ model, the state of o at position k is described by a probability vector { G i ( k ) } i M o . In Q-ESDM, the corresponding quantum state is a vector
| ψ o ( k ) = i M o α i ( k ) | i , i | α i ( k ) | 2 = 1 ,
with complex amplitudes α i ( k ) . The squared moduli | α i ( k ) | 2 play the role of mode probabilities, but the relative phases carry additional information and enable interference between semantic modes.
The total Hilbert space for an expression e with occurrences o 1 , , o L of ambiguous operators is constructed as a tensor product
H sem = = 1 L H o H res ,
where H res is an output or “result” register. The classical semantic state Σ ( τ ) and EEG-derived component B can be treated as external parameters that control Hamiltonians acting on H sem , or, in a more elaborate variant, be included in an enlarged Hilbert space that also encodes identity and content.

5.2. Hamiltonian Encoding of Entropic Preferences

In the classical model, the entropic gate at position k uses the expected entropy drop Δ H i ( k ) to define mode energies
E i ( k ) = E 0 κ Δ H i ( k ) ,
and realizes a softmax over modes via classical Boltzmann sampling. In Q-ESDM, we use the same quantities to define a local semantic Hamiltonian acting on the mode space at position k:
H sem ( k ) = i M o E i ( k ) | i i | .
Modes that are expected to lead to a greater entropy drop (i.e. clearer judgements) are assigned lower energies.
Starting from an initial superposition, for instance a uniform state
| ψ o ( k ) ( 0 ) = 1 M o i M o | i ,
one can implement an entropic gate in at least two canonical ways:
  • Adiabatic scheme: define a time-dependent Hamiltonian H ( k ) ( s ) that interpolates between a simple initial Hamiltonian with known ground state and H sem ( k ) in (26). Under sufficiently slow evolution, the instantaneous state follows a low-energy eigenstate that concentrates amplitude on modes i with low E i ( k ) , i.e. high Δ H i ( k ) .
  • Variational scheme: construct a parametrized unitary U ( k ) ( θ ) (e.g. in the style of QAOA) acting on H o , with layers that alternate between H sem ( k ) and a mixing Hamiltonian. The parameters θ can be optimized offline or adapted online to align the final measurement statistics in the basis { | i } with the desired distribution over modes.
In both cases, the quantum entropic gate uses the same classical information Δ H i ( k ) to shape a Hamiltonian landscape that favours semantically advantageous modes, but it does so through unitary dynamics and interference rather than through classical thermal fluctuations.

5.3. Quantum Expression Dynamics and Commit

At the level of the entire expression, we replace the classical family of maps U e τ : S S by a family of unitaries
U e ( τ ) : H sem H sem , τ [ 0 , 1 ] ,
with U e ( 0 ) = I and U e ( 1 ) = U e . The global quantum state at internal time τ is
| Ψ ( τ ) = U e ( τ ) | Ψ ( 0 ) ,
where | Ψ ( 0 ) encodes the initial superpositions over semantic modes and, optionally, over identity and content degrees of freedom.
As in the classical case, we are interested in the behaviour of the induced output statistics over Y . Let { M x } x Y be a POVM acting on the result register H res (and possibly on parts of H sem ) that models an ideal readout of output tokens. The time-dependent output distribution is then
p e Q ( x | Ψ ( τ ) ) = Ψ ( τ ) | M x | Ψ ( τ ) , x Y ,
and we can define a quantum analogue of semantic output entropy, either as a classical Shannon entropy of p e Q or as a von Neumann entropy of an appropriate reduced state.
We say that a quantum semantic commit has occurred at τ c when the reduced state of the result register is close to a pure state associated with a particular output: for instance, when its density matrix ρ res ( τ c ) satisfies
ρ res ( τ c ) | x * x * | + ε ,
with ε small in trace norm. At that point, even if the full state | Ψ ( τ c ) remains coherent and entangled with other registers, the statistics of an eventual measurement in the output basis are already nearly fixed.

5.4. Measurement and Blind Reveal in Q-ESDM

The blind reveal mechanism carries over naturally to the quantum extension. At a final time τ f 1 , the result register is measured in the token basis { | x : x Y } , or more generally via the POVM { M x } . The observed token
Y out Q Y
is a random variable whose distribution is sharply peaked at x * if semantic commit has occurred, but which remains underdetermined if the system has failed to converge to a low-entropy regime.
From the external observer’s perspective, the process is operationally similar to the classical SMTJ-based blind reveal: a rich, hidden internal dynamics (now unitary and possibly entangled) culminates in a single discrete outcome that exposes only one branch of an underlying superposition. The distinction lies in the representational power of the quantum state, which can encode coherent superpositions of semantic modes and exploit interference in selecting trajectories.

5.5. Classical ESDM–SMTJ as a Limit

The original ESDM–SMTJ model can be recovered as a restriction or limit of Q-ESDM in which:
  • All density matrices are constrained to be diagonal in the mode basis { | i } , so that only classical probabilities over modes are represented and phase information is discarded.
  • The semantic Hamiltonians H sem ( k ) are used only to define Boltzmann weights exp ( E i ( k ) / k B T ) for classical sampling, rather than to generate unitary evolution.
  • The dynamics of Σ ( τ ) and of the SMTJ arrays are treated as classical stochastic processes, without maintaining coherence across different semantic trajectories.
In this sense, ESDM–SMTJ can be viewed as the “thermal” or decohered limit of a more general quantum semantic dynamics. The Q-ESDM framework suggests a possible route for embedding entropic semantic computation into future quantum hardware, while the classical SMTJ implementation remains more immediately accessible for near-term experiments and applications to BCI/EEG-based authentication.
A detailed analysis of concrete quantum circuits, noise models and potential algorithmic advantages of Q-ESDM is left for future work. Here we simply note that the core concepts of multi-modal semantics, entropic preference and commit/reveal admit a natural reformulation in the language of quantum states and Hamiltonians.

6. Conclusion and Outlook

We have introduced ESDM–SMTJ, an Entropic Semantic Dynamics Model that couples a high-level semantic state trajectory Σ ( τ ) with a low-level SMTJ-based p-bit substrate. Operators such as “+” are treated as genuinely multi-modal, with the effective mode at each occurrence selected by an entropic gate that favours trajectories expected to reduce output entropy. This preference is realized physically by mapping predicted entropy drops Δ H i ( k ) to energy levels E i ( k ) in small SMTJ networks, whose Boltzmann statistics implement a softmax over semantic modes.
Internal processing unfolds as a trajectory Σ ( τ ) with a structured, time-varying entropy profile H e ( τ ) , rather than as an instantaneous jump from premise to conclusion. Peano-consistent judgements (e.g. 2 + 2 4 ) correspond to attractors in the joint semantic–SMTJ dynamics; rare anomalies (e.g. 2 + 2 5 ) arise when the gate operates in a near-indifferent regime with a small entropic margin Δ H N Δ H R and an elevated thermal-shock index Θ shock (as in our C++ demo with Δ H N 0.49 , Δ H R 0.37 and softmax probabilities P N 0.53 , P R 0.47 ), so that correlated SMTJ noise can push the entropic softmax to sample the anomalous mode even though the normal mode retains a slight average advantage.
Conceptually, the framework is novel in three ways. First, it internalizes polysemic operators within the state space itself, rather than treating ambiguity as a purely syntactic or external phenomenon. Second, it gives a concrete dynamical account of rumination, insight and stabilization via the entropy profile H e ( τ ) and the notions of semantic commit and blind reveal. Third, it anchors these ideas in a realistic hardware model (SMTJs) while also admitting a quantum extension (Q-ESDM) in which semantic modes become basis states of a Hamiltonian.
Our main contributions can be summarized as follows:
  • We formalize a semantic state space with internal time, ambiguous operators and an entropic gate that selects semantic modes based on predicted entropy reduction.
  • We show how this gate can be implemented in SMTJ-based p-bit hardware by mapping entropy drops to energy levels, so that Boltzmann statistics realize a softmax over modes.
  • We introduce the notions of semantic commit and blind reveal as a way to separate hidden judgement formation from overt token emission.
  • We provide a toy arithmetic case study ( 2 + 2 = ? 4 ) that illustrates both Peano-consistent attractors and controlled anomalous judgements ( 2 + 2 5 ) arising from mis-tuned parameters.
  • We outline a quantum extension (Q-ESDM) where semantic modes live in a Hilbert space and the entropic preferences are encoded in a semantic Hamiltonian.
Beyond abstract reasoning, the ESDM–SMTJ framework connects naturally to BCI and cognitive modelling. By embedding EEG-derived feature vectors into the semantic state and allowing the identity component B to modulate the entropic gate, the model can represent both a slow fingerprint-like structure and fast, task-related content. This opens a route to EEG-based authentication schemes in which access to specific semantic attractors (e.g. a stable arithmetical mapping or a particular decision pattern) depends on the match between an incoming neural fingerprint and a stored reference. More broadly, the entropy profile H e ( τ ) and the commit/reveal mechanism provide a compact language to describe cognitive phenomena such as hesitation, sudden insight and confidence in a way that is directly tied to hardware dynamics.
Compared to quantum annealing platforms such as D-Wave, the ESDM–SMTJ approach trades tunnelling-based search in large Ising landscapes for a more flexible, room-temperature architecture explicitly designed for semantic dynamics. Rather than competing directly on asymptotic optimization power, it offers a complementary direction: implementing structured, entropy-guided semantic computation and cognitive-like trajectories in compact classical hardware, with a clear theoretical pathway to a future quantum version (Q-ESDM).
Several open problems remain for future work:
  • Expressivity and identifiability. Characterizing which classes of semantic mappings and cognitive processes can be represented in ESDM–SMTJ, and under what conditions the fingerprint and content components are identifiable from behaviour or EEG.
  • Learning rules. Deriving concrete online or offline learning rules for updating the predictive model of Δ H i ( k ) and the embedding f BCI from data.
  • Hardware prototypes. Designing and fabricating small SMTJ arrays that demonstrably implement the entropic gate and the commit/reveal mechanism, and benchmarking them against purely digital implementations.
  • Comparison with quantum annealing. Analysing systematically when classical ESDM–SMTJ, its quantum extension Q-ESDM and existing quantum annealers differ in performance, robustness and interpretability on semantic and cognitive tasks.
  • Cognitive validation. Testing whether the predicted entropy profiles and commit times correlate with behavioural measures of hesitation, confidence and error patterns in human subjects.
We hope that ESDM–SMTJ and Q-ESDM can serve as a bridge between formal logic, probabilistic hardware and neurocognitive data, and as a starting point for a broader study of entropic semantic computation in both classical and quantum devices.

6.1. Relation to Prior SMTJ p-Bit BOLTZMANN Machines

Our work is closely related to prior demonstrations of superparamagnetic tunnel junction (sMTJ) based p-bit hardware, in particular the hardware-aware in situ learning of Kaiser et al. [5]. Those systems and ESDM–SMTJ share a common physical substrate, but differ in their objective and abstraction level. Here we briefly highlight the main points of contact and departure.
  • Common substrate and statistics. Both approaches use sMTJ devices operated in a thermally activated regime as stochastic binary units (p-bits), and exploit their Boltzmann statistics. In Kaiser et al., the p-bits implement the neurons of a fully connected Boltzmann machine, sampling from P ( m ) exp ( β E ( m ) ) with E ( m ) = i j W i j m i m j . In ESDM–SMTJ, the same physics is used at a different granularity: small p-bit blocks realize softmax choices over a finite set of semantic modes, with energies E i ( k ) derived from predicted entropy drops rather than from a fixed Ising cost function.
  • Learning objective versus semantic dynamics. The focus of Kaiser et al. is hardware-aware in situ learning of a Boltzmann machine: device imperfections and parameter spreads are compensated by an analog learning rule so that the hardware reproduces a target distribution (e.g. the truth table of a full adder). By contrast, ESDM–SMTJ does not aim to approximate an external distribution over bit patterns. Instead, it defines a semantic state trajectory Σ ( τ ) , multi-modal operators and a notion of graded acceptability, and uses sMTJ blocks as entropic gates that steer Σ ( τ ) toward low-entropy semantic attractors.
  • Energy landscape versus entropy-drop gate. In prior p-bit Boltzmann machines, the energy landscape is specified by synaptic weights W i j and biases; the hardware samples from the corresponding equilibrium distribution, and learning adjusts W i j . In ESDM–SMTJ, energies are not static parameters of a global Ising model. For each occurrence of an ambiguous operator, we compute mode-dependent entropy drops Δ H i ( k ) predicted from the semantic model and map them to local energies E i ( k ) = E 0 κ Δ H i ( k ) . The sMTJ block thus implements a softmax over future expected clarity rather than over a fixed cost function, enabling context-dependent semantic mode selection.
  • Blind reveal and commit semantics. Prior work evaluates hardware quality by comparing sampled bit distributions to a target distribution (e.g. via KL divergence) and does not distinguish between internal computation and exposure of results. ESDM–SMTJ introduces an explicit separation between a hidden semantic commit at internal time τ c (when H e ( τ ) becomes low and stable) and a later blind reveal implemented by an output sMTJ block decoding to discrete tokens. This commit/reveal view is tailored to modelling symbolic judgements rather than generic sampling.
  • BCI/EEG and cognitive modelling. To our knowledge, existing sMTJ p-bit machines have not been applied to neural or biometric data. ESDM–SMTJ explicitly embeds EEG-derived feature vectors into the semantic state as a slow-varying fingerprint component B and a fast content component C ( τ ) , and uses sMTJ-based gates for identity-dependent authentication and semantic computation. This connects probabilistic spintronic hardware to BCI and cognitive modelling in a way that goes beyond the scope of prior Boltzmann-machine demonstrations.
In summary, ESDM–SMTJ can be viewed as a semantic and neurocognitive layer built on top of the same p-bit physics exploited by earlier sMTJ Boltzmann machines, with entropic mode selection and blind reveal as its central organizing principles rather than pure distribution learning.

References

  1. Shunsuke Fukami. Spintronics probabilistic computer: Realization of proposals by feynman and hinton using a spintronics device. JSAP Reviews 2025, 2025, 250204.
  2. Matthew, W. Daniels, Advait Madhavan, Philippe Talatchian, Alice Mizrahi, and Mark D. Stiles. Energy-efficient stochastic computing with superparamagnetic tunnel junctions. Physical Review Applied 2020, 13, 034016. [Google Scholar]
  3. Baofang Cai, Yihan He, Yue Xin, Zhengping Yuan, Xue Zhang, Zhifeng Zhu, and Gengchiau Liang. Unconventional computing based on magnetic tunnel junction. Applied Physics A 2023, 129, 236.
  4. Leo Alexander Schnitzspan. Superparamagnetic Tunnel Junctions—True Randomness, Electrical Coupling and Neuromorphic Computing. Ph.d. thesis, Johannes Gutenberg-Universität Mainz, Mainz, Germany, 2023.
  5. Jan Kaiser, William A. Borders, Kerem Y. Camsari, Shunsuke Fukami, Hideo Ohno, and Supriyo Datta. Hardware-aware in situ learning based on stochastic magnetic tunnel junctions. Phys. Rev. Applied 2022, 17, 014016.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated