Chapter 5: Micro Motifs
5.1 Scope and Objectives
This chapter analyzes the behavior of an individual Neo at the smallest structural scales. We focus on nodes, edges, Lex, stochasticity, energy transitions, mutation effects, and In-Life learning rules, independent of population-level dynamics.
5.2 Minimal Neo Structures
Before analyzing general Neo behavior or introducing simulation-based and macro-level analytical tools, it is essential to understand the smallest possible Neo configurations in full mathematical detail. These minimal structures allow us to introduce the exact formal treatment used throughout this chapter—state evolution, transition matrices, and stationary distributions—while also revealing the fundamental building blocks of Neo dynamics. Even the simplest Neos already display the core mechanisms of computation, stochasticity, memory, and stability that will later reappear in more complex systems.
The goal of this section is twofold. First, we provide concrete numerical examples that make the Lex update rule, noise-driven stochasticity, and state transitions completely explicit. Second, by deriving transition matrices and stationary distributions for each small case, we establish the analytical vocabulary that will be necessary for larger-scale reasoning. These elementary analyses form the conceptual bridge between the micro-level Neo defined earlier and the simulation- or macro-level treatments that follow in later chapters.
A Neo, being a finite-state stochastic dynamical system, evolves according to a Markov process once the input is fixed. At any time step, the internal state determines the distribution of the next state. For small Neos, this process can be expressed exactly through a transition matrix, which is a table containing all probabilities of moving from any current state to any future state in one time step. Formally, if the Neo has K possible internal configurations, the transition matrix is a K×K matrix
P(i,j)=P(X(t+1)=j∣X(t)=i),
where X(t) denotes the internal state at time t. The transition matrix completely determines the dynamics of the Neo under fixed external input and fixed parameters.
Once we have the transition matrix, we can study the stationary distribution, which is the long-run probability with which the Neo occupies each state. A stationary distribution is any probability vector π satisfying
πP=π,∑iπi=1.
This equation expresses the idea that the distribution does not change over time when the system is already in its long-run regime. If the Markov chain is irreducible (all states eventually communicate) and aperiodic, the stationary distribution is unique and describes the asymptotic behavior of the Neo independent of its initial condition. In simple configurations, the stationary distribution reveals whether the Neo stabilizes on a particular state, oscillates between several states, or continuously explores the entire state space.
With these foundations, the next subsection analyzes a two-node recurrent Neo structure in detail, including a numerical example, the complete derivation of transition probabilities, the full transition matrix, and the stationary distribution. Additional minimal Neo cases (zero-node, single-node, and other two-node topologies) are provided in Appendix A.
5.2.1 Two-Node Recurrent Neo
We now analyze a minimal recurrent Neo consisting of two binary nodes that feed back into each other while both receive the same external input bit. This gives the simplest nontrivial closed micro-dynamics that can already exhibit stability, oscillation, and noise-driven switching, and it will later serve as a canonical building block for larger Neos. The construction here is consistent with the general Neo definition given in Chapter 2 of the main document.
At tick t, the internal state of this two-node Neo is the binary pair
Xt=(Vt(1),Vt(2))∈{0,1}2,
and the external percept is a single bit
Ut∈{0,1}.
Node 1 receives Ut and the current state of Node 2, while Node 2 receives Ut and the current state of Node 1. The resulting graph is:
Node 1's local input vector is
z1(t)=[Ut,Vt(2)],
and Node 2's local input vector is
z2(t)=[Ut,Vt(1)].
Each node follows the same stochastic threshold rule used in the general Neo model. For node i∈{1,2}, with weight vector wi, bias bi, and noise scale αi, we define the deterministic pre-activation
si=wi⊤zi(t)+bi,
and the noisy activation
ai(t)=si+αiηi(t),
where ηi(t) is a Bernoulli noise term. The updated node state is
Vt+1(i)=H(ai(t)),
with H(⋅) the Heaviside step. In the simplified piecewise-probabilistic form we use here, the noise is summarized by a "soft band" around zero: when si is outside this band the node behaves almost deterministically, and when si lies inside it, the node fires with probability one half. Concretely, the firing probabilities are
p1(u,v2)=⎩⎨⎧1,0,21,s1≥α1,s1≤−α1,−α1<s1<α1,
p2(u,v1)=⎩⎨⎧1,0,21,s2≥α2,s2≤−α2,−α2<s2<α2,
where s1 and s2 are evaluated at the corresponding inputs.
5.2.1.1 Dynamics with Concrete Parameters
To make these abstract rules concrete, we now choose explicit parameters and examine how the two-node Neo behaves step by step. Let the weight vectors and biases be
w1=[1.2,−0.8],b1=−0.3,α1=0.4,
w2=[0.6,1.1],b2=−0.2,α2=0.3.
Recall that the inputs are ordered as
z1(t)=[Ut,Vt(2)],z2(t)=[Ut,Vt(1)].
We consider a fixed external input Ut=1, and an initial internal state
Vt(1)=0,Vt(2)=1.
For this tick, suppose the noise samples are η1(t)=1 and η2(t)=0. Node 1 computes
s1=1.2⋅Ut−0.8⋅Vt(2)+b1=1.2⋅1−0.8⋅1−0.3=0.1,
and then
a1(t)=s1+α1η1(t)=0.1+0.4⋅1=0.5.
Since a1(t)≥0, Node 1 switches to the ON state,
Vt+1(1)=1.
Node 2, with the same input but a different feedback term, computes
s2=0.6⋅Ut+1.1⋅Vt(1)+b2=0.6⋅1+1.1⋅0−0.2=0.4,
and
a2(t)=s2+α2η2(t)=0.4+0.3⋅0=0.4.
Again a2(t)≥0, so Node 2 is also ON:
Vt+1(2)=1.
In this example the feedback and shared input quickly drive the system to the joint ON state
Xt+1=(Vt+1(1),Vt+1(2))=(1,1),
with randomness playing a decisive role only for Node 1 through the positive noise spike. As we now show, once the parameters are fixed we can summarize all such updates by a four-state Markov chain and compute its stationary distribution.
5.2.1.2 Computing Transition Probabilities
We now fix the external input permanently to Ut=1 for all t. Under this assumption, the system becomes a homogeneous Markov chain on the four states
(0,0),(0,1),(1,0),(1,1),
which we abbreviate as 00,01,10,11. For any current state (v1,v2) we set
z1=[1,v2],z2=[1,v1],
and compute the deterministic parts
s1(1,v2)=1.2⋅1−0.8v2−0.3,
s2(1,v1)=0.6⋅1+1.1v1−0.2.
For our chosen parameters the resulting values are
s1(1,0)=0.9,s1(1,1)=0.1,
s2(1,0)=0.4,s2(1,1)=1.5.
With α1=0.4 and α2=0.3, we see that
s1(1,0)=0.9≥α1⇒p1(1,0)=1,
s1(1,1)=0.1∈(−α1,α1)=(−0.4,0.4)⇒p1(1,1)=21,
s2(1,0)=0.4≥α2⇒p2(1,0)=1,
s2(1,1)=1.5≥α2⇒p2(1,1)=1.
Thus, for this specific two-node Neo under constant input Ut=1, Node 2 is effectively deterministic and always fires, while Node 1 is deterministic when Vt(2)=0 and probabilistic when Vt(2)=1.
Given p1(u,v2) and p2(u,v1), the joint transition probability factorizes because the local noises are independent:
P(Xt+1=(v1′,v2′)∣Xt=(v1,v2),Ut=1)=P(Vt+1(1)=v1′∣v2,1)⋅P(Vt+1(2)=v2′∣v1,1).
Using the usual Bernoulli splitting, this expands to
P(Xt+1=(1,1)∣v1,v2)=p1(1,v2)⋅p2(1,v1),
P(Xt+1=(1,0)∣v1,v2)=p1(1,v2)⋅(1−p2(1,v1)),
P(Xt+1=(0,1)∣v1,v2)=(1−p1(1,v2))⋅p2(1,v1),
P(Xt+1=(0,0)∣v1,v2)=(1−p1(1,v2))⋅(1−p2(1,v1)).
With the numeric values for p1 and p2, we can now write the full transition matrix.
5.2.1.3 The 4×4 Transition Matrix
We order the states as 00,01,10,11. For each row, we plug the appropriate p1(1,v2) and p2(1,v1) into the expressions above.
From 00 we have v1=0, v2=0, thus p1=p1(1,0)=1 and p2=p2(1,0)=1. The next state is deterministically 11:
P(00→11)=1,P(00→00)=P(00→01)=P(00→10)=0.
From 01 we have v1=0, v2=1, so p1=p1(1,1)=21 and p2=p2(1,0)=1. Hence
P(01→11)=p1p2=21,
P(01→01)=(1−p1)p2=21,
and transitions to 00 and 10 have probability zero.
From 10 we have v1=1, v2=0, hence p1=p1(1,0)=1 and p2=p2(1,1)=1. Again the next state is deterministically 11:
P(10→11)=1,
all other outcomes zero.
From 11 we have v1=1, v2=1, so p1=p1(1,1)=21 and p2=p2(1,1)=1. This is analogous to the 01 case:
P(11→11)=p1p2=21,P(11→01)=(1−p1)p2=21,
with all other transitions zero.
Collecting these results, the transition matrix P in the state order [00,01,10,11] is
P=000001/201/2000011/211/2.
States 00 and 10 are transient: both flow deterministically into 11 and can never be revisited. The long-run behavior is confined to the two-state subsystem {01,11}.
5.2.1.4 Stationary Distribution for Fixed Input
The stationary distribution π for this Markov chain satisfies
π=πP,∑x∈{00,01,10,11}πx=1,
where πx=P(Xt=x) in the long-run equilibrium. Because the first and third columns of P are identically zero, any stationary distribution must have
π00=π10=0.
Hence the stationary mass lives entirely on 01 and 11. Let
π01=x,π11=1−x.
We only need to enforce the stationarity condition on these two states. For state 01 we have
π01=π01P(01→01)+π11P(11→01)=x⋅21+(1−x)⋅21=21.
This immediately gives x=21. The probability of state 11 then follows from normalization:
π11=1−π01=21.
Thus the unique stationary distribution for this two-node Neo under constant input Ut=1 is
π00=0,π01=21,π10=0,π11=21.
In other words, in the long run the Neo spends half of its time in the partially active configuration (0,1) and half of its time fully active in (1,1). The OFF–OFF and mixed (1,0) states are only visited transiently, on the way into this two-state attractor. Even this minimal recurrent Neo therefore exhibits a nontrivial equilibrium structure: sustained stochastic switching between a "semi-on" and a "fully-on" configuration, shaped jointly by feedback and local noise. This equilibrium behavior will later connect directly to macro-level notions such as stationary distributions over larger Neo populations and the emergence of stable computational motifs.
Although this explicit transition-matrix analysis is useful for understanding the micro-dynamics of very small Neos, it becomes rapidly impractical for larger topologies. A Neo with only ten binary nodes already has a state space of size 210=1024, and the full transition matrix has 220 entries, making exact computation infeasible. In fact, determining stationary distributions or attractors in general recurrent Boolean systems is NP-hard, even without stochasticity. When noise, weighted inputs, and structural asymmetries are included—as they necessarily are in realistic Neos—the combinatorial explosion becomes even more severe. For this reason, while the analysis above is valuable for intuition, we cannot rely on brute-force enumeration for larger structures. In later sections, we therefore shift to more scalable analytical tools such as probabilistic Boolean networks, mean-field approximations, and Ising-model-based energy formulations. These methods allow us to characterize stability, entropy flow, and emergent motifs in larger Neos without computing full exponential-scale transition dynamics.
5.2.2 Scalable Analytical Tools for Larger Neos
This section introduces three analytical frameworks—Probabilistic Boolean Networks, Mean-Field Analysis, and Ising-Model Energy Formulations—that allow us to study Neo dynamics beyond the regime of exact enumeration.
Probabilistic Boolean Networks (PBNs)
A Neo with binary nodes and stochastic activations can be naturally viewed as a probabilistic Boolean network [@shmulevich2002probabilistic]. In the PBN framework, each node i updates according to a Boolean function fi that is chosen probabilistically from a finite family. In Neosis, the stochasticity arises not from switching among Boolean functions but from noise inside each node's activation rule. Nevertheless, the effective update rule can be written in the PBN form
P(Vt+1(i)=1∣Vt=x)=Fi(x),
where the function
Fi(x)=P(ai(t)≥0∣zi(t)=x[inputs(i)])
plays the role of a probabilistic Boolean function. For a threshold node with noise amplitude αi, we can express this as
Fi(x)=⎩⎨⎧1,0,21,si(x)≥αi,si(x)≤−αi,−αi<si(x)<αi,
where
si(x)=wi⊤zi(x)+bi.
In larger Neos, this allows us to treat the global dynamics as a Markov chain specified by the collection of local probabilistic rules Fi. Transition probabilities remain exponentially large in principle, but PBN theory provides tools for analyzing long-term behavior through structure-based reductions, such as dependency graphs and influence measures [@shmulevich2002gene]. These allow us to identify stable motifs, absorbing sets, and highly influential nodes without enumerating the entire state space. This will be essential when analyzing medium-sized Neos (5–20 nodes), where exact methods are impossible but local dependency patterns still convey meaningful structure.
Mean-Field Analysis (MFA)
For even larger Neos, local dependency structures are insufficient, and we instead approximate node variables as weakly correlated random variables [@opper2001advanced]. The mean-field approach replaces the exact binary node values by their expectations. Let
mi(t)=E[Vt(i)]∈[0,1]
denote the probability that node i is ON at tick t. Under mean-field assumptions, the expected update satisfies
mi(t+1)=E[Fi(Vt)]≈Fi(m1(t),m2(t),…,mn(t)),
where the input vector to Fi is replaced by the vector of marginal activation probabilities. For a weighted threshold node, this approximation yields
mi(t+1)≈Φ(αiwi⊤m(t)+bi),
where Φ is a smoothed Heaviside-like transfer function induced by the noise. In the Neosis case with uniform Bernoulli noise, Φ reduces to a clipped linear segment:
Φ(x)=⎩⎨⎧1,0,2x+1,x≥1,x≤−1,−1<x<1.
The mean-field map
m(t+1)=M(m(t))
defines a low-dimensional dynamical system on [0,1]n. Fixed points of this map approximate stationary distributions of the original high-dimensional Neo. Stability of these fixed points provides insight into whether the Neo sustains persistent activation, converges to quiescence, or supports multiple attractors. While mean-field approximations ignore correlations between nodes, they provide tractable approximations for networks with tens or even hundreds of nodes, especially when connections are dense or exhibit weak pairwise correlations.
Ising-Model Energy Formulations
When a Neo's weighted interactions are mostly symmetric—i.e., wij≈wji—the dynamics resemble those of an Ising system [@hopfield1982neural; @amit1989modeling]. By mapping node states to spins via
σi=2Vt(i)−1∈{−1,+1},
we can define an effective energy function
E(σ)=−∑i<jJijσiσj−∑ihiσi,
where the couplings Jij approximate the symmetric part of the weight matrix and the fields hi reflect biases and input effects. In the presence of stochasticity at each node, the Neo performs a noisy relaxation on this energy landscape. The probability of a configuration σ under a stationary distribution often takes a Boltzmann-like form
π(σ)∝exp(−TE(σ)),
where the effective temperature T is related to the noise amplitudes αi. This mapping is not exact unless weights are symmetric and noise is small, but even in approximate form it provides a powerful tool for analyzing phase transitions, stable modes, and metastable attractors in large Neos. Energy minima correspond to stable motifs, while the barrier heights determine the switching dynamics induced by noise. In later sections, we will use Ising-style approximations to analyze specific structured Neo topologies such as chains, rings, stars, and locally modular subgraphs.
5.2.3 Canonical Micro-Motifs
Purpose: Identify recurring low-level patterns.
Expectation: Use chains, fan-in, fan-out, and loops as computational building blocks for larger Neos.
5.3 Lex and Local Computation
5.3.1 Lex Dynamics
Purpose: Formalize deterministic and stochastic transitions induced by the Lex rule.
Expectation: Analyze the influence of weights, bias, and the stochastic term on node updates.
5.3.2 Effect of Stochasticity
Purpose: Study how randomness modifies micro-scale behavior.
Expectation: Show variability, exploration, and divergence across identical initialized Neos.
5.3.3 Micro-Level Expressive Capacity
Purpose: Assess the representational power of small fixed structures.
Expectation: Describe the deterministic and stochastic input–output mappings achievable by one- and two-node Neos.
5.4 Energy Trajectories at Micro Scale
5.4.1 Tick-Level Energy Flow
Purpose: Examine energy changes during a single cycle.
Expectation: Detail computation cost, reward acquisition, and the resulting energy update.
5.4.2 Lifetime and Vitality in Simple Structures
Purpose: Quantify survival properties of minimal Neos.
Expectation: Compare deterministic, stochastic, and recurrent motifs in terms of energy trajectories and survival.
5.5 Micro-Level Mutation Experiments
5.5.1 Isolated Mutation Types
Purpose: Analyze the effect of each mutation primitive separately.
Expectation: Show structural and behavioral results for node+, node−, edge+, edge−, and paramf.
5.5.2 Mutation Cost and Trade-Offs
Purpose: Relate mutation outcomes to energy budget.
Expectation: Demonstrate scenarios where beneficial mutations fail due to cost and scenarios where small modifications outperform structural changes.
5.5.3 Comparative Mutation Strategies
Purpose: Compare alternative mutation strategies on identical initial conditions.
Expectation: Identify strategies that maximize accuracy, stability, or survival at the micro scale.
5.6 In-Life Learning at the Micro Level
5.6.1 In-Life Learning vs Mutation
Purpose: Clarify conceptual separation between In-Life learning rules and evolutionary mutation.
Expectation: Show why In-Life learning must be pattern-triggered rather than error-driven.
5.6.2 Minimal In-Life Learning Schemes
Purpose: Introduce simple local In-Life learning mechanisms.
Expectation: Propose conditional param adjustments and evaluate their behavior in one- and two-node systems.
5.6.3 Effects of In-Life Learning on Micro Dynamics
Purpose: Analyze situations where In-Life learning helps or harms.
Expectation: Present simulations illustrating successful adaptation versus destabilizing drift.
5.7 Role of Stochasticity in Micro Evolution
5.7.1 Fixed-Structure Stochastic Behavior
Purpose: Understand the influence of noise on stable structures.
Expectation: Demonstrate divergence in predictions and internal states across runs.
5.7.2 Stochasticity as Exploration Under Mutation
Purpose: Show how noise facilitates discovery of structural variations.
Expectation: Illustrate how stochasticity interacts with Evo to produce divergent evolutionary paths.
5.8 Summary of Micro-Level Insights
Purpose: Consolidate micro-scale results.
Expectation: Summarize patterns in structural motifs, mutation tendencies, In-Life learning interactions, and the role of stochasticity.
Appendix A: Additional Minimal Neo Structures
This appendix provides detailed analysis of additional minimal Neo configurations that complement the two-node recurrent Neo presented in Section 4.2.1.
A.1 Zero-Node and Degenerate Cases
A zero-node Neo has no internal state. Formally, its state vector is empty:
V(t)=∅.
There is only one possible configuration, which we call state 0. No Lex update is applied, because there are no nodes and therefore no activations to compute. The system can only emit a fixed output or some externally defined constant, and it cannot store any information about past inputs.
Because there is only one state, the internal dynamics form a trivial Markov chain with a single state.
Transition matrix
Let the state space be
S={0}.
The one-step transition matrix is
P=[1],
meaning that if the system is in state 0 at time t, it remains in state 0 at time t+1 with probability 1.
Stationary distribution
The stationary distribution is a row vector
π=[π0],
with normalization
π0=1.
The stationarity condition
πP=π
expands to
[π0][1]=[π0],
which is satisfied for π0=1. Thus the unique stationary distribution is
π=[1].
This confirms that a zero-node Neo has no representational capacity and its internal dynamics are trivial: there is a single state that is always occupied.
A.2 Single-Node Neo
We now consider the smallest non-degenerate Neo: a single internal state node. This Neo can already pass through deterministic and stochastic regimes and store a single bit of memory. The output is always taken directly from the state node:
Y(t)=V(t).
We model Lex for a single node using a linear activation perturbed by symmetric noise. Let the input U(t)∈{0,1} be fixed to a constant value u so that the resulting Markov chain is time-homogeneous.
The activation is
a(t)=aUU(t)+aVV(t)+b+αη(t),
where:
aU is the weight from the external input,
aV is the self-connection weight (memory term),
b is a bias,
α scales the stochastic perturbation,
η(t)∈{−1,+1} is a Rademacher noise variable with
P(η(t)=+1)=P(η(t)=−1)=21.
The state update is
V(t+1)=H(a(t)),
where H(x)=1 if x≥0 and 0 otherwise. The output is simply
Y(t)=V(t).
To obtain a concrete Markov chain and show how the probabilities are derived, we fix the input to U(t)=1 and choose specific parameter values.
Numerical Lex specification
Choose:
aU=1,aV=1,b=−0.7,α=1.0,U(t)=1 for all t.
Then the deterministic part of the activation is
s(V(t))=aUU(t)+aVV(t)+b=1⋅1+1⋅V(t)−0.7=0.3+V(t).
So we have:
If V(t)=0, then
s0=0.3.
If V(t)=1, then
s1=1.3.
The full activation for each case is
a(t)=s(V(t))+αη(t)=s(V(t))+1.0⋅η(t).
We now compute the transition probabilities explicitly.
Computing transition probabilities
We want P(V(t+1)=1∣V(t)=v) for v∈{0,1}.
Because η(t)∈{−1,+1} with equal probability, we can write:
P(V(t+1)=1∣V(t)=v)=P(H(a(t))=1∣V(t)=v)=P(a(t)≥0∣V(t)=v)
=P(s(v)+η(t)≥0)=211{s(v)+1≥0}+211{s(v)−1≥0},
where 1{⋅} is the indicator function.
We apply this to each state.
Case 1: V(t)=0.
Here s0=0.3. The two possible activations are:
a+=s0+1=0.3+1=1.3,a−=s0−1=0.3−1=−0.7.
So:
a+≥0⇒H(a+)=1,
a−<0⇒H(a−)=0.
Thus:
P(V(t+1)=1∣V(t)=0)=21⋅1+21⋅0=21.
Similarly,
P(V(t+1)=0∣V(t)=0)=1−21=21.
Case 2: V(t)=1.
Here s1=1.3. The two possible activations are:
a+=s1+1=1.3+1=2.3,a−=s1−1=1.3−1=0.3.
Both are nonnegative, so:
H(a+)=1,
H(a−)=1.
Therefore:
P(V(t+1)=1∣V(t)=1)=21⋅1+21⋅1=1,
and
P(V(t+1)=0∣V(t)=1)=0.
Transition matrix
Order the states as [0,1]. The transition matrix P(1) has entries
Pij(1)=P(V(t+1)=j∣V(t)=i).
From the probabilities we derived:
From state 0:
P(0→0)=0.5,P(0→1)=0.5.
From state 1:
P(1→0)=0,P(1→1)=1.
Thus
P(1)=[0.500.51].
Stationary distribution
Let the stationary distribution be
π=[π0,π1],
with
π0+π1=1.
The stationarity condition is
πP(1)=π.
Compute the left-hand side:
πP(1)=[π0,π1][0.500.51]=[π0⋅0.5+π1⋅0,π0⋅0.5+π1⋅1]=[0.5π0,0.5π0+π1].
Setting this equal to [π0,π1] yields:
For the first component:
0.5π0=π0⇒0.5π0=0⇒π0=0.
Normalization then forces:
π1=1.
So the unique stationary distribution is
π=(0,1).
In words, under this Lex and fixed input, the single-node Neo eventually spends almost all its time in the state V=1: once it flips to 1 it never returns to 0. The stochasticity only affects the transient path from 0 to 1.
A.3 Two-Node Neo
We now consider the simplest interacting Neo: two binary state nodes
V(t)=(V1(t),V2(t))∈{0,1}2.
We will analyze three basic topologies:
Feedforward: V1→V2.
Parallel: both nodes driven independently by the same input.
Feedback: V1↔V2.
In each case, we specify Lex for both nodes using simple numeric parameters and derive the full transition matrix over the four states, followed by the stationary distribution.
We order the joint states as:
1≡(0,0),2≡(0,1),3≡(1,0),4≡(1,1).
A.3.A Feedforward (V1 → V2)
We first analyze a two-node feedforward motif where V1 depends only on the external input U and noise, and V2 depends only on V1 and noise. This already yields richer dynamics than the single-node case.
Diagram:
We again fix the input to a constant value U(t)=u to obtain a time-homogeneous Markov chain. The randomness then comes solely from internal noise.
Lex specification
Node 1 activation:
a1(t)=1.0⋅U(t)−0.2+0.7⋅η1(t),
with η1(t)∈{−1,+1} and P(η1=±1)=21.
Node 2 activation:
a2(t)=1.4⋅V1(t)−0.1+0.9⋅η2(t),
with η2(t)∈{−1,+1} independent of η1(t).
The state updates are
V1(t+1)=H(a1(t)),V2(t+1)=H(a2(t)).
We will compute the transition matrix for two fixed inputs: u=0 and u=1.
Case u = 0: computing node-wise probabilities
Set U(t)=0.
For node 1, the deterministic part is:
s1=1.0⋅0−0.2=−0.2.
The two activations are:
a1,+=s1+0.7=−0.2+0.7=0.5,a1,−=s1−0.7=−0.2−0.7=−0.9.
So:
H(a1,+)=1,
H(a1,−)=0,
and
P(V1(t+1)=1∣U=0)=21⋅1+21⋅0=0.5.
Thus
P(V1′=1∣U=0)=0.5,P(V1′=0∣U=0)=0.5.
For node 2, the deterministic part depends on V1(t):
s2(V1)=1.4⋅V1−0.1.
If V1(t)=0:
s2(0)=−0.1,
so
a2,+=−0.1+0.9=0.8,a2,−=−0.1−0.9=−1.0.
Hence:
H(a2,+)=1,
H(a2,−)=0,
and
P(V2(t+1)=1∣V1(t)=0)=21.
If V1(t)=1:
s2(1)=1.4−0.1=1.3,
so
a2,+=1.3+0.9=2.2,a2,−=1.3−0.9=0.4.
Both are nonnegative, so
P(V2(t+1)=1∣V1(t)=1)=1.
In summary (for u=0):
P(V1′=1)=0.5,P(V1′=0)=0.5,
P(V2′=1∣V1=0)=0.5,P(V2′=1∣V1=1)=1.
We now build the 4-state transition matrix.
Transition probabilities for each joint state (u = 0)
Let X(t)=(V1(t),V2(t)). Because V1′ depends only on U and V2′ depends only on V1(t) and noise, and noise for node 1 and 2 are independent, we have
P(X(t+1)=(v1′,v2′)∣X(t)=(v1,v2),U=0)=P(V1′=v1′∣U=0)⋅P(V2′=v2′∣V1=v1).
We compute row by row.
Row 1: current state (0,0).
Here V1=0, but note that P(V1′∣U=0) does not depend on the current state. We have:
P(V1′=0)=0.5,P(V1′=1)=0.5.
For V2′:
If V1=0, then P(V2′=1)=0.5, P(V2′=0)=0.5.
Therefore:
P((0,0)→(0,0))=P(V1′=0)⋅P(V2′=0∣V1=0)=0.5⋅0.5=0.25,
P((0,0)→(0,1))=0.5⋅0.5=0.25,
P((0,0)→(1,0))=P(V1′=1)⋅P(V2′=0∣V1=0)=0.5⋅0.5=0.25,
P((0,0)→(1,1))=0.5⋅0.5=0.25.
The row sums to 1.
Row 2: current state (0,1).
The current value of V2 does not influence V1′ or V2′ (which depends on V1, not V2). So the same reasoning applies:
P((0,1)→j)=P((0,0)→j)
for each joint next state j. So the second row is identical to the first:
(0.25,0.25,0.25,0.25).
Row 3: current state (1,0).
Now V1=1. The distribution of V1′ is still independent of the current state and remains
P(V1′=0)=0.5,P(V1′=1)=0.5.
However, V2′ conditional probabilities change because they depend on V1(t), which is 1 here:
P(V2′=1∣V1=1)=1,P(V2′=0∣V1=1)=0.
Thus:
P((1,0)→(0,0))=P(V1′=0)⋅P(V2′=0∣V1=1)=0.5⋅0=0,
P((1,0)→(0,1))=0.5⋅1=0.5,
P((1,0)→(1,0))=P(V1′=1)⋅P(V2′=0∣V1=1)=0.5⋅0=0,
P((1,0)→(1,1))=0.5⋅1=0.5.
So the third row is:
(0,0.5,0,0.5).
Row 4: current state (1,1).
Same as row 3, because V1=1 and V2 is irrelevant to V1′ and V2′:
(0,0.5,0,0.5).
Transition matrix (u = 0)
Collecting all rows:
Pu=0FF=0.250.25000.250.250.50.50.250.25000.250.250.50.5.
Stationary distribution (u = 0)
Let the stationary distribution be
π=(π1,π2,π3,π4),
with
π1+π2+π3+π4=1.
By symmetry of rows (1 and 2 identical; 3 and 4 identical) and columns, it is natural to look for a solution with
π1=π3=a,π2=π4=b,a+b=21.
We now impose stationarity on, say, π1:
π1′=π1=∑i=14πiPu=0FF(i,1)=π1⋅0.25+π2⋅0.25+π3⋅0+π4⋅0.
Substitute π1=a,π2=b:
a=0.25a+0.25b.
Using the constraint a+b=0.5, we solve:
a=0.25(a+b)=0.25⋅0.5=0.125=81.
Therefore
b=0.5−a=0.5−0.125=0.375=83.
Thus
πu=0FF=(81,83,81,83).
Case u = 1
We now set U(t)=1.
For node 1, deterministic part:
s1=1.0⋅1−0.2=0.8.
Then:
a1,+=0.8+0.7=1.5,a1,−=0.8−0.7=0.1.
Both nonnegative, so
P(V1′=1∣U=1)=1.
For node 2, deterministic part:
s2(V1)=1.4V1−0.1.
As before, if V1=0, P(V2′=1)=0.5, and if V1=1, P(V2′=1)=1. However, under U=1, V1′ is always 1, so after one step the system will behave as if V1(t)=1 at all future times.
To keep the Markov chain defined in terms of current state, we again use:
P(X(t+1)=(v1′,v2′)∣X(t)=(v1,v2),U=1)=P(V1′=v1′∣U=1)P(V2′=v2′∣V1=v1).
But now P(V1′=0)=0 and P(V1′=1)=1, so transitions can only go to states with V1′=1 (i.e., states 3 or 4).
Carrying through the same logic as before yields:
Pu=1FF=000000000.50.5000.50.511.
The last state (1,1) is absorbing, and all other states have a nonzero probability of eventually reaching it. Thus the unique stationary distribution is:
πu=1FF=(0,0,0,1).
A.3.B Parallel (U → V1 and U → V2)
Now we consider a parallel topology where both nodes receive the same input but do not directly interact. This is the smallest multi-bit output that can still be treated as two independent single-node Neos.
Diagram:
Assume both nodes obey the same single-node Lex rule used in Section A.2, and that their noise processes are independent. For a fixed input u, each node has the same transition probabilities as the single-node case, and the joint transition matrix is the Kronecker product of the single-node matrix with itself.
For simplicity, consider U(t)=0 with the single-node parameters chosen so that
P(Vi′=1∣U=0)=0.5,i∈{1,2}.
Then each node at each step is an independent Bernoulli(0.5) random variable, independent of its previous state. The joint distribution over the pair is uniform over the four states.
Thus for any current state,
P(X(t+1)=(v1′,v2′))=0.25
for each of the four configurations.
The transition matrix is therefore:
Pu=0PAR=0.250.250.250.250.250.250.250.250.250.250.250.250.250.250.250.25.
The stationary distribution π must satisfy πPu=0PAR=π. Because every row is identical and uniform, the unique normalized solution is
πu=0PAR=(0.25,0.25,0.25,0.25).
The parallel 2-node Neo under symmetric stochastic input is therefore a uniform random generator over the four joint states.
For U(t)=1 and our earlier single-node parameters, each node deterministically goes to state 1. Hence (1,1) is absorbing, and the stationary distribution is again
πu=1PAR=(0,0,0,1).
A.3.C Feedback (V1 ↔ V2)
Finally, we consider a feedback topology where each node depends on the other's previous state. This is the smallest genuine recurrent Neo and can already exhibit nontrivial stochastic dynamics, including concentration on particular states.
Diagram:
We again fix U(t)=u and choose explicit parameters.
Lex specification
Node 1:
a1(t)=0.7⋅U(t)+1.0⋅V2(t)−0.5+0.6⋅η1(t).
Node 2:
a2(t)=−0.4⋅U(t)+1.5⋅V1(t)−0.2+0.7⋅η2(t).
Both noise variables η1,η2 take values {−1,+1} with probability 0.5 and are independent.
Updates:
V1(t+1)=H(a1(t)),V2(t+1)=H(a2(t)).
We again consider two cases: u=0 and u=1.
Case u = 0
Set U(t)=0. The deterministic parts are:
s1(V2)=0.7⋅0+1.0V2−0.5=V2−0.5,
s2(V1)=−0.4⋅0+1.5V1−0.2=1.5V1−0.2.
We compute node-wise probabilities.
Node 1:
If V2=0:
s1(0)=−0.5,
so
a1,+=−0.5+0.6=0.1,a1,−=−0.5−0.6=−1.1.
Thus
P(V1′=1∣V2=0,U=0)=21⋅1+21⋅0=0.5.
If V2=1:
s1(1)=1−0.5=0.5,
so
a1,+=0.5+0.6=1.1,a1,−=0.5−0.6=−0.1.
Again,
P(V1′=1∣V2=1,U=0)=0.5.
So for u=0, node 1 is symmetric:
P(V1′=1∣V2=0)=P(V1′=1∣V2=1)=0.5.
Node 2:
If V1=0:
s2(0)=−0.2,
a2,+=−0.2+0.7=0.5,a2,−=−0.2−0.7=−0.9.
So
P(V2′=1∣V1=0,U=0)=0.5.
If V1=1:
s2(1)=1.5−0.2=1.3,
a2,+=1.3+0.7=2.0,a2,−=1.3−0.7=0.6.
Both nonnegative, so
P(V2′=1∣V1=1,U=0)=1.
We now build the transition matrix.
State (0,0): here V1=0,V2=0.
Node1: P(V1′=1∣V2=0)=0.5, P(V1′=0)=0.5.
Node2: P(V2′=1∣V1=0)=0.5, P(V2′=0)=0.5.
Joint transitions:
P((0,0)→(0,0))=0.5⋅0.5=0.25,
P((0,0)→(0,1))=0.5⋅0.5=0.25,
P((0,0)→(1,0))=0.5⋅0.5=0.25,
P((0,0)→(1,1))=0.5⋅0.5=0.25.
State (0,1): V1=0,V2=1.
Node1: P(V1′=1∣V2=1)=0.5, P(V1′=0)=0.5.
Node2: P(V2′=1∣V1=0)=0.5, P(V2′=0)=0.5.
Thus the row is identical to state (0,0):
(0.25,0.25,0.25,0.25).
State (1,0): V1=1,V2=0.
Node1: P(V1′=1∣V2=0)=0.5.
Node2: P(V2′=1∣V1=1)=1, P(V2′=0)=0.
So:
P((1,0)→(0,0))=P(V1′=0)⋅P(V2′=0∣V1=1)=0.5⋅0=0,
P((1,0)→(0,1))=0.5⋅1=0.5,
P((1,0)→(1,0))=0.5⋅0=0,
P((1,0)→(1,1))=0.5⋅1=0.5.
State (1,1): V1=1,V2=1.
Node1: P(V1′=1∣V2=1)=0.5.
Node2: P(V2′=1∣V1=1)=1.
So again:
(1,1)→(0,0)=0,(1,1)→(0,1)=0.5,(1,1)→(1,0)=0,(1,1)→(1,1)=0.5.
Collecting these rows, we obtain:
Pu=0FB=0.250.25000.250.250.50.50.250.25000.250.250.50.5.
This matrix is identical to the feedforward case with u=0, so the stationary distribution is again:
πu=0FB=(81,83,81,83).
Case u = 1 (feedback)
With U=1, the deterministic parts change, and the Markov chain becomes more biased toward the state (1,1). A similar step-by-step analysis (omitted here to avoid redundancy) yields:
Pu=1FB=0.250000.2500.500.250.5000.250.50.51.
The state (1,1) is absorbing and is reachable from every other state under the dynamics, so the unique stationary distribution is
πu=1FB=(0,0,0,1).
This two-node feedback Neo therefore converges, under this Lex choice, to a fully "activated" state in the long run.
Last updated