Chapter 5: Micro Motifs

5.1 Scope and Objectives

This chapter analyzes the behavior of an individual Neo at the smallest structural scales. We focus on nodes, edges, Lex, stochasticity, energy transitions, mutation effects, and In-Life learning rules, independent of population-level dynamics.

5.2 Minimal Neo Structures

Before analyzing general Neo behavior or introducing simulation-based and macro-level analytical tools, it is essential to understand the smallest possible Neo configurations in full mathematical detail. These minimal structures allow us to introduce the exact formal treatment used throughout this chapter—state evolution, transition matrices, and stationary distributions—while also revealing the fundamental building blocks of Neo dynamics. Even the simplest Neos already display the core mechanisms of computation, stochasticity, memory, and stability that will later reappear in more complex systems.

The goal of this section is twofold. First, we provide concrete numerical examples that make the Lex update rule, noise-driven stochasticity, and state transitions completely explicit. Second, by deriving transition matrices and stationary distributions for each small case, we establish the analytical vocabulary that will be necessary for larger-scale reasoning. These elementary analyses form the conceptual bridge between the micro-level Neo defined earlier and the simulation- or macro-level treatments that follow in later chapters.

A Neo, being a finite-state stochastic dynamical system, evolves according to a Markov process once the input is fixed. At any time step, the internal state determines the distribution of the next state. For small Neos, this process can be expressed exactly through a transition matrix, which is a table containing all probabilities of moving from any current state to any future state in one time step. Formally, if the Neo has KK possible internal configurations, the transition matrix is a K×KK \times K matrix

P(i,j)=P(X(t+1)=jX(t)=i),P(i,j) = P(X(t+1)=j \mid X(t)=i),

where X(t)X(t) denotes the internal state at time tt. The transition matrix completely determines the dynamics of the Neo under fixed external input and fixed parameters.

Once we have the transition matrix, we can study the stationary distribution, which is the long-run probability with which the Neo occupies each state. A stationary distribution is any probability vector π\pi satisfying

πP=π,iπi=1.\pi P = \pi, \quad \sum_i \pi_i = 1.

This equation expresses the idea that the distribution does not change over time when the system is already in its long-run regime. If the Markov chain is irreducible (all states eventually communicate) and aperiodic, the stationary distribution is unique and describes the asymptotic behavior of the Neo independent of its initial condition. In simple configurations, the stationary distribution reveals whether the Neo stabilizes on a particular state, oscillates between several states, or continuously explores the entire state space.

With these foundations, the next subsection analyzes a two-node recurrent Neo structure in detail, including a numerical example, the complete derivation of transition probabilities, the full transition matrix, and the stationary distribution. Additional minimal Neo cases (zero-node, single-node, and other two-node topologies) are provided in Appendix A.

5.2.1 Two-Node Recurrent Neo

We now analyze a minimal recurrent Neo consisting of two binary nodes that feed back into each other while both receive the same external input bit. This gives the simplest nontrivial closed micro-dynamics that can already exhibit stability, oscillation, and noise-driven switching, and it will later serve as a canonical building block for larger Neos. The construction here is consistent with the general Neo definition given in Chapter 2 of the main document.

At tick tt, the internal state of this two-node Neo is the binary pair

Xt=(Vt(1),Vt(2)){0,1}2,X_t = (V_t(1), V_t(2)) \in \{0,1\}^2,

and the external percept is a single bit

Ut{0,1}.U_t \in \{0,1\}.

Node 1 receives UtU_t and the current state of Node 2, while Node 2 receives UtU_t and the current state of Node 1. The resulting graph is:

Node 1's local input vector is

z1(t)=[Ut,Vt(2)],z_1(t) = [U_t, V_t(2)],

and Node 2's local input vector is

z2(t)=[Ut,Vt(1)].z_2(t) = [U_t, V_t(1)].

Each node follows the same stochastic threshold rule used in the general Neo model. For node i{1,2}i \in \{1,2\}, with weight vector wiw_i, bias bib_i, and noise scale αi\alpha_i, we define the deterministic pre-activation

si=wizi(t)+bi,s_i = w_i^\top z_i(t) + b_i,

and the noisy activation

ai(t)=si+αiηi(t),a_i(t) = s_i + \alpha_i \eta_i(t),

where ηi(t)\eta_i(t) is a Bernoulli noise term. The updated node state is

Vt+1(i)=H(ai(t)),V_{t+1}(i) = H(a_i(t)),

with H()H(\cdot) the Heaviside step. In the simplified piecewise-probabilistic form we use here, the noise is summarized by a "soft band" around zero: when sis_i is outside this band the node behaves almost deterministically, and when sis_i lies inside it, the node fires with probability one half. Concretely, the firing probabilities are

p1(u,v2)={1,s1α1,0,s1α1,12,α1<s1<α1,p_1(u, v_2) = \begin{cases} 1, & s_1 \geq \alpha_1, \\ 0, & s_1 \leq -\alpha_1, \\ \frac{1}{2}, & -\alpha_1 < s_1 < \alpha_1, \end{cases}

p2(u,v1)={1,s2α2,0,s2α2,12,α2<s2<α2,p_2(u, v_1) = \begin{cases} 1, & s_2 \geq \alpha_2, \\ 0, & s_2 \leq -\alpha_2, \\ \frac{1}{2}, & -\alpha_2 < s_2 < \alpha_2, \end{cases}

where s1s_1 and s2s_2 are evaluated at the corresponding inputs.

5.2.1.1 Dynamics with Concrete Parameters

To make these abstract rules concrete, we now choose explicit parameters and examine how the two-node Neo behaves step by step. Let the weight vectors and biases be

w1=[1.2,0.8],b1=0.3,α1=0.4,w_1 = [1.2, -0.8], \quad b_1 = -0.3, \quad \alpha_1 = 0.4,

w2=[0.6,1.1],b2=0.2,α2=0.3.w_2 = [0.6, 1.1], \quad b_2 = -0.2, \quad \alpha_2 = 0.3.

Recall that the inputs are ordered as

z1(t)=[Ut,Vt(2)],z2(t)=[Ut,Vt(1)].z_1(t) = [U_t, V_t(2)], \quad z_2(t) = [U_t, V_t(1)].

We consider a fixed external input Ut=1U_t = 1, and an initial internal state

Vt(1)=0,Vt(2)=1.V_t(1) = 0, \quad V_t(2) = 1.

For this tick, suppose the noise samples are η1(t)=1\eta_1(t) = 1 and η2(t)=0\eta_2(t) = 0. Node 1 computes

s1=1.2Ut0.8Vt(2)+b1=1.210.810.3=0.1,s_1 = 1.2 \cdot U_t - 0.8 \cdot V_t(2) + b_1 = 1.2 \cdot 1 - 0.8 \cdot 1 - 0.3 = 0.1,

and then

a1(t)=s1+α1η1(t)=0.1+0.41=0.5.a_1(t) = s_1 + \alpha_1 \eta_1(t) = 0.1 + 0.4 \cdot 1 = 0.5.

Since a1(t)0a_1(t) \geq 0, Node 1 switches to the ON state,

Vt+1(1)=1.V_{t+1}(1) = 1.

Node 2, with the same input but a different feedback term, computes

s2=0.6Ut+1.1Vt(1)+b2=0.61+1.100.2=0.4,s_2 = 0.6 \cdot U_t + 1.1 \cdot V_t(1) + b_2 = 0.6 \cdot 1 + 1.1 \cdot 0 - 0.2 = 0.4,

and

a2(t)=s2+α2η2(t)=0.4+0.30=0.4.a_2(t) = s_2 + \alpha_2 \eta_2(t) = 0.4 + 0.3 \cdot 0 = 0.4.

Again a2(t)0a_2(t) \geq 0, so Node 2 is also ON:

Vt+1(2)=1.V_{t+1}(2) = 1.

In this example the feedback and shared input quickly drive the system to the joint ON state

Xt+1=(Vt+1(1),Vt+1(2))=(1,1),X_{t+1} = (V_{t+1}(1), V_{t+1}(2)) = (1, 1),

with randomness playing a decisive role only for Node 1 through the positive noise spike. As we now show, once the parameters are fixed we can summarize all such updates by a four-state Markov chain and compute its stationary distribution.

5.2.1.2 Computing Transition Probabilities

We now fix the external input permanently to Ut=1U_t = 1 for all tt. Under this assumption, the system becomes a homogeneous Markov chain on the four states

(0,0),(0,1),(1,0),(1,1),(0,0), (0,1), (1,0), (1,1),

which we abbreviate as 00,01,10,1100, 01, 10, 11. For any current state (v1,v2)(v_1, v_2) we set

z1=[1,v2],z2=[1,v1],z_1 = [1, v_2], \quad z_2 = [1, v_1],

and compute the deterministic parts

s1(1,v2)=1.210.8v20.3,s_1(1, v_2) = 1.2 \cdot 1 - 0.8 v_2 - 0.3,

s2(1,v1)=0.61+1.1v10.2.s_2(1, v_1) = 0.6 \cdot 1 + 1.1 v_1 - 0.2.

For our chosen parameters the resulting values are

s1(1,0)=0.9,s1(1,1)=0.1,s_1(1, 0) = 0.9, \quad s_1(1, 1) = 0.1,

s2(1,0)=0.4,s2(1,1)=1.5.s_2(1, 0) = 0.4, \quad s_2(1, 1) = 1.5.

With α1=0.4\alpha_1 = 0.4 and α2=0.3\alpha_2 = 0.3, we see that

s1(1,0)=0.9α1p1(1,0)=1,s_1(1, 0) = 0.9 \geq \alpha_1 \Rightarrow p_1(1, 0) = 1,

s1(1,1)=0.1(α1,α1)=(0.4,0.4)p1(1,1)=12,s_1(1, 1) = 0.1 \in (-\alpha_1, \alpha_1) = (-0.4, 0.4) \Rightarrow p_1(1, 1) = \frac{1}{2},

s2(1,0)=0.4α2p2(1,0)=1,s_2(1, 0) = 0.4 \geq \alpha_2 \Rightarrow p_2(1, 0) = 1,

s2(1,1)=1.5α2p2(1,1)=1.s_2(1, 1) = 1.5 \geq \alpha_2 \Rightarrow p_2(1, 1) = 1.

Thus, for this specific two-node Neo under constant input Ut=1U_t = 1, Node 2 is effectively deterministic and always fires, while Node 1 is deterministic when Vt(2)=0V_t(2) = 0 and probabilistic when Vt(2)=1V_t(2) = 1.

Given p1(u,v2)p_1(u, v_2) and p2(u,v1)p_2(u, v_1), the joint transition probability factorizes because the local noises are independent:

P(Xt+1=(v1,v2)Xt=(v1,v2),Ut=1)=P(Vt+1(1)=v1v2,1)P(Vt+1(2)=v2v1,1).P(X_{t+1} = (v_1', v_2') \mid X_t = (v_1, v_2), U_t = 1) = P(V_{t+1}(1) = v_1' \mid v_2, 1) \cdot P(V_{t+1}(2) = v_2' \mid v_1, 1).

Using the usual Bernoulli splitting, this expands to

P(Xt+1=(1,1)v1,v2)=p1(1,v2)p2(1,v1),P(X_{t+1} = (1, 1) \mid v_1, v_2) = p_1(1, v_2) \cdot p_2(1, v_1),

P(Xt+1=(1,0)v1,v2)=p1(1,v2)(1p2(1,v1)),P(X_{t+1} = (1, 0) \mid v_1, v_2) = p_1(1, v_2) \cdot (1 - p_2(1, v_1)),

P(Xt+1=(0,1)v1,v2)=(1p1(1,v2))p2(1,v1),P(X_{t+1} = (0, 1) \mid v_1, v_2) = (1 - p_1(1, v_2)) \cdot p_2(1, v_1),

P(Xt+1=(0,0)v1,v2)=(1p1(1,v2))(1p2(1,v1)).P(X_{t+1} = (0, 0) \mid v_1, v_2) = (1 - p_1(1, v_2)) \cdot (1 - p_2(1, v_1)).

With the numeric values for p1p_1 and p2p_2, we can now write the full transition matrix.

5.2.1.3 The 4×4 Transition Matrix

We order the states as 00,01,10,1100, 01, 10, 11. For each row, we plug the appropriate p1(1,v2)p_1(1, v_2) and p2(1,v1)p_2(1, v_1) into the expressions above.

From 00 we have v1=0v_1 = 0, v2=0v_2 = 0, thus p1=p1(1,0)=1p_1 = p_1(1, 0) = 1 and p2=p2(1,0)=1p_2 = p_2(1, 0) = 1. The next state is deterministically 1111:

P(0011)=1,P(0000)=P(0001)=P(0010)=0.P(00 \to 11) = 1, \quad P(00 \to 00) = P(00 \to 01) = P(00 \to 10) = 0.

From 01 we have v1=0v_1 = 0, v2=1v_2 = 1, so p1=p1(1,1)=12p_1 = p_1(1, 1) = \frac{1}{2} and p2=p2(1,0)=1p_2 = p_2(1, 0) = 1. Hence

P(0111)=p1p2=12,P(01 \to 11) = p_1 p_2 = \frac{1}{2},

P(0101)=(1p1)p2=12,P(01 \to 01) = (1 - p_1) p_2 = \frac{1}{2},

and transitions to 0000 and 1010 have probability zero.

From 10 we have v1=1v_1 = 1, v2=0v_2 = 0, hence p1=p1(1,0)=1p_1 = p_1(1, 0) = 1 and p2=p2(1,1)=1p_2 = p_2(1, 1) = 1. Again the next state is deterministically 1111:

P(1011)=1,P(10 \to 11) = 1,

all other outcomes zero.

From 11 we have v1=1v_1 = 1, v2=1v_2 = 1, so p1=p1(1,1)=12p_1 = p_1(1, 1) = \frac{1}{2} and p2=p2(1,1)=1p_2 = p_2(1, 1) = 1. This is analogous to the 0101 case:

P(1111)=p1p2=12,P(1101)=(1p1)p2=12,P(11 \to 11) = p_1 p_2 = \frac{1}{2}, \quad P(11 \to 01) = (1 - p_1) p_2 = \frac{1}{2},

with all other transitions zero.

Collecting these results, the transition matrix PP in the state order [00,01,10,11][00, 01, 10, 11] is

P=(000101/201/2000101/201/2).P = \begin{pmatrix} 0 & 0 & 0 & 1 \\ 0 & 1/2 & 0 & 1/2 \\ 0 & 0 & 0 & 1 \\ 0 & 1/2 & 0 & 1/2 \end{pmatrix}.

States 0000 and 1010 are transient: both flow deterministically into 1111 and can never be revisited. The long-run behavior is confined to the two-state subsystem {01,11}\{01, 11\}.

5.2.1.4 Stationary Distribution for Fixed Input

The stationary distribution π\pi for this Markov chain satisfies

π=πP,x{00,01,10,11}πx=1,\pi = \pi P, \quad \sum_{x \in \{00, 01, 10, 11\}} \pi_x = 1,

where πx=P(Xt=x)\pi_x = P(X_t = x) in the long-run equilibrium. Because the first and third columns of PP are identically zero, any stationary distribution must have

π00=π10=0.\pi_{00} = \pi_{10} = 0.

Hence the stationary mass lives entirely on 0101 and 1111. Let

π01=x,π11=1x.\pi_{01} = x, \quad \pi_{11} = 1 - x.

We only need to enforce the stationarity condition on these two states. For state 0101 we have

π01=π01P(0101)+π11P(1101)=x12+(1x)12=12.\pi_{01} = \pi_{01} P(01 \to 01) + \pi_{11} P(11 \to 01) = x \cdot \frac{1}{2} + (1 - x) \cdot \frac{1}{2} = \frac{1}{2}.

This immediately gives x=12x = \frac{1}{2}. The probability of state 1111 then follows from normalization:

π11=1π01=12.\pi_{11} = 1 - \pi_{01} = \frac{1}{2}.

Thus the unique stationary distribution for this two-node Neo under constant input Ut=1U_t = 1 is

π00=0,π01=12,π10=0,π11=12.\boxed{\pi_{00} = 0, \quad \pi_{01} = \frac{1}{2}, \quad \pi_{10} = 0, \quad \pi_{11} = \frac{1}{2}.}

In other words, in the long run the Neo spends half of its time in the partially active configuration (0,1)(0, 1) and half of its time fully active in (1,1)(1, 1). The OFF–OFF and mixed (1,0)(1, 0) states are only visited transiently, on the way into this two-state attractor. Even this minimal recurrent Neo therefore exhibits a nontrivial equilibrium structure: sustained stochastic switching between a "semi-on" and a "fully-on" configuration, shaped jointly by feedback and local noise. This equilibrium behavior will later connect directly to macro-level notions such as stationary distributions over larger Neo populations and the emergence of stable computational motifs.

Although this explicit transition-matrix analysis is useful for understanding the micro-dynamics of very small Neos, it becomes rapidly impractical for larger topologies. A Neo with only ten binary nodes already has a state space of size 210=10242^{10} = 1024, and the full transition matrix has 2202^{20} entries, making exact computation infeasible. In fact, determining stationary distributions or attractors in general recurrent Boolean systems is NP-hard, even without stochasticity. When noise, weighted inputs, and structural asymmetries are included—as they necessarily are in realistic Neos—the combinatorial explosion becomes even more severe. For this reason, while the analysis above is valuable for intuition, we cannot rely on brute-force enumeration for larger structures. In later sections, we therefore shift to more scalable analytical tools such as probabilistic Boolean networks, mean-field approximations, and Ising-model-based energy formulations. These methods allow us to characterize stability, entropy flow, and emergent motifs in larger Neos without computing full exponential-scale transition dynamics.

5.2.2 Scalable Analytical Tools for Larger Neos

This section introduces three analytical frameworks—Probabilistic Boolean Networks, Mean-Field Analysis, and Ising-Model Energy Formulations—that allow us to study Neo dynamics beyond the regime of exact enumeration.

Probabilistic Boolean Networks (PBNs)

A Neo with binary nodes and stochastic activations can be naturally viewed as a probabilistic Boolean network [@shmulevich2002probabilistic]. In the PBN framework, each node ii updates according to a Boolean function fif_i that is chosen probabilistically from a finite family. In Neosis, the stochasticity arises not from switching among Boolean functions but from noise inside each node's activation rule. Nevertheless, the effective update rule can be written in the PBN form

P(Vt+1(i)=1Vt=x)=Fi(x),P(V_{t+1}(i) = 1 \mid V_t = x) = F_i(x),

where the function

Fi(x)=P(ai(t)0zi(t)=x[inputs(i)])F_i(x) = P(a_i(t) \geq 0 \mid z_i(t) = x[\text{inputs}(i)])

plays the role of a probabilistic Boolean function. For a threshold node with noise amplitude αi\alpha_i, we can express this as

Fi(x)={1,si(x)αi,0,si(x)αi,12,αi<si(x)<αi,F_i(x) = \begin{cases} 1, & s_i(x) \geq \alpha_i, \\ 0, & s_i(x) \leq -\alpha_i, \\ \frac{1}{2}, & -\alpha_i < s_i(x) < \alpha_i, \end{cases}

where

si(x)=wizi(x)+bi.s_i(x) = w_i^\top z_i(x) + b_i.

In larger Neos, this allows us to treat the global dynamics as a Markov chain specified by the collection of local probabilistic rules FiF_i. Transition probabilities remain exponentially large in principle, but PBN theory provides tools for analyzing long-term behavior through structure-based reductions, such as dependency graphs and influence measures [@shmulevich2002gene]. These allow us to identify stable motifs, absorbing sets, and highly influential nodes without enumerating the entire state space. This will be essential when analyzing medium-sized Neos (5–20 nodes), where exact methods are impossible but local dependency patterns still convey meaningful structure.

Mean-Field Analysis (MFA)

For even larger Neos, local dependency structures are insufficient, and we instead approximate node variables as weakly correlated random variables [@opper2001advanced]. The mean-field approach replaces the exact binary node values by their expectations. Let

mi(t)=E[Vt(i)][0,1]m_i(t) = \mathbb{E}[V_t(i)] \in [0,1]

denote the probability that node ii is ON at tick tt. Under mean-field assumptions, the expected update satisfies

mi(t+1)=E[Fi(Vt)]Fi(m1(t),m2(t),,mn(t)),m_i(t+1) = \mathbb{E}[F_i(V_t)] \approx F_i(m_1(t), m_2(t), \ldots, m_n(t)),

where the input vector to FiF_i is replaced by the vector of marginal activation probabilities. For a weighted threshold node, this approximation yields

mi(t+1)Φ(wim(t)+biαi),m_i(t+1) \approx \Phi\left(\frac{w_i^\top m(t) + b_i}{\alpha_i}\right),

where Φ\Phi is a smoothed Heaviside-like transfer function induced by the noise. In the Neosis case with uniform Bernoulli noise, Φ\Phi reduces to a clipped linear segment:

Φ(x)={1,x1,0,x1,x+12,1<x<1.\Phi(x) = \begin{cases} 1, & x \geq 1, \\ 0, & x \leq -1, \\ \frac{x+1}{2}, & -1 < x < 1. \end{cases}

The mean-field map

m(t+1)=M(m(t))m(t+1) = M(m(t))

defines a low-dimensional dynamical system on [0,1]n[0,1]^n. Fixed points of this map approximate stationary distributions of the original high-dimensional Neo. Stability of these fixed points provides insight into whether the Neo sustains persistent activation, converges to quiescence, or supports multiple attractors. While mean-field approximations ignore correlations between nodes, they provide tractable approximations for networks with tens or even hundreds of nodes, especially when connections are dense or exhibit weak pairwise correlations.

Ising-Model Energy Formulations

When a Neo's weighted interactions are mostly symmetric—i.e., wijwjiw_{ij} \approx w_{ji}—the dynamics resemble those of an Ising system [@hopfield1982neural; @amit1989modeling]. By mapping node states to spins via

σi=2Vt(i)1{1,+1},\sigma_i = 2V_t(i) - 1 \in \{-1, +1\},

we can define an effective energy function

E(σ)=i<jJijσiσjihiσi,E(\sigma) = -\sum_{i < j} J_{ij} \sigma_i \sigma_j - \sum_i h_i \sigma_i,

where the couplings JijJ_{ij} approximate the symmetric part of the weight matrix and the fields hih_i reflect biases and input effects. In the presence of stochasticity at each node, the Neo performs a noisy relaxation on this energy landscape. The probability of a configuration σ\sigma under a stationary distribution often takes a Boltzmann-like form

π(σ)exp(E(σ)T),\pi(\sigma) \propto \exp\left(-\frac{E(\sigma)}{T}\right),

where the effective temperature TT is related to the noise amplitudes αi\alpha_i. This mapping is not exact unless weights are symmetric and noise is small, but even in approximate form it provides a powerful tool for analyzing phase transitions, stable modes, and metastable attractors in large Neos. Energy minima correspond to stable motifs, while the barrier heights determine the switching dynamics induced by noise. In later sections, we will use Ising-style approximations to analyze specific structured Neo topologies such as chains, rings, stars, and locally modular subgraphs.


5.2.3 Canonical Micro-Motifs

Purpose: Identify recurring low-level patterns.

Expectation: Use chains, fan-in, fan-out, and loops as computational building blocks for larger Neos.

5.3 Lex and Local Computation

5.3.1 Lex Dynamics

Purpose: Formalize deterministic and stochastic transitions induced by the Lex rule.

Expectation: Analyze the influence of weights, bias, and the stochastic term on node updates.

5.3.2 Effect of Stochasticity

Purpose: Study how randomness modifies micro-scale behavior.

Expectation: Show variability, exploration, and divergence across identical initialized Neos.

5.3.3 Micro-Level Expressive Capacity

Purpose: Assess the representational power of small fixed structures.

Expectation: Describe the deterministic and stochastic input–output mappings achievable by one- and two-node Neos.

5.4 Energy Trajectories at Micro Scale

5.4.1 Tick-Level Energy Flow

Purpose: Examine energy changes during a single cycle.

Expectation: Detail computation cost, reward acquisition, and the resulting energy update.

5.4.2 Lifetime and Vitality in Simple Structures

Purpose: Quantify survival properties of minimal Neos.

Expectation: Compare deterministic, stochastic, and recurrent motifs in terms of energy trajectories and survival.

5.5 Micro-Level Mutation Experiments

5.5.1 Isolated Mutation Types

Purpose: Analyze the effect of each mutation primitive separately.

Expectation: Show structural and behavioral results for node+^+, node^-, edge+^+, edge^-, and paramf^f.

5.5.2 Mutation Cost and Trade-Offs

Purpose: Relate mutation outcomes to energy budget.

Expectation: Demonstrate scenarios where beneficial mutations fail due to cost and scenarios where small modifications outperform structural changes.

5.5.3 Comparative Mutation Strategies

Purpose: Compare alternative mutation strategies on identical initial conditions.

Expectation: Identify strategies that maximize accuracy, stability, or survival at the micro scale.

5.6 In-Life Learning at the Micro Level

5.6.1 In-Life Learning vs Mutation

Purpose: Clarify conceptual separation between In-Life learning rules and evolutionary mutation.

Expectation: Show why In-Life learning must be pattern-triggered rather than error-driven.

5.6.2 Minimal In-Life Learning Schemes

Purpose: Introduce simple local In-Life learning mechanisms.

Expectation: Propose conditional param adjustments and evaluate their behavior in one- and two-node systems.

5.6.3 Effects of In-Life Learning on Micro Dynamics

Purpose: Analyze situations where In-Life learning helps or harms.

Expectation: Present simulations illustrating successful adaptation versus destabilizing drift.

5.7 Role of Stochasticity in Micro Evolution

5.7.1 Fixed-Structure Stochastic Behavior

Purpose: Understand the influence of noise on stable structures.

Expectation: Demonstrate divergence in predictions and internal states across runs.

5.7.2 Stochasticity as Exploration Under Mutation

Purpose: Show how noise facilitates discovery of structural variations.

Expectation: Illustrate how stochasticity interacts with Evo to produce divergent evolutionary paths.

5.8 Summary of Micro-Level Insights

Purpose: Consolidate micro-scale results.

Expectation: Summarize patterns in structural motifs, mutation tendencies, In-Life learning interactions, and the role of stochasticity.


Appendix A: Additional Minimal Neo Structures

This appendix provides detailed analysis of additional minimal Neo configurations that complement the two-node recurrent Neo presented in Section 4.2.1.

A.1 Zero-Node and Degenerate Cases

A zero-node Neo has no internal state. Formally, its state vector is empty:

V(t)=.V(t) = \varnothing.

There is only one possible configuration, which we call state 0. No Lex update is applied, because there are no nodes and therefore no activations to compute. The system can only emit a fixed output or some externally defined constant, and it cannot store any information about past inputs.

Because there is only one state, the internal dynamics form a trivial Markov chain with a single state.

Transition matrix

Let the state space be

S={0}.\mathcal{S} = \{0\}.

The one-step transition matrix is

P=[1],P = \begin{bmatrix} 1 \end{bmatrix},

meaning that if the system is in state 0 at time tt, it remains in state 0 at time t+1t+1 with probability 1.

Stationary distribution

The stationary distribution is a row vector

π=[π0],\pi = [\pi_0],

with normalization

π0=1.\pi_0 = 1.

The stationarity condition

πP=π\pi P = \pi

expands to

[π0][1]=[π0],[\pi_0] \begin{bmatrix} 1 \end{bmatrix} = [\pi_0],

which is satisfied for π0=1\pi_0 = 1. Thus the unique stationary distribution is

π=[1].\boxed{\pi = [1].}

This confirms that a zero-node Neo has no representational capacity and its internal dynamics are trivial: there is a single state that is always occupied.


A.2 Single-Node Neo

We now consider the smallest non-degenerate Neo: a single internal state node. This Neo can already pass through deterministic and stochastic regimes and store a single bit of memory. The output is always taken directly from the state node:

Y(t)=V(t).Y(t) = V(t).

We model Lex for a single node using a linear activation perturbed by symmetric noise. Let the input U(t){0,1}U(t) \in \{0,1\} be fixed to a constant value uu so that the resulting Markov chain is time-homogeneous.

The activation is

a(t)=aUU(t)+aVV(t)+b+αη(t),a(t) = a_U\,U(t) + a_V\,V(t) + b + \alpha\,\eta(t),

where:

  • aUa_U is the weight from the external input,

  • aVa_V is the self-connection weight (memory term),

  • bb is a bias,

  • α\alpha scales the stochastic perturbation,

  • η(t){1,+1}\eta(t) \in \{-1, +1\} is a Rademacher noise variable with

    P(η(t)=+1)=P(η(t)=1)=12.\mathbb{P}(\eta(t) = +1) = \mathbb{P}(\eta(t) = -1) = \tfrac12.

The state update is

V(t+1)=H(a(t)),V(t+1) = H(a(t)),

where H(x)=1H(x) = 1 if x0x \ge 0 and 00 otherwise. The output is simply

Y(t)=V(t).Y(t) = V(t).

To obtain a concrete Markov chain and show how the probabilities are derived, we fix the input to U(t)=1U(t) = 1 and choose specific parameter values.

Numerical Lex specification

Choose:

aU=1,aV=1,b=0.7,α=1.0,U(t)=1 for all t.a_U = 1,\quad a_V = 1,\quad b = -0.7,\quad \alpha = 1.0,\quad U(t) = 1 \text{ for all } t.

Then the deterministic part of the activation is

s(V(t))=aUU(t)+aVV(t)+b=11+1V(t)0.7=0.3+V(t).s(V(t)) = a_U\,U(t) + a_V\,V(t) + b = 1\cdot 1 + 1\cdot V(t) - 0.7 = 0.3 + V(t).

So we have:

  • If V(t)=0V(t) = 0, then

    s0=0.3.s_0 = 0.3.

  • If V(t)=1V(t) = 1, then

    s1=1.3.s_1 = 1.3.

The full activation for each case is

a(t)=s(V(t))+αη(t)=s(V(t))+1.0η(t).a(t) = s(V(t)) + \alpha\,\eta(t) = s(V(t)) + 1.0 \cdot \eta(t).

We now compute the transition probabilities explicitly.

Computing transition probabilities

We want P(V(t+1)=1V(t)=v)P(V(t+1) = 1 \mid V(t) = v) for v{0,1}v \in \{0,1\}.

Because η(t){1,+1}\eta(t) \in \{-1, +1\} with equal probability, we can write:

P(V(t+1)=1V(t)=v)=P(H(a(t))=1V(t)=v)=P(a(t)0V(t)=v)P(V(t+1)=1 \mid V(t) = v) = P(H(a(t))=1 \mid V(t)=v) = P(a(t) \ge 0 \mid V(t)=v)

=P(s(v)+η(t)0)=121{s(v)+10}+121{s(v)10},= P(s(v) + \eta(t) \ge 0) = \tfrac12 \mathbb{1}\{s(v) + 1 \ge 0\} + \tfrac12 \mathbb{1}\{s(v) - 1 \ge 0\},

where 1{}\mathbb{1}\{\cdot\} is the indicator function.

We apply this to each state.

Case 1: V(t)=0V(t) = 0.

Here s0=0.3s_0 = 0.3. The two possible activations are:

a+=s0+1=0.3+1=1.3,a=s01=0.31=0.7.a_+ = s_0 + 1 = 0.3 + 1 = 1.3,\quad a_- = s_0 - 1 = 0.3 - 1 = -0.7.

So:

  • a+0H(a+)=1a_+ \ge 0 \Rightarrow H(a_+) = 1,

  • a<0H(a)=0a_- < 0 \Rightarrow H(a_-) = 0.

Thus:

P(V(t+1)=1V(t)=0)=121+120=12.P(V(t+1)=1 \mid V(t)=0) = \tfrac12 \cdot 1 + \tfrac12 \cdot 0 = \tfrac12.

Similarly,

P(V(t+1)=0V(t)=0)=112=12.P(V(t+1)=0 \mid V(t)=0) = 1 - \tfrac12 = \tfrac12.

Case 2: V(t)=1V(t) = 1.

Here s1=1.3s_1 = 1.3. The two possible activations are:

a+=s1+1=1.3+1=2.3,a=s11=1.31=0.3.a_+ = s_1 + 1 = 1.3 + 1 = 2.3,\quad a_- = s_1 - 1 = 1.3 - 1 = 0.3.

Both are nonnegative, so:

  • H(a+)=1H(a_+) = 1,

  • H(a)=1H(a_-) = 1.

Therefore:

P(V(t+1)=1V(t)=1)=121+121=1,P(V(t+1)=1 \mid V(t)=1) = \tfrac12 \cdot 1 + \tfrac12 \cdot 1 = 1,

and

P(V(t+1)=0V(t)=1)=0.P(V(t+1)=0 \mid V(t)=1) = 0.

Transition matrix

Order the states as [0,1][0, 1]. The transition matrix P(1)P^{(1)} has entries

Pij(1)=P(V(t+1)=jV(t)=i).P^{(1)}_{ij} = P(V(t+1) = j \mid V(t) = i).

From the probabilities we derived:

  • From state 0:

    P(00)=0.5,P(01)=0.5.P(0\to 0) = 0.5,\quad P(0\to 1) = 0.5.

  • From state 1:

    P(10)=0,P(11)=1.P(1\to 0) = 0,\quad P(1\to 1) = 1.

Thus

P(1)=[0.50.501].P^{(1)} = \begin{bmatrix} 0.5 & 0.5 \\ 0 & 1 \end{bmatrix}.

Stationary distribution

Let the stationary distribution be

π=[π0,π1],\pi = [\pi_0, \pi_1],

with

π0+π1=1.\pi_0 + \pi_1 = 1.

The stationarity condition is

πP(1)=π.\pi P^{(1)} = \pi.

Compute the left-hand side:

πP(1)=[π0,π1][0.50.501]=[π00.5+π10,  π00.5+π11]=[0.5π0,  0.5π0+π1].\pi P^{(1)} = [\pi_0, \pi_1] \begin{bmatrix} 0.5 & 0.5 \\ 0 & 1 \end{bmatrix} = [\pi_0 \cdot 0.5 + \pi_1 \cdot 0,\; \pi_0 \cdot 0.5 + \pi_1 \cdot 1] = [0.5\pi_0,\; 0.5\pi_0 + \pi_1].

Setting this equal to [π0,π1][\pi_0, \pi_1] yields:

  1. For the first component:

    0.5π0=π0    0.5π0=0    π0=0.0.5\pi_0 = \pi_0 \;\Rightarrow\; 0.5\pi_0 = 0 \;\Rightarrow\; \pi_0 = 0.

  2. Normalization then forces:

    π1=1.\pi_1 = 1.

So the unique stationary distribution is

π=(0,1).\boxed{\pi = (0, 1).}

In words, under this Lex and fixed input, the single-node Neo eventually spends almost all its time in the state V=1V=1: once it flips to 1 it never returns to 0. The stochasticity only affects the transient path from 0 to 1.


A.3 Two-Node Neo

We now consider the simplest interacting Neo: two binary state nodes

V(t)=(V1(t),V2(t)){0,1}2.V(t) = (V_1(t), V_2(t)) \in \{0,1\}^2.

We will analyze three basic topologies:

  1. Feedforward: V1V2V_1 \to V_2.

  2. Parallel: both nodes driven independently by the same input.

  3. Feedback: V1V2V_1 \leftrightarrow V_2.

In each case, we specify Lex for both nodes using simple numeric parameters and derive the full transition matrix over the four states, followed by the stationary distribution.

We order the joint states as:

1(0,0),2(0,1),3(1,0),4(1,1).1 \equiv (0,0), \quad 2 \equiv (0,1), \quad 3 \equiv (1,0), \quad 4 \equiv (1,1).


A.3.A Feedforward (V1 → V2)

We first analyze a two-node feedforward motif where V1V_1 depends only on the external input UU and noise, and V2V_2 depends only on V1V_1 and noise. This already yields richer dynamics than the single-node case.

Diagram:

We again fix the input to a constant value U(t)=uU(t) = u to obtain a time-homogeneous Markov chain. The randomness then comes solely from internal noise.

Lex specification

Node 1 activation:

a1(t)=1.0U(t)0.2+0.7η1(t),a_1(t) = 1.0 \cdot U(t) - 0.2 + 0.7 \cdot \eta_1(t),

with η1(t){1,+1}\eta_1(t) \in \{-1,+1\} and P(η1=±1)=12\mathbb{P}(\eta_1 = \pm 1) = \tfrac12.

Node 2 activation:

a2(t)=1.4V1(t)0.1+0.9η2(t),a_2(t) = 1.4 \cdot V_1(t) - 0.1 + 0.9 \cdot \eta_2(t),

with η2(t){1,+1}\eta_2(t) \in \{-1,+1\} independent of η1(t)\eta_1(t).

The state updates are

V1(t+1)=H(a1(t)),V2(t+1)=H(a2(t)).V_1(t+1) = H(a_1(t)), \quad V_2(t+1) = H(a_2(t)).

We will compute the transition matrix for two fixed inputs: u=0u = 0 and u=1u = 1.


Case u = 0: computing node-wise probabilities

Set U(t)=0U(t) = 0.

For node 1, the deterministic part is:

s1=1.000.2=0.2.s_1 = 1.0 \cdot 0 - 0.2 = -0.2.

The two activations are:

a1,+=s1+0.7=0.2+0.7=0.5,a1,=s10.7=0.20.7=0.9.a_{1,+} = s_1 + 0.7 = -0.2 + 0.7 = 0.5, \quad a_{1,-} = s_1 - 0.7 = -0.2 - 0.7 = -0.9.

So:

  • H(a1,+)=1H(a_{1,+}) = 1,

  • H(a1,)=0H(a_{1,-}) = 0,

and

P(V1(t+1)=1U=0)=121+120=0.5.P(V_1(t+1) = 1 \mid U = 0) = \tfrac12 \cdot 1 + \tfrac12 \cdot 0 = 0.5.

Thus

P(V1=1U=0)=0.5,P(V1=0U=0)=0.5.P(V_1' = 1 \mid U = 0) = 0.5, \quad P(V_1' = 0 \mid U = 0) = 0.5.

For node 2, the deterministic part depends on V1(t)V_1(t):

s2(V1)=1.4V10.1.s_2(V_1) = 1.4 \cdot V_1 - 0.1.

If V1(t)=0V_1(t) = 0:

s2(0)=0.1,s_2(0) = -0.1,

so

a2,+=0.1+0.9=0.8,a2,=0.10.9=1.0.a_{2,+} = -0.1 + 0.9 = 0.8, \quad a_{2,-} = -0.1 - 0.9 = -1.0.

Hence:

  • H(a2,+)=1H(a_{2,+}) = 1,

  • H(a2,)=0H(a_{2,-}) = 0,

and

P(V2(t+1)=1V1(t)=0)=12.P(V_2(t+1) = 1 \mid V_1(t) = 0) = \tfrac12.

If V1(t)=1V_1(t) = 1:

s2(1)=1.40.1=1.3,s_2(1) = 1.4 - 0.1 = 1.3,

so

a2,+=1.3+0.9=2.2,a2,=1.30.9=0.4.a_{2,+} = 1.3 + 0.9 = 2.2, \quad a_{2,-} = 1.3 - 0.9 = 0.4.

Both are nonnegative, so

P(V2(t+1)=1V1(t)=1)=1.P(V_2(t+1) = 1 \mid V_1(t) = 1) = 1.

In summary (for u=0u = 0):

P(V1=1)=0.5,P(V1=0)=0.5,P(V_1' = 1) = 0.5, \quad P(V_1' = 0) = 0.5,

P(V2=1V1=0)=0.5,P(V2=1V1=1)=1.P(V_2' = 1 \mid V_1 = 0) = 0.5, \quad P(V_2' = 1 \mid V_1 = 1) = 1.

We now build the 4-state transition matrix.

Transition probabilities for each joint state (u = 0)

Let X(t)=(V1(t),V2(t))X(t) = (V_1(t), V_2(t)). Because V1V_1' depends only on UU and V2V_2' depends only on V1(t)V_1(t) and noise, and noise for node 1 and 2 are independent, we have

P(X(t+1)=(v1,v2)X(t)=(v1,v2),U=0)=P(V1=v1U=0)P(V2=v2V1=v1).P(X(t+1) = (v_1', v_2') \mid X(t) = (v_1, v_2), U = 0) = P(V_1' = v_1' \mid U = 0) \cdot P(V_2' = v_2' \mid V_1 = v_1).

We compute row by row.

Row 1: current state (0,0).

Here V1=0V_1 = 0, but note that P(V1U=0)P(V_1' \mid U = 0) does not depend on the current state. We have:

P(V1=0)=0.5,P(V1=1)=0.5.P(V_1' = 0) = 0.5, \quad P(V_1' = 1) = 0.5.

For V2V_2':

  • If V1=0V_1 = 0, then P(V2=1)=0.5P(V_2' = 1) = 0.5, P(V2=0)=0.5P(V_2' = 0) = 0.5.

Therefore:

  • P((0,0)(0,0))=P(V1=0)P(V2=0V1=0)=0.50.5=0.25,P((0,0) \to (0,0)) = P(V_1' = 0) \cdot P(V_2' = 0 \mid V_1 = 0) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(0,1))=0.50.5=0.25,P((0,0) \to (0,1)) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(1,0))=P(V1=1)P(V2=0V1=0)=0.50.5=0.25,P((0,0) \to (1,0)) = P(V_1' = 1) \cdot P(V_2' = 0 \mid V_1 = 0) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(1,1))=0.50.5=0.25.P((0,0) \to (1,1)) = 0.5 \cdot 0.5 = 0.25.

The row sums to 1.

Row 2: current state (0,1).

The current value of V2V_2 does not influence V1V_1' or V2V_2' (which depends on V1V_1, not V2V_2). So the same reasoning applies:

P((0,1)j)=P((0,0)j)P((0,1) \to j) = P((0,0) \to j)

for each joint next state jj. So the second row is identical to the first:

(0.25,0.25,0.25,0.25).(0.25, 0.25, 0.25, 0.25).

Row 3: current state (1,0).

Now V1=1V_1 = 1. The distribution of V1V_1' is still independent of the current state and remains

P(V1=0)=0.5,P(V1=1)=0.5.P(V_1' = 0) = 0.5, \quad P(V_1' = 1) = 0.5.

However, V2V_2' conditional probabilities change because they depend on V1(t)V_1(t), which is 1 here:

P(V2=1V1=1)=1,P(V2=0V1=1)=0.P(V_2' = 1 \mid V_1 = 1) = 1, \quad P(V_2' = 0 \mid V_1 = 1) = 0.

Thus:

  • P((1,0)(0,0))=P(V1=0)P(V2=0V1=1)=0.50=0,P((1,0) \to (0,0)) = P(V_1' = 0) \cdot P(V_2' = 0 \mid V_1 = 1) = 0.5 \cdot 0 = 0,

  • P((1,0)(0,1))=0.51=0.5,P((1,0) \to (0,1)) = 0.5 \cdot 1 = 0.5,

  • P((1,0)(1,0))=P(V1=1)P(V2=0V1=1)=0.50=0,P((1,0) \to (1,0)) = P(V_1' = 1) \cdot P(V_2' = 0 \mid V_1 = 1) = 0.5 \cdot 0 = 0,

  • P((1,0)(1,1))=0.51=0.5.P((1,0) \to (1,1)) = 0.5 \cdot 1 = 0.5.

So the third row is:

(0,0.5,0,0.5).(0, 0.5, 0, 0.5).

Row 4: current state (1,1).

Same as row 3, because V1=1V_1 = 1 and V2V_2 is irrelevant to V1V_1' and V2V_2':

(0,0.5,0,0.5).(0, 0.5, 0, 0.5).

Transition matrix (u = 0)

Collecting all rows:

Pu=0FF=[0.250.250.250.250.250.250.250.2500.500.500.500.5].P^{FF}_{u=0} = \begin{bmatrix} 0.25 & 0.25 & 0.25 & 0.25 \\ 0.25 & 0.25 & 0.25 & 0.25 \\ 0 & 0.5 & 0 & 0.5 \\ 0 & 0.5 & 0 & 0.5 \end{bmatrix}.

Stationary distribution (u = 0)

Let the stationary distribution be

π=(π1,π2,π3,π4),\pi = (\pi_1, \pi_2, \pi_3, \pi_4),

with

π1+π2+π3+π4=1.\pi_1 + \pi_2 + \pi_3 + \pi_4 = 1.

By symmetry of rows (1 and 2 identical; 3 and 4 identical) and columns, it is natural to look for a solution with

π1=π3=a,π2=π4=b,a+b=12.\pi_1 = \pi_3 = a, \quad \pi_2 = \pi_4 = b, \quad a + b = \tfrac12.

We now impose stationarity on, say, π1\pi_1:

π1=π1=i=14πiPu=0FF(i,1)=π10.25+π20.25+π30+π40.\pi_1' = \pi_1 = \sum_{i=1}^4 \pi_i P^{FF}_{u=0}(i,1) = \pi_1 \cdot 0.25 + \pi_2 \cdot 0.25 + \pi_3 \cdot 0 + \pi_4 \cdot 0.

Substitute π1=a,π2=b\pi_1 = a, \pi_2 = b:

a=0.25a+0.25b.a = 0.25 a + 0.25 b.

Using the constraint a+b=0.5a + b = 0.5, we solve:

a=0.25(a+b)=0.250.5=0.125=18.a = 0.25 (a + b) = 0.25 \cdot 0.5 = 0.125 = \tfrac18.

Therefore

b=0.5a=0.50.125=0.375=38.b = 0.5 - a = 0.5 - 0.125 = 0.375 = \tfrac38.

Thus

πu=0FF=(18,38,18,38).\boxed{\pi^{FF}_{u=0} = \Big(\tfrac18, \tfrac38, \tfrac18, \tfrac38\Big).}


Case u = 1

We now set U(t)=1U(t) = 1.

For node 1, deterministic part:

s1=1.010.2=0.8.s_1 = 1.0 \cdot 1 - 0.2 = 0.8.

Then:

a1,+=0.8+0.7=1.5,a1,=0.80.7=0.1.a_{1,+} = 0.8 + 0.7 = 1.5, \quad a_{1,-} = 0.8 - 0.7 = 0.1.

Both nonnegative, so

P(V1=1U=1)=1.P(V_1' = 1 \mid U = 1) = 1.

For node 2, deterministic part:

s2(V1)=1.4V10.1.s_2(V_1) = 1.4 V_1 - 0.1.

As before, if V1=0V_1 = 0, P(V2=1)=0.5P(V_2' = 1) = 0.5, and if V1=1V_1 = 1, P(V2=1)=1P(V_2' = 1) = 1. However, under U=1U = 1, V1V_1' is always 1, so after one step the system will behave as if V1(t)=1V_1(t) = 1 at all future times.

To keep the Markov chain defined in terms of current state, we again use:

P(X(t+1)=(v1,v2)X(t)=(v1,v2),U=1)=P(V1=v1U=1)P(V2=v2V1=v1).P(X(t+1) = (v_1', v_2') \mid X(t) = (v_1, v_2), U = 1) = P(V_1' = v_1' \mid U = 1) P(V_2' = v_2' \mid V_1 = v_1).

But now P(V1=0)=0P(V_1' = 0) = 0 and P(V1=1)=1P(V_1' = 1) = 1, so transitions can only go to states with V1=1V_1' = 1 (i.e., states 3 or 4).

Carrying through the same logic as before yields:

Pu=1FF=[000.50.5000.50.500010001].P^{FF}_{u=1} = \begin{bmatrix} 0 & 0 & 0.5 & 0.5 \\ 0 & 0 & 0.5 & 0.5 \\ 0 & 0 & 0 & 1 \\ 0 & 0 & 0 & 1 \end{bmatrix}.

The last state (1,1) is absorbing, and all other states have a nonzero probability of eventually reaching it. Thus the unique stationary distribution is:

πu=1FF=(0,0,0,1).\boxed{\pi^{FF}_{u=1} = (0,0,0,1).}


A.3.B Parallel (U → V1 and U → V2)

Now we consider a parallel topology where both nodes receive the same input but do not directly interact. This is the smallest multi-bit output that can still be treated as two independent single-node Neos.

Diagram:

Assume both nodes obey the same single-node Lex rule used in Section A.2, and that their noise processes are independent. For a fixed input uu, each node has the same transition probabilities as the single-node case, and the joint transition matrix is the Kronecker product of the single-node matrix with itself.

For simplicity, consider U(t)=0U(t) = 0 with the single-node parameters chosen so that

P(Vi=1U=0)=0.5,i{1,2}.P(V_i' = 1 \mid U = 0) = 0.5, \quad i \in \{1,2\}.

Then each node at each step is an independent Bernoulli(0.5) random variable, independent of its previous state. The joint distribution over the pair is uniform over the four states.

Thus for any current state,

P(X(t+1)=(v1,v2))=0.25P(X(t+1) = (v_1', v_2')) = 0.25

for each of the four configurations.

The transition matrix is therefore:

Pu=0PAR=[0.250.250.250.250.250.250.250.250.250.250.250.250.250.250.250.25].P^{PAR}_{u=0} = \begin{bmatrix} 0.25 & 0.25 & 0.25 & 0.25 \\ 0.25 & 0.25 & 0.25 & 0.25 \\ 0.25 & 0.25 & 0.25 & 0.25 \\ 0.25 & 0.25 & 0.25 & 0.25 \end{bmatrix}.

The stationary distribution π\pi must satisfy πPu=0PAR=π\pi P^{PAR}_{u=0} = \pi. Because every row is identical and uniform, the unique normalized solution is

πu=0PAR=(0.25,0.25,0.25,0.25).\boxed{\pi^{PAR}_{u=0} = (0.25, 0.25, 0.25, 0.25).}

The parallel 2-node Neo under symmetric stochastic input is therefore a uniform random generator over the four joint states.

For U(t)=1U(t) = 1 and our earlier single-node parameters, each node deterministically goes to state 1. Hence (1,1) is absorbing, and the stationary distribution is again

πu=1PAR=(0,0,0,1).\pi^{PAR}_{u=1} = (0,0,0,1).


A.3.C Feedback (V1 ↔ V2)

Finally, we consider a feedback topology where each node depends on the other's previous state. This is the smallest genuine recurrent Neo and can already exhibit nontrivial stochastic dynamics, including concentration on particular states.

Diagram:

We again fix U(t)=uU(t) = u and choose explicit parameters.

Lex specification

Node 1:

a1(t)=0.7U(t)+1.0V2(t)0.5+0.6η1(t).a_1(t) = 0.7 \cdot U(t) + 1.0 \cdot V_2(t) - 0.5 + 0.6 \cdot \eta_1(t).

Node 2:

a2(t)=0.4U(t)+1.5V1(t)0.2+0.7η2(t).a_2(t) = -0.4 \cdot U(t) + 1.5 \cdot V_1(t) - 0.2 + 0.7 \cdot \eta_2(t).

Both noise variables η1,η2\eta_1, \eta_2 take values {1,+1}\{-1,+1\} with probability 0.5 and are independent.

Updates:

V1(t+1)=H(a1(t)),V2(t+1)=H(a2(t)).V_1(t+1) = H(a_1(t)), \quad V_2(t+1) = H(a_2(t)).

We again consider two cases: u=0u = 0 and u=1u = 1.


Case u = 0

Set U(t)=0U(t) = 0. The deterministic parts are:

s1(V2)=0.70+1.0V20.5=V20.5,s_1(V_2) = 0.7 \cdot 0 + 1.0 V_2 - 0.5 = V_2 - 0.5,

s2(V1)=0.40+1.5V10.2=1.5V10.2.s_2(V_1) = -0.4 \cdot 0 + 1.5 V_1 - 0.2 = 1.5 V_1 - 0.2.

We compute node-wise probabilities.

Node 1:

If V2=0V_2 = 0:

s1(0)=0.5,s_1(0) = -0.5,

so

a1,+=0.5+0.6=0.1,a1,=0.50.6=1.1.a_{1,+} = -0.5 + 0.6 = 0.1, \quad a_{1,-} = -0.5 - 0.6 = -1.1.

Thus

P(V1=1V2=0,U=0)=121+120=0.5.P(V_1' = 1 \mid V_2 = 0, U = 0) = \tfrac12 \cdot 1 + \tfrac12 \cdot 0 = 0.5.

If V2=1V_2 = 1:

s1(1)=10.5=0.5,s_1(1) = 1 - 0.5 = 0.5,

so

a1,+=0.5+0.6=1.1,a1,=0.50.6=0.1.a_{1,+} = 0.5 + 0.6 = 1.1, \quad a_{1,-} = 0.5 - 0.6 = -0.1.

Again,

P(V1=1V2=1,U=0)=0.5.P(V_1' = 1 \mid V_2 = 1, U = 0) = 0.5.

So for u=0u = 0, node 1 is symmetric:

P(V1=1V2=0)=P(V1=1V2=1)=0.5.P(V_1' = 1 \mid V_2 = 0) = P(V_1' = 1 \mid V_2 = 1) = 0.5.

Node 2:

If V1=0V_1 = 0:

s2(0)=0.2,s_2(0) = -0.2,

a2,+=0.2+0.7=0.5,a2,=0.20.7=0.9.a_{2,+} = -0.2 + 0.7 = 0.5, \quad a_{2,-} = -0.2 - 0.7 = -0.9.

So

P(V2=1V1=0,U=0)=0.5.P(V_2' = 1 \mid V_1 = 0, U = 0) = 0.5.

If V1=1V_1 = 1:

s2(1)=1.50.2=1.3,s_2(1) = 1.5 - 0.2 = 1.3,

a2,+=1.3+0.7=2.0,a2,=1.30.7=0.6.a_{2,+} = 1.3 + 0.7 = 2.0, \quad a_{2,-} = 1.3 - 0.7 = 0.6.

Both nonnegative, so

P(V2=1V1=1,U=0)=1.P(V_2' = 1 \mid V_1 = 1, U = 0) = 1.

We now build the transition matrix.

State (0,0): here V1=0,V2=0V_1 = 0, V_2 = 0.

  • Node1: P(V1=1V2=0)=0.5P(V_1' = 1 \mid V_2 = 0) = 0.5, P(V1=0)=0.5P(V_1' = 0) = 0.5.

  • Node2: P(V2=1V1=0)=0.5P(V_2' = 1 \mid V_1 = 0) = 0.5, P(V2=0)=0.5P(V_2' = 0) = 0.5.

Joint transitions:

  • P((0,0)(0,0))=0.50.5=0.25,P((0,0) \to (0,0)) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(0,1))=0.50.5=0.25,P((0,0) \to (0,1)) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(1,0))=0.50.5=0.25,P((0,0) \to (1,0)) = 0.5 \cdot 0.5 = 0.25,

  • P((0,0)(1,1))=0.50.5=0.25.P((0,0) \to (1,1)) = 0.5 \cdot 0.5 = 0.25.

State (0,1): V1=0,V2=1V_1 = 0, V_2 = 1.

  • Node1: P(V1=1V2=1)=0.5P(V_1' = 1 \mid V_2 = 1) = 0.5, P(V1=0)=0.5P(V_1' = 0) = 0.5.

  • Node2: P(V2=1V1=0)=0.5P(V_2' = 1 \mid V_1 = 0) = 0.5, P(V2=0)=0.5P(V_2' = 0) = 0.5.

Thus the row is identical to state (0,0):

(0.25,0.25,0.25,0.25).(0.25, 0.25, 0.25, 0.25).

State (1,0): V1=1,V2=0V_1 = 1, V_2 = 0.

  • Node1: P(V1=1V2=0)=0.5P(V_1' = 1 \mid V_2 = 0) = 0.5.

  • Node2: P(V2=1V1=1)=1P(V_2' = 1 \mid V_1 = 1) = 1, P(V2=0)=0P(V_2' = 0) = 0.

So:

  • P((1,0)(0,0))=P(V1=0)P(V2=0V1=1)=0.50=0,P((1,0) \to (0,0)) = P(V_1' = 0) \cdot P(V_2' = 0 \mid V_1 = 1) = 0.5 \cdot 0 = 0,

  • P((1,0)(0,1))=0.51=0.5,P((1,0) \to (0,1)) = 0.5 \cdot 1 = 0.5,

  • P((1,0)(1,0))=0.50=0,P((1,0) \to (1,0)) = 0.5 \cdot 0 = 0,

  • P((1,0)(1,1))=0.51=0.5.P((1,0) \to (1,1)) = 0.5 \cdot 1 = 0.5.

State (1,1): V1=1,V2=1V_1 = 1, V_2 = 1.

  • Node1: P(V1=1V2=1)=0.5P(V_1' = 1 \mid V_2 = 1) = 0.5.

  • Node2: P(V2=1V1=1)=1P(V_2' = 1 \mid V_1 = 1) = 1.

So again:

(1,1)(0,0)=0,(1,1)(0,1)=0.5,(1,1)(1,0)=0,(1,1)(1,1)=0.5.(1,1) \to (0,0) = 0, \quad (1,1) \to (0,1) = 0.5, \quad (1,1) \to (1,0) = 0, \quad (1,1) \to (1,1) = 0.5.

Collecting these rows, we obtain:

Pu=0FB=[0.250.250.250.250.250.250.250.2500.500.500.500.5].P^{FB}_{u=0} = \begin{bmatrix} 0.25 & 0.25 & 0.25 & 0.25 \\ 0.25 & 0.25 & 0.25 & 0.25 \\ 0 & 0.5 & 0 & 0.5 \\ 0 & 0.5 & 0 & 0.5 \end{bmatrix}.

This matrix is identical to the feedforward case with u=0u = 0, so the stationary distribution is again:

πu=0FB=(18,38,18,38).\boxed{\pi^{FB}_{u=0} = \Big(\tfrac18, \tfrac38, \tfrac18, \tfrac38\Big).}

Case u = 1 (feedback)

With U=1U = 1, the deterministic parts change, and the Markov chain becomes more biased toward the state (1,1). A similar step-by-step analysis (omitted here to avoid redundancy) yields:

Pu=1FB=[0.250.250.250.25000.50.500.500.50001].P^{FB}_{u=1} = \begin{bmatrix} 0.25 & 0.25 & 0.25 & 0.25 \\ 0 & 0 & 0.5 & 0.5 \\ 0 & 0.5 & 0 & 0.5 \\ 0 & 0 & 0 & 1 \end{bmatrix}.

The state (1,1) is absorbing and is reachable from every other state under the dynamics, so the unique stationary distribution is

πu=1FB=(0,0,0,1).\boxed{\pi^{FB}_{u=1} = (0,0,0,1).}

This two-node feedback Neo therefore converges, under this Lex choice, to a fully "activated" state in the long run.

Last updated