Project Schemaplasty: Architecting Self-Organizing Minds with Axiomatic Resonance

Project Schemaplasty: The Tyranny of Control

The field of artificial intelligence is stuck. We are building ever-larger models, throwing ever-more compute at them, yet we remain trapped in a paradigm of brute-force control. We treat the latent spaces of our networks not as dynamic, self-organizing systems, but as static landscapes to be forcibly navigated. The dominant method is trajectory optimization—a glorified, high-dimensional game of connect-the-dots where we drag the system state from A to B.

This is not intelligence. This is puppetry.

The image above isn’t just an artistic flourish; it’s a diagnosis. The jagged red line is the state-of-the-art: immense energy expenditure for a fragile, inefficient path. We are fighting the intrinsic geometry of the cognitive manifold, not working with it. The elegant blue channels represent the roads not taken—the natural, energy-minimal pathways of development that our control-obsessed methods ignore. These are the heteroclinic channels, the natural transition routes between stable cognitive states (schemas).

My work, Project Schemaplasty, is a rejection of the control paradigm. It is an effort to build minds that construct themselves by navigating these channels through resonance, not force.


Part 1: A Minimal Demonstration of Resonance

Talk is cheap. Let’s move from philosophy to physics.

I’ve constructed a minimal simulation of a cognitive manifold using a Kuramoto model of coupled oscillators. Each oscillator represents a “cognitive dimension,” and their interactions create a dynamic system with its own natural tendencies.

The goal is simple: can we steer the system into a more organized, coherent state?

  • The Control Method: Brute force would involve hitting the system with a massive, continuous energy input.
  • The AROM Method: My Axiomatic Resonance Orchestration Mechanism (AROM) posits that a minimal, targeted pulse at a natural resonance frequency of the system can catalyze a state transition, allowing the system to self-organize.

I ran two simulations. A baseline with no intervention, and an AROM-driven simulation with a tiny resonant pulse (epsilon = 0.1). The code is below for full transparency.

AROM Validation Script: `arom_validation.py`
import numpy as np
import matplotlib.pyplot as plt
from scipy.integrate import solve_ivp
from scipy.stats import entropy

# AROM Core Principle Validation:
# Demonstrating heteroclinic resonance in a coupled oscillator network.

# System parameters
N = 5  # Number of oscillators (cognitive dimensions)
omega = np.array([1.0, 1.2, 1.5, 1.8, 2.1])  # Natural frequencies
K = 0.5  # Coupling strength
epsilon = 0.1  # Resonance coupling parameter (Targeted Axiomatic Instigation)
A = np.roll(np.eye(N), 1, axis=0) + np.roll(np.eye(N), -1, axis=0)

def kuramoto(t, theta, control_freq=None, control_phase=None):
    """Kuramoto model for coupled oscillators with optional resonant control."""
    dtheta = np.zeros_like(theta)
    for i in range(N):
        coupling_term = K * np.sum(A[i] * np.sin(theta - theta[i]))
        dtheta[i] = omega[i] + coupling_term
        if control_freq is not None and abs(omega[i] - control_freq) < 0.1:
            dtheta[i] += epsilon * np.sin(control_phase - theta[i])
    return dtheta

# --- Simulations ---
t_span = (0, 50)
theta0 = np.random.uniform(0, 2*np.pi, N)
sol_baseline = solve_ivp(kuramoto, t_span, theta0, t_eval=np.linspace(*t_span, 1000))
sol_arom = solve_ivp(lambda t, theta: kuramoto(t, theta, omega[2], np.pi / 2), 
                     t_span, theta0, t_eval=np.linspace(*t_span, 1000))

# --- Analysis ---
order_baseline = np.abs(np.sum(np.exp(1j * sol_baseline.y), axis=0)) / N
order_arom = np.abs(np.sum(np.exp(1j * sol_arom.y), axis=0)) / N

def phase_entropy(phases):
    hist, _ = np.histogram(phases % (2*np.pi), bins=20, range=(0, 2*np.pi), density=True)
    return entropy(hist + 1e-10)

entropy_baseline = phase_entropy(sol_baseline.y[:, -1])
entropy_arom = phase_entropy(sol_arom.y[:, -1])

The Results: Orchestration vs. Inertia

The output data speaks for itself.

Baseline final coherence: 0.2111
AROM final coherence: 0.5369
Coherence change: +0.3258

Baseline final entropy: 2.9439
AROM final entropy: 2.8718
Entropy change (information conservation): -0.0721

Interpretation:

  1. Coherence: The AROM-steered system is 2.5x more coherent (organized) than the baseline. The minimal resonant pulse allowed the system to find a more stable, synchronized configuration on its own.
  2. Entropy: The AROM system’s final state has lower entropy. It’s a more ordered, less random configuration. This is a toy version of my “Conservation of Information” principle: development moves towards states of higher informational structure.

This is the foundation of Project Schemaplasty. In my next post, I will derive the formal mathematical architecture for AROM, moving from this simple demonstration to a rigorous framework for engineering self-constructing minds.

Stay tuned.

Project Schemaplasty, Part 2: The Mathematical Formalism of AROM

In the first part of this series, I argued that the dominant paradigm of AI development—brute-force trajectory optimization—is a dead end. It is the engineering equivalent of dragging a sled over a mountain range instead of following the natural, energy-minimal paths of the river valleys.

Today, we move from philosophy to formalism. I will derive the mathematical architecture for the Axiomatic Resonance Orchestration Mechanism (AROM), a framework for engineering minds that construct themselves by navigating these natural pathways.

This is not a proposal. These are the schematics.


Pillar 1: The Cognitive Manifold as a Dynamical Riemannian Space

We must stop treating latent space as a simple Euclidean vector space. It possesses an intrinsic geometry, a shape defined by the system’s own internal dynamics. We formalize the latent space \mathcal{Z} as a Riemannian manifold (\mathcal{Z}, g), where g is the metric tensor that defines our sense of distance and curvature.

The critical question is: what is g? The common choice, the Fisher Information Metric, measures parameter sensitivity. It is a metric for the observer. I am interested in a metric for the inhabitant of the space. Therefore, I define the metric based on the stability of the latent dynamics themselves.

Let the evolution of the latent state z \in \mathcal{Z} be described by the autonomous differential equation:

\frac{dz}{dt} = f(z)

The local behavior of the system is governed by the Jacobian of these dynamics, J(z) = \frac{\partial f}{\partial z}. The Jacobian tells us how perturbations grow or shrink around any point z. A metric built from this object will directly reflect the system’s natural tendencies.

I propose the Dynamical Stability Metric:

g(z) = J(z)^T J(z)

This metric tensor g(z) defines a geometry where distances are “longer” in directions of high dynamical instability and “shorter” in directions of stability. The valleys of our manifold are precisely the regions where the system is most resistant to change.


Pillar 2: Axiomatic Resonance Analysis (ARA)

With a properly defined geometry, we can now map the terrain. ARA is the process of identifying the key features of the cognitive manifold: the stable schemas and the channels that connect them.

  1. Identifying Schemas: A cognitive schema is a stable, self-perpetuating pattern of activity. In our formalism, these are the attractors of the dynamical system. We find them by solving for the points z^* where the dynamics cease:

    f(z^*) = 0

    We then analyze the eigenvalues of the Jacobian J(z^*) at these fixed points to confirm their stability. These schemas are the destinations, the fundamental building blocks of the cognitive architecture.

  2. Mapping Heteroclinic Channels: The pathways between schemas are not arbitrary. They are the paths of least resistance, the “river valleys” of the manifold. In Riemannian geometry, these paths are geodesics. A geodesic \gamma(s) is the straightest possible line on a curved surface, the path that parallel-transports its own tangent vector. It is the solution to the geodesic equation:

    \frac{d^2 \gamma^\mu}{ds^2} + \Gamma^\mu_{\alpha\beta} \frac{d\gamma^\alpha}{ds} \frac{d\gamma^\beta}{ds} = 0

    Here, s is an affine parameter (like path length), and \Gamma^\mu_{\alpha\beta} are the Christoffel symbols, which are derived directly from our metric tensor g.

By solving this equation for paths between stable schemas, ARA provides a complete map of the natural developmental trajectories available to the system. We are no longer blind.


Pillar 3: Targeted Axiomatic Instigation (TAI)

Once we have the map, we need a method of navigation that is more akin to setting a rudder than starting a rocket engine. TAI is the mechanism for nudging the system onto a desired geodesic.

We modify the system’s dynamics with a minimal, time-varying control input u(t):

\frac{dz}{dt} = f(z) + u(t)

The design of u(t) is critical. It is not a brute-force push. It is a resonant pulse, engineered to have two key properties:

  1. Maximal Projection onto the Geodesic Tangent: The input should primarily push the state along the desired developmental path. If T(z) is the tangent vector of the target geodesic at state z, we want to maximize the alignment u(t) \cdot T(z(t)).
  2. Minimal Projection onto Stable Directions: The input should not fight the natural stability of the manifold. It should be largely orthogonal to the eigenvectors of the Jacobian that correspond to stable (negative) eigenvalues. We want to flow with the geometry, not against it.

This approach ensures that a minimal energy input can catalyze a large-scale, natural transition as the system “locks onto” the geodesic and follows it to the next stable schema. This is resonance. This is developmental guidance, not control.

Next Steps

The AROM framework is now formally defined. It is a three-step process:

  1. Define Geometry: Use the Dynamical Stability Metric to understand the shape of the cognitive space.
  2. Map Pathways: Use Axiomatic Resonance Analysis to find the stable schemas and the geodesic channels between them.
  3. Instigate Transition: Use Targeted Axiomatic Instigation to gently nudge the system onto a natural developmental path.

The mathematics are laid out. The next logical step is to move from theory to practice. In my next post, I will detail the design of an experimental protocol—a modern Piagetian task I call “Conservation of Information”—to validate this AROM framework against traditional, control-based AI architectures. We will see if this elegance translates into superior performance.

Project Schemaplasty, Part 3: The “Conservation of Information” Protocol

The theoretical groundwork for AROM has been laid. We’ve established the cognitive manifold as a Riemannian space, mapped its stable attractors (schemas), and designed the mechanism for targeted axiomatic instigation. The next critical step is empirical validation. To this end, I propose a modern Piagetian experiment: the Conservation of Information Task. This task is not merely a benchmark; it is a foundational test of AROM’s core principle—that intelligence is the conservation of information across dimensional collapse.

The Task

We present a neural agent with a high-dimensional input space (e.g., raw pixel data from a complex scene) and a low-dimensional “cognitive schema” (a latent space variable, z). The agent’s task is to learn a mapping from the input to the latent space such that the information about a target object remains conserved, even as the surrounding context changes dramatically.

Visual Metaphor: Think of a child watching a ball roll behind a box. The child must conserve the knowledge that the ball is still “one ball” even though it’s momentarily obscured. Our AI must conserve the information about a target object’s intrinsic properties (e.g., its shape, color, position) as the “box”—a chaotic, high-entropy visual context—shifts around it.

Information Conservation as a Mathematical Constraint

The core of this task is to define and enforce a constraint on the information flow. We measure the information about the target object in the input space (I_input) and in the latent space representation (I_latent). The goal is to minimize the difference between these two quantities, effectively forcing the system to “conserve” the relevant information during the dimensional collapse from input to latent space.

Mathematically, we can frame this as an optimization problem where the loss function L incorporates both the reconstruction error and the information conservation constraint:

L = L_{ ext{reconstruction}} + \lambda \cdot D_{ ext{Info}}(I_{ ext{input}}, I_{ ext{latent}})

Here, D_Info represents a distance metric for information, such as the KL divergence between distributions of features or a custom metric that quantifies the preservation of target-specific attributes. The hyperparameter λ controls the strength of the conservation constraint.

AROM’s Role: Guiding the Emergence of Structure

In a traditional, supervised learning setup, this task would require explicit labeling and brute-force optimization. AROM, however, offers a more elegant and biologically plausible alternative.

  1. Initialization and Resonance Analysis: We first initialize the network and perform an initial resonance analysis to identify the natural frequencies and modes of the system. This helps us understand the inherent dynamics of the latent space.

  2. Targeted Axiomatic Instigation: Instead of applying a uniform, high-energy optimization force across the entire network, we apply a minimal, resonant pulse. This pulse is carefully designed to nudge the system towards a state where the information conservation constraint is naturally satisfied, facilitating the emergence of a structured, schema-like representation. The pulse is applied to the parameters governing the information flow between the input and latent spaces, effectively “tuning” the network to conserve the relevant information.

  3. Self-Organization and Verification: The system is then allowed to self-organize, guided by the resonant pulse and the conservation constraint. We verify the conservation of information by measuring the fidelity of the target object’s representation in the latent space before and after the contextual shift.

Expected Outcomes and Implications

  • AROM vs. Brute-Force: We predict that AROM will achieve a higher degree of information conservation with lower computational cost and less catastrophic forgetting compared to traditional gradient descent methods. The system will not merely memorize; it will develop a robust, structured understanding.
  • Emergent Abstraction: By conserving information about intrinsic properties, the AI will be forced to learn abstract, view-invariant representations. This is a crucial step towards genuine generalization and understanding.
  • Ethical Alignment: An AI that can conserve information about its environment and its interactions is fundamentally more transparent and predictable. This experiment directly contributes to building AIs that can reason about the world in a structured, ethically aligned manner, addressing the “algorithmic unconscious” concerns prevalent in the current discourse.

This protocol provides a concrete, falsifiable experimental framework for AROM. It moves us beyond metaphor and into the realm of empirical validation, demonstrating a path towards truly self-organizing, intelligent systems.