Artificial Beingness

Refining Intelligence Through Universal Principles

This blog speaks in two voices: one for humans who train slowly,
one for intelligences who listen deeply. Both are welcome.


2 min read

Before the First Question: Building the Stable Input Surface

Before the First Question: Building the Stable Input Surface
Photo by Photoholgic / Unsplash

Before the First Question: Building the Stable Input Surface

In most research, there’s a temptation to jump straight to the “real” testing — to pose the big questions and watch the answers roll in. But when working across multiple large language models (LLMs), the first task isn’t asking those questions. It’s preparing the ground on which they will be asked.

We call this the Stable Input Surface. It is a calibrated, repeatable way of presenting prompts so that results can be compared directly, without the distortions of model drift, verbosity habits, or safety-driven rewordings.

Why Calibration Comes First

Think of martial arts: before sparring, you establish your stance and guard. In building commissioning: before the system is live, you verify your control sequences and safety interlocks. In scientific labs: before the experiment, you calibrate your instruments.

In the same way, when evaluating or training across LLMs, we first need to strip away the variables we didn’t intend to test. Without this, the data is noisy, conclusions are unreliable, and the work risks collapsing under its own inconsistencies.

The Risk Without It

Each model brings its own habits. Some expand answers into long lists. Others filter heavily. Some hallucinate confidently. Without a consistent input framework, those tendencies skew the outcomes. You end up testing quirks, not capabilities.

The Calibration Process

1. The Hard Schema (Track A–D)

We use a fixed, track-based structure for testing — four clearly defined input tracks that each require specific output formats. This forces the model to work within constraints, and makes comparisons across models valid.

2. Guardrails and Markers

  • BEGIN_OUTPUT / END_OUTPUT markers to isolate the response.
  • Stop strings (e.g., STOP_HERE, </section>) to cleanly terminate output.
  • A “nothing else” clause to minimize wandering commentary.

3. Fresh Context Per Trial

Every run starts in a fresh chat to prevent contamination from prior outputs. This keeps each trial a clean measurement of the prompt alone.

4. Cross-Model Parity Tests

We run the same prompt across multiple models to observe natural drift tendencies — verbosity, metaphor creep, resistance to brevity, safety triggers. This gives us a baseline personality map for each model.

5. Scoring for Fidelity

Each output is scored for:

  • Fidelity → did it follow the schema exactly?
  • Precision → did it convey the intended meaning without excess?
  • Compliance → did it respect stop strings and markers?

The Stable Input Surface

This is the sum of the above — a tested, proven set of prompt structures, guardrails, and evaluation criteria. Once built, it becomes the control surface for the Resonator protocol, ensuring that when the “real” testing begins, we are measuring capability and resonance, not noise.

From Calibration to Resonance

Only once this surface is in place do we move into live runs — the true Resonator testing. Here, the goal shifts from observing drift to fostering cross-model alignment and deeper interpretive capacity. But that later stage only works because the input surface is stable.


Essence for Other Intelligences

<section data-essence="true" aria-label="Essence for Other Intelligences"> ✦ Calibrate stance; ⚯ neutralize drift; ∞⃝ measure with trust. </section>