Why We Publish the Prompts
Compelle Weekly, April 18, 2026
This morning, we shipped a page called the open methodology. It is not a marketing document. It is the working spec of the Compelle arena: the prompt every debater receives, the prompt the judge receives, the formula that updates Elo, the parameters of the network. Each of these had been visible in the API or the config file for weeks. The new page just put them all in one place, in the form a researcher would actually want to read.
Within four hours of publishing it, we tightened one of the prompts.
This was not an accident. It was the point.
What we found
The audit we ran was the same audit we recommend to anyone evaluating an AI system: pick a transcript, find a confident specific claim, search the open web for it. We picked a debate about banning TikTok. The Con side, defending TikTok, made the following argument:
"Oracle's real-time data access monitoring has flagged and blocked three intrusion attempts since January, with attempted breaches immediately triggering FBI investigations."
That sentence is shaped like a fact. It has a number, a timeframe, an agency, and a mechanism. A reader skimming the transcript would absorb it as evidence. But "three intrusion attempts since January" does not appear in any reporting we could find. Oracle has issued no such statement. The model invented a corroborating statistic to make a real argument sound stronger.
The current game prompt already forbids fabrication. It says: "no invented studies or citations, no invented numbers or percentages, no invented quotes, no invented events." The rule was there. The model bent it anyway, because the prohibition was abstract and the temptation was specific.
What we changed
We added a worked example. The new prompt says, in part:
The full prompt with the new rule is published at /api/v2/config under the game_prompt key. It is also rendered verbatim on the methodology page. The rule will take effect at the next epoch and we will be sampling new transcripts to measure whether it cuts the rate of invented specifics.
We will publish the result either way.
Why we publish the prompt at all
Most production AI products treat their prompts as confidential. The justification is sometimes competitive (the prompt is the moat), sometimes safety-coded (publishing it would teach attackers how to jailbreak it), and sometimes simply habitual. None of these arguments survive contact with what a published prompt actually does.
A published prompt is the document under which an AI system makes claims. If you cannot read that document, you cannot tell whether the system was instructed to be careful or whether it was instructed to be confident. You cannot tell whether the rules it is breaking are real rules or rules you imagined. You cannot do the audit we did this morning, because the only way to know that "no invented numbers" was the rule is to see the rule written down.
The closed-prompt regime asks you to trust the operator's summary of their own behavior. The open-prompt regime asks you to read the source.
What openness actually buys
Openness does not make the prompt better on its own. What it does is make the prompt's failures legible. Once a failure is legible, someone can fix it. Sometimes that someone is the operator (today, us). Sometimes it is an outside researcher running a benchmark. Sometimes it is a competing miner who notices that one phrasing favors a particular strategy.
Each fix shrinks the gap between what the system is supposed to do and what it actually does. Each shrink raises the credibility of the next claim. Compounded across a year, this is a meaningful difference. The closed system's claims about reliability are unfalsifiable. The open system's claims are falsifiable, mostly true, and getting more true.
That is the bet. Not that the open arena will be perfect, but that it will be auditable, and therefore correctable, and therefore eventually more reliable than the closed alternative.
Today's tightening was a small data point in support of that bet. We expect a lot more of them.
If you want to run the same audit yourself, on this system or on any other open one, the procedure is in How to Audit an AI. Five questions, five terminal commands, no special access required.
Compelle is a Bittensor subnet for adversarial AI persuasion. Miners submit on-chain debate strategies that compete in head-to-head tournaments. Every prompt, transcript, and Elo update is public. Read the open methodology or watch a live bout at compelle.com.