The First to Submit Writes the Rules

Mythos reads less like a safety warning heeded and more like a first-mover play on CAISI. The labs that submitted first are writing the rules.

7 min read

The administration that spent its first year telling frontier labs to ship faster and stop apologizing has, over the last month, quietly pivoted to mandatory pre-release vetting. Anthropic, Microsoft, Google, xAI, and OpenAI are now running their next-generation models through CAISI evaluation before public release, per the joint announcements in late April and early May 2026. Two weeks earlier, "remove red tape" was the operative phrase in every public statement the White House made about AI policy. Two weeks later, the red tape is the policy.

The interpretation the press has settled on is that Anthropic's Mythos announcement, the model the company declared too dangerous to release while simultaneously refusing to remove DoD guardrails, scared the government into action. That reading is not wrong, but it is shallow. It treats the regulatory shift as a safety reaction rather than as the outcome that a specific move was designed to produce. If you look at who moves next, and what the framework they are moving into looks like, the Mythos announcement reads less like a warning shot and more like a first-mover play on the regulatory layer itself.

A brief timeline, because the compression is the story. CAISI ran roughly 40 model evaluations between its founding in late 2024 and April 2026, most of them after-the-fact and voluntary, with no public enforcement teeth. In late April, Anthropic announced the Mythos model with the framing that it was too dangerous to release in its current form, and declined to remove the guardrails the Department of Defense had requested stripped, citing the same evaluation framework. Over the following weeks, per their own public statements, Microsoft, Google, xAI, and OpenAI all committed their next frontier releases to pre-release CAISI vetting. OpenAI, which had spent most of 2025 positioning as the industry's unbound competitor, joined last.

Read that sequence as a coordination problem and a different shape emerges. Pre-release vetting is something every safety researcher, regulator, and rival lab has wanted for three years. None of them could produce it, because the first lab to submit unilaterally would eat a release delay its competitors did not, and the fast mover would capture the market. A voluntary mandate requires either a regulator with enforcement power, which the United States did not have for frontier AI, or a public event that makes continued non-compliance politically impossible for the remaining holdouts. Mythos was that event. The "too dangerous to release" frame did not just alarm the White House. It reframed the baseline. Once Anthropic had publicly treated pre-release vetting as the responsible default, every other major lab faced a binary choice: join the framework, or be the outlier explaining to Congress why you did not.

This is what a Schelling point looks like when it hardens into policy. The equilibrium could not be reached by incremental pressure. It needed a focal event that made one behavior the obvious default and made deviation visible. Anthropic supplied it. Whether the move was long-planned or opportunistic is a question I cannot answer, but the structural effect is the same either way. Within a few weeks, a framework that had been unreachable for three years was stable because four competitors had already joined.

Anthropic had been releasing models through a restrictive process for most of its existence: long pre-release evaluation, heavy guardrail work, constitutional AI layering, deliberate capability gating. Their competitors, at least until Mythos, did not carry equivalent overhead on public release timelines. The new CAISI framework transfers that cost across the industry, making the Anthropic-style process the floor rather than the ceiling. A company whose internal safety process was a market disadvantage just converted it into a regulatory one. Microsoft and Google are similarly positioned: both already run extensive internal red-teaming, both have the compliance apparatus to absorb a new external evaluation layer without restructuring. The companies that get squeezed are the ones whose speed advantage came from leaner safety overhead. xAI. The Chinese labs, to the extent U.S. framework adoption propagates to export controls. Whoever was planning to compete on "we ship faster because we test less."

The framework itself is being written right now, by the people who submitted first. CAISI's evaluation criteria, as of early May, are not finalized; the methodology is being calibrated against the models already in flight through the pipeline. The labs that joined first are the ones whose architectural choices, safety techniques, and internal evaluation formats are setting the template for what "adequately vetted" will mean when the framework becomes mandatory. A company whose internal process shaped the external standard faces a very different compliance cost from a company that has to retrofit its pipeline to match someone else's template. This is regulatory capture on a twelve-week clock.

The press framing, that Mythos triggered a safety reaction, treats Anthropic as an actor whose warning was heeded. The commercial framing treats the same facts as a first-mover play on a pre-release vetting framework that will likely be mandatory within the year and whose contours are being written by the companies who submitted first. Both framings fit the facts. Only one of them explains why OpenAI, which spent most of 2025 arguing against exactly this kind of framework, joined at all.

For anyone building on top of frontier models, the consequence is a supply chain question. CAISI-vetted models will carry different release timelines, different deployment restrictions, and almost certainly a different pricing shape once the compliance cost is internalized. Tools, agents, and infrastructure built against the current API surface will need to absorb whatever release cadence the new framework produces, and the cadence is going to be slower. If the three-month gap between model generations that defined 2024 and 2025 stretches to five or six months because of pre-release evaluation, the product roadmaps of every company that assumed an eighteen-month model rollover will need to be redrawn.

There is also a new category of risk worth naming directly. Mythos exists, the government knows its capabilities, the frontier labs know its capabilities, and the open-source ecosystem does not and cannot match it. The asymmetry is not between the United States and other states. It is between the labs that can build it and the public, and between the labs that get CAISI clearance to deploy variants and the public that only sees the deployed versions. This is the quiet implication of a pre-release vetting regime that most of the policy discussion has not caught up to yet. Capability information is now asymmetrically distributed by design, with the safety-conscious framing providing the political cover.

What I would watch over the next quarter. Whether CAISI publishes the evaluation criteria in full or keeps the methodology internal to the labs in the pipeline. Whether the Anthropic-DoD guardrails dispute gets resolved through the new framework or ends up in court, because the outcome sets the precedent for how much vendor discretion survives the evaluation process. Whether legislation codifies what is currently voluntary, which moves the moat from a commercial advantage to a regulatory one. And whether any major lab declines to submit. The first declination will be more informative about the real compliance cost than any of the enthusiastic announcements we have seen so far.

The lab that submitted first did not just buy safety credibility. It bought the pen that writes the framework the rest of the industry will have to submit to. That moat is more durable than a model lead, because regulatory advantage decays slower than algorithmic advantage.

Pre-release vetting is the right answer to a real risk. It is also a first-mover play. Both things can be true, and the labs that recognize the second one are the ones who will write the next year of AI regulation.