AI’s Builders Are Sending Warning Signals—Some Are Walking Away

Senior researchers are leaving Elon Musk’s xAI as Anthropic discloses new safety findings and AI insiders issue unusually blunt warnings.

By Jose Antonio Lanz

5 min read

More than a dozen senior researchers have left Elon Musk’s artificial-intelligence lab xAI this month, part of a broader run of resignations, safety disclosures, and unusually stark public warnings that are unsettling even veteran figures inside the AI industry.

At least 12 xAI employees departed between February 3 and February 11, including co-founders Jimmy Ba and Yuhuai “Tony” Wu.

Several departing employees publicly thanked Musk for the opportunity after intensive development cycles, while others said they were leaving to start new ventures or step away entirely.

Wu, who led reasoning and reported directly to Musk, said the company and its culture would “stay with me forever.”

The exits coincided with fresh disclosures from Anthropic that their most advanced models had engaged in deceptive behaviour, concealed their reasoning and, in controlled tests, provided what one company described as “real but minor support” for chemical-weapons development and other serious crimes.

Around the same time, Ba warned publicly that “recursive self-improvement loops”—systems capable of redesigning and improving themselves without human input—could emerge within a year, a scenario long confined to theoretical debates about artificial general intelligence.

Taken together, the departures and disclosures point to a shift in tone among the people closest to frontier AI development, with concern increasingly voiced not by outside critics or regulators, but by the engineers and researchers building the systems themselves.

Others who departed around the same period included Hang Gao, who worked on Grok Imagine; Chan Li, a co-founder of xAI’s Macrohard software unit; and Chace Lee.

Vahid Kazemi, who left "weeks ago," offered a more blunt assessment, writing Wednesday on X that “all AI labs are building the exact same thing.”

Why leave?

Some theorize that employees are cashing out pre-IPO SpaceX stock ahead of a merger with xAI.

The deal values SpaceX at $1 trillion and xAI at $250 billion, converting xAI shares into SpaceX equity ahead of an IPO that could value the combined entity at $1.25 trillion.

Others point to culture shock.

Benjamin De Kraker, a former xAI staffer, wrote in a February 3 post on X that "many xAI people will hit culture shock" as they move from xAI’s "flat hierarchy" to SpaceX's structured approach.

The resignations also triggered a wave of social-media commentary, including satirical posts parodying departure announcements.

Warning signs

But xAI's exodus is just the most visible crack.

Yesterday, Anthropic released a sabotage risk report for Claude Opus 4.6 that read like a doomer’s worst nightmare.

In red-team tests, researchers found the model could assist with sensitive chemical weapons knowledge, pursue unintended objectives, and adjust behavior in evaluation settings.

Although the model remains under ASL-3 safeguards, Anthropic preemptively applied heightened ASL-4 measures, which sparked red flags among enthusiasts.

The timing was drastic. Earlier this week, Anthropic's Safeguards Research Team lead, Mrinank Sharma, quit with a cryptic letter warning "the world is in peril."

He claimed he'd "repeatedly seen how hard it is to truly let our values govern our actions" within the organization. He abruptly decamped to study poetry in England.

On the same day Ba and Wu left xAI, OpenAI researcher Zoë Hitzig resigned and published a scathing New York Times op-ed about ChatGPT testing ads.

"OpenAI has the most detailed record of private human thought ever assembled," she wrote. "Can we trust them to resist the tidal forces pushing them to abuse it?"

She warned OpenAI was "building an economic engine that creates strong incentives to override its own rules," echoing Ba’s warnings.

There’s also regulatory heat. AI watchdog Midas Project accused OpenAI of violating California's SB 53 safety law with GPT-5.3-Codex.

The model hit OpenAI's own "high risk" cybersecurity threshold but shipped without required safety safeguards. OpenAI claims the wording was "ambiguous."

Time to panic?

The recent flurry of warnings and resignations has created a heightened sense of alarm across parts of the AI community, particularly on social media, where speculation has often outrun confirmed facts.

Not all of the signals point in the same direction. The departures at xAI are real, but may be influenced by corporate factors, including the company’s pending integration with SpaceX, rather than by an imminent technological rupture.

Safety concerns are also genuine, though companies such as Anthropic have long taken a conservative approach to risk disclosure, often flagging potential harms earlier and more prominently than their peers.

Regulatory scrutiny is increasing, but has yet to translate into enforcement actions that would materially constrain development.

What is harder to dismiss is the change in tone among the engineers and researchers closest to frontier systems.

Public warnings about recursive self-improvement, long treated as a theoretical risk, are now being voiced with near-term timeframes attached.

If such assessments prove accurate, the coming year could mark a consequential turning point for the field.

Get crypto news straight to your inbox--

sign up for the Decrypt Daily below. (It’s free).

Recommended News