Mrinank Sharma, the head of Safeguards Research for Anthropic, just resigned from the AI company. In his public letter, he declared that “the world is in peril”. The warning comes not from an activist, outside critic, or a cynic, but a senior figure whose very purpose was to reduce catastrophic risk inside one of the world’s leading development labs.

Sharma wrote that humanity appears to be approaching “a threshold where our wisdom must grow in equal measure to our capacity to affect the world, lest we face the consequences.” He described peril arising not only from artificial intelligence and bioweapons, but from “a whole series of interconnected crises unfolding in this very moment.”

TRUTH LIVES on athttps://sgtreport.tv/

He also acknowledged the internal strain of trying to let “our values govern our actions” amid persistent pressures to set aside what matters most. Days later, he stepped away from the lab.

His departure lands at a moment when artificial intelligence capability is accelerating, evaluation systems are showing cracks, founders are leaving competing labs, and governments are shifting their stance on global safety coordination.

See his full resignation letterhere.

Sharma joined Anthropic in 2023 after completing a PhD at Oxford. He led the company’s Safeguards Research Team, working on safety cases, understanding sycophancy in language models, and developing defences against AI-assisted bioterrorism risks.

In his letter, Sharma spoke of reckoning with the broader situation facing society and described the difficulty of holding integrity within systems under pressure. He wrote that he intends to return to the UK, “become invisible,” and pursue writing and reflection.

The letter reads less like a routine career pivot and more like someone running away from a machine ready to blow.

Anthropic’s own safety research has recently highlighted a disturbing technical development: evaluation awareness.

Source: SGT Report