
Mrinank Sharma, who led Anthropic’s Safeguards Research Team, has resigned from the AI company with a cryptic letter hinting at internal tensions over AI safety and broader concerns about global crises.
Key Details of the Resignation
Sharma, who joined Anthropic in 2023 and led its Safeguards Research Team since early last year, announced his departure in a vague but concerning letter shared with colleagues. During his tenure, he worked on addressing AI sycophancy, developed defenses against potential bioterrorism facilitated by AI, and authored one of the company’s first AI safety cases.
His resignation letter, while lacking specific details, suggests internal conflicts regarding the company’s values and priorities. “Throughout my time here, I’ve repeatedly seen how hard it is to truly let our values govern our actions,” Sharma stated, adding that employees “constantly face pressures to set aside what matters most.”
Broader Warnings and Context
Sharma’s letter contains ominous warnings about global conditions, stating that “the world is in peril” not only from artificial intelligence and bioweapons but from “a whole series of interconnected crises.” He suggests humanity is approaching a threshold where wisdom must match our technological capabilities.
The resignation comes amid significant market reactions to Anthropic’s release of Claude Cowork, which triggered stock market declines over concerns about potential job automation, particularly in white-collar sectors like legal services.
Internal Concerns at Anthropic
According to reporting by The Telegraph, Anthropic employees have expressed private concerns about their own technology’s potential impact on the labor market. Internal survey responses revealed anxieties such as “It kind of feels like I’m coming to work every day to put myself out of a job” and “In the long term, I think AI will end up doing everything and make me and many others irrelevant.”
Industry Pattern of Safety-Related Resignations
Sharma’s departure follows a pattern in the AI industry where high-profile resignations over safety concerns have become increasingly common. A notable example includes a former member of OpenAI’s disbanded “Superalignment” team who quit after concluding the company was prioritizing product releases over user safety.
Significance and Implications
This resignation highlights ongoing tensions within AI companies between rapid technological advancement and responsible development. It also underscores the growing concerns among AI researchers and developers about the broader societal implications of the technologies they’re creating.
While Sharma’s letter is notably self-exonerating—a common characteristic of such public resignations—it adds to mounting evidence of internal conflicts within leading AI organizations about the direction and pace of AI development.


GIPHY App Key not set. Please check settings