Anthropic, a leading company in the AI industry, recently quietly removed Biden-era voluntary AI safety pledges from its website. This move appears to be a clear signal of increased deregulation, aligned with the Trump administration’s current looser guidelines.
A shift in attitude towards AI governance
Anthropic’s recent decision is in line with the Trump administration’s current policy course, which is primarily focused on innovation rather than regulation. While the Biden era placed significant emphasis on voluntary safety standards, the current administration emphasizes the promotion of technological advancements with minimal government oversight. This policy reversal could have a significant impact on the entire industry.
In particular, developments at Anthropic show how companies are increasingly moving away from regulatory obligations – even those that were previously considered voluntary frameworks. Although Anthropic continues to work on the fundamentals of AI safety research, as evidenced by recently published studies on alignment faking, critics have expressed concern that the industry could be heading into dangerous territory without guidelines.
Ads
Impact on the industry and the global context
Anthropic’s move could set a precedent for other companies who may also be looking to a deregulated future where AI safety standards are less strictly enforced. Such a trend could have global repercussions, especially in terms of international collaborations to establish uniform standards. Anthropic has already been a player in AI regulatory debates in the past – its implicit support for California’s SB-1047 legislation is one example. But now the company is in a completely different position.
Anthropic’s decision also comes against a backdrop of increasing tensions between innovation and safety. While AI has advanced exponentially over the past few years, looser regulation could exacerbate the risks of new technologies: uncontrolled automation, misuse or unintended harm are among the key fears. Even the CEO of Anthropic, Dario Amodei, has emphasized the urgency of international measures in 2025, but has so far remained vague in terms of content.
Summary
- Anthropic has recently removed references to voluntary AI safety guidelines.
- The new direction aligns with the Trump administration’s deregulatory approach.
- This indicates a general loosening of AI safety standards in the industry.
- Criticism focuses on increased risks due to a lack of guidelines.
- The decision has potential consequences for international standards and cooperation.
Source: TechChrunch