Saturday, 29 November 2025
30.9 C
Singapore
29.9 C
Thailand
27.5 C
Indonesia
28.6 C
Philippines

Anthropic updates AI rules to address rising safety concerns

Anthropic updates Claude AI rules with stricter bans on weapons and cybersecurity misuse while easing restrictions on political content.

Anthropic has introduced new rules for the use of its Claude AI chatbot, aiming to address growing fears about misuse in an increasingly dangerous digital landscape. The updated policy outlines stricter cybersecurity requirements and places clear restrictions on the development of some of the world’s most dangerous weapons.

Expanded restrictions on weapons development

Previously, Anthropic banned the use of Claude to “produce, modify, design, market, or distribute weapons, explosives, dangerous materials or other systems designed to cause harm to or loss of human life.” The new version strengthens this rule by naming specific categories of weapons. Users are now prohibited from developing high-yield explosives, as well as biological, nuclear, chemical, and radiological (CBRN) weapons, with the assistance of Claude.

The move follows the company’s introduction of “AI Safety Level 3” protections in May, alongside its Claude Opus 4 model. These measures are designed to make the system more resistant to jailbreak attempts and to block attempts at using the technology in the design or creation of CBRN weapons.

Addressing cybersecurity and agentic AI risks

Anthropic also highlighted concerns about the risks posed by more advanced and autonomous AI tools. These include “Computer Use”, a feature that allows Claude to control a user’s computer directly, and “Claude Code”, which integrates the system into a developer’s terminal.

“These powerful capabilities introduce new risks, including potential for scaled abuse, malware creation, and cyber attacks,” the company wrote in its policy update.

In response, a new section has been added to the rules, titled “Do Not Compromise Computer or Network Systems.” This prohibits users from employing Claude to identify or exploit security vulnerabilities, create or distribute malware, or develop tools for denial-of-service attacks. By expanding its cybersecurity policy, Anthropic is aiming to minimise risks linked to hacking, fraud, and large-scale digital abuse.

Adjustments to political content rules

While restrictions have tightened in some areas, Anthropic has also eased its stance on political content. Previously, all campaign-related and lobbying content was banned. Under the new guidelines, the company will only prohibit use cases that are “deceptive or disruptive to democratic processes, or involve voter and campaign targeting.”

Anthropic further clarified its rules for “high-risk” use cases. These requirements apply only when Claude is being used in consumer-facing scenarios, rather than internal business applications. This distinction aims to provide businesses with greater flexibility when deploying AI in professional settings, while still safeguarding individuals from potential harm.

The changes reflect Anthropic’s effort to strike a balance between innovation and safety as AI systems become increasingly powerful and widely available. By tightening rules around weapons and cybersecurity while refining its political guidelines, the company is aiming to prevent misuse without overly limiting legitimate uses of Claude.

Hot this week

Andika Rama returns to claim TGR Asia Esports GT Championship 2025 title

Indonesia’s Andika Rama wins the TGR Asia Esports GT Championship 2025 as his team seals both individual and country titles.

Sumsub reports sharp rise in synthetic personal data fraud in APAC

Sumsub reports a sharp rise in synthetic identity fraud and deepfake attacks across APAC as AI-driven scams become more sophisticated.

Crunchyroll brings world-first premieres and major anime showcases to AFA Singapore 2025

Crunchyroll brings exclusive premieres, guest panels and a large interactive booth to AFA Singapore 2025.

Valve offers strongest hint yet on expected Steam Machine pricing

Valve hints that the Steam Machine may be priced close to a similarly powerful DIY PC, but external factors keep final costs uncertain.

Southeast Asia’s Agnes AI partners with Agora to launch real-time AI workspace

Agnes AI and Agora launch a real-time AI workspace that connects human teams and AI agents for collaborative work at scale.

SMRT upgrades Bishan Depot with automation to double train overhaul capacity

SMRT upgrades Bishan Depot with automation to double overhaul capacity and enhance safety, efficiency, and workforce sustainability.

Apple is expected to overtake Samsung as the world’s leading smartphone maker

Apple is projected to overtake Samsung as the world’s top smartphone maker, driven by strong iPhone 17 demand and upcoming device launches.

Singapore orders Apple and Google to stop spoofed government identities on messaging apps

Singapore orders Apple and Google to block spoofed government identities on messaging apps to curb rising impersonation scams.

Nintendo acquires Bandai Namco Studios Singapore

Nintendo acquires Bandai Namco Studios Singapore to boost game development and expand its subsidiary network.

Related Articles

Popular Categories