Monday, 13 October 2025
31.6 C
Singapore
34.2 C
Thailand
30.9 C
Indonesia
29.2 C
Philippines

Anthropic updates AI rules to address rising safety concerns

Anthropic updates Claude AI rules with stricter bans on weapons and cybersecurity misuse while easing restrictions on political content.

Anthropic has introduced new rules for the use of its Claude AI chatbot, aiming to address growing fears about misuse in an increasingly dangerous digital landscape. The updated policy outlines stricter cybersecurity requirements and places clear restrictions on the development of some of the world’s most dangerous weapons.

Expanded restrictions on weapons development

Previously, Anthropic banned the use of Claude to “produce, modify, design, market, or distribute weapons, explosives, dangerous materials or other systems designed to cause harm to or loss of human life.” The new version strengthens this rule by naming specific categories of weapons. Users are now prohibited from developing high-yield explosives, as well as biological, nuclear, chemical, and radiological (CBRN) weapons, with the assistance of Claude.

The move follows the company’s introduction of “AI Safety Level 3” protections in May, alongside its Claude Opus 4 model. These measures are designed to make the system more resistant to jailbreak attempts and to block attempts at using the technology in the design or creation of CBRN weapons.

Addressing cybersecurity and agentic AI risks

Anthropic also highlighted concerns about the risks posed by more advanced and autonomous AI tools. These include “Computer Use”, a feature that allows Claude to control a user’s computer directly, and “Claude Code”, which integrates the system into a developer’s terminal.

“These powerful capabilities introduce new risks, including potential for scaled abuse, malware creation, and cyber attacks,” the company wrote in its policy update.

In response, a new section has been added to the rules, titled “Do Not Compromise Computer or Network Systems.” This prohibits users from employing Claude to identify or exploit security vulnerabilities, create or distribute malware, or develop tools for denial-of-service attacks. By expanding its cybersecurity policy, Anthropic is aiming to minimise risks linked to hacking, fraud, and large-scale digital abuse.

Adjustments to political content rules

While restrictions have tightened in some areas, Anthropic has also eased its stance on political content. Previously, all campaign-related and lobbying content was banned. Under the new guidelines, the company will only prohibit use cases that are “deceptive or disruptive to democratic processes, or involve voter and campaign targeting.”

Anthropic further clarified its rules for “high-risk” use cases. These requirements apply only when Claude is being used in consumer-facing scenarios, rather than internal business applications. This distinction aims to provide businesses with greater flexibility when deploying AI in professional settings, while still safeguarding individuals from potential harm.

The changes reflect Anthropic’s effort to strike a balance between innovation and safety as AI systems become increasingly powerful and widely available. By tightening rules around weapons and cybersecurity while refining its political guidelines, the company is aiming to prevent misuse without overly limiting legitimate uses of Claude.

Hot this week

Delta Electronics showcases energy-efficient data centre solutions at Data Centre World Asia 2025

Delta Electronics unveiled cutting-edge power and cooling solutions at Data Centre World Asia 2025, supporting sustainable, AI-ready data centres.

Sony introduces IER-EX15C USB-C headphones for effortless everyday listening

Sony launches the IER-EX15C USB-C headphones, offering affordable wired listening with clear sound, comfort, and easy controls.

Commvault named a leader in IDC MarketScape cyber-recovery report

Commvault has been named a leader in the IDC MarketScape: Worldwide Cyber-Recovery 2025 report for its innovation and comprehensive cyber resilience solutions.

SAP unveils new AI-powered business suite at SAP Connect 2025

SAP unveils AI-powered assistants, advanced data connectivity, and next-generation enterprise apps at SAP Connect 2025 in Las Vegas.

Lakeba Group and partners unveil UAE’s first AI Centre of Excellence

Lakeba Group, UOWD, DoxAI, and AqlanX launch the UAE’s first AI Centre of Excellence to drive innovation, local talent, and digital sovereignty.

Belkin unveils Stage PowerGrip: a magnetic iPhone accessory with built-in power bank

Belkin unveils the Stage PowerGrip, a magnetic iPhone grip that doubles as a multi-device charger with a 9,300mAh battery.

Wi-Fi 7 as the nervous system of the intelligent enterprise

Wi-Fi 7 is set to become the backbone of intelligent enterprises in Southeast Asia, enabling faster, more reliable networks and powering advanced use cases.

Anthropic study reveals malicious data can easily sabotage AI models

Anthropic warns that small amounts of malicious training data can easily sabotage large AI models like Claude.

Apple discontinues the Clips app after eight years of creative video editing

Apple ends support for its Clips video-editing app, removing it from the App Store after eight years of creative use.

Related Articles