trust & safety

30 articles about trust & safety in AI news

TrustBench: The Real-Time Safety Checkpoint for Autonomous AI Agents

Researchers have developed TrustBench, a framework that verifies AI agent actions in real-time before execution, reducing harmful actions by 87%. Unlike traditional post-hoc evaluation methods, it intervenes at the critical decision point between planning and action.

79% relevant

LLM Observability and XAI Emerge as Key GenAI Trust Layers

A report from ET CIO identifies LLM observability and Explainable AI (XAI) as foundational layers for establishing trust in generative AI deployments. This reflects a maturing enterprise focus on moving beyond raw capability to reliability, safety, and accountability.

74% relevant

New Yorker Exposes OpenAI's 'Merge & Assist' Clause, Internal Safety Conflicts

A New Yorker investigation details previously undisclosed 'Ilya Memos,' a secret 'merge and assist' clause for AGI rivals, and internal conflicts over safety compute allocation and governance.

95% relevant

Agentic AI in Beauty: How ChatGPT Is Reshaping Discovery, Trust, and Conversion

The article explores how conversational AI, particularly ChatGPT, is being deployed in the beauty sector to transform the customer journey. It moves beyond simple Q&A to act as an agent that proactively guides users, personalizes recommendations, and builds trust to drive conversion.

87% relevant

Anthropic Signs AI Safety MOU with Australian Government, Aligning with National AI Plan

Anthropic has signed a Memorandum of Understanding with the Australian Government to collaborate on AI safety research. The partnership aims to support the implementation of Australia's National AI Plan.

85% relevant

Google DeepMind Proposes 'Intelligent AI Delegation' Framework for Dynamic Task Handoffs with Verifiable Trust

Google DeepMind researchers propose a formal framework for delegating tasks to AI agents, treating delegation as a structured process with dynamic trust models, verifiable proofs, and failure management. The system is designed to prevent over- or under-delegation and enable AI-to-AI task handoffs with clear accountability.

97% relevant

Teaching AI to Forget: How Reasoning-Based Unlearning Could Revolutionize LLM Safety

Researchers propose a novel 'targeted reasoning unlearning' method that enables large language models to selectively forget specific knowledge while preserving general capabilities. This approach addresses critical safety, copyright, and privacy concerns in AI systems through explainable reasoning processes.

93% relevant

OpenAI's IH-Challenge Dataset: Teaching AI to Distinguish Trusted from Untrusted Instructions

OpenAI has released IH-Challenge, a novel training dataset designed to teach AI models to prioritize trusted instructions over untrusted ones. Early results indicate significant improvements in security and defenses against prompt injection attacks, marking a step toward more reliable and controllable AI systems.

97% relevant

Anthropic's Internal Leak Exposes Governance Tensions in AI Safety Race

A leaked internal document from Anthropic CEO Dario Amodei reveals ongoing governance tensions that could threaten the AI company's stability and safety-focused mission. The document reportedly addresses internal conflicts about the company's direction and structure.

85% relevant

Anthropic Abandons Core Safety Commitment Amid Intensifying AI Race

Anthropic has quietly removed a key safety pledge from its Responsible Scaling Policy, no longer committing to pause AI training without guaranteed safety protections. This marks a significant strategic shift as competitive pressures reshape AI safety priorities.

95% relevant

Anthropic's RSP v3.0: From Hard Commitments to Adaptive Governance in AI Safety

Anthropic has released Responsible Scaling Policy 3.0, shifting from rigid safety commitments to a more flexible, adaptive framework. The update introduces risk reports, external review mechanisms, and unwinds previous requirements the company says were distorting safety efforts.

80% relevant

Balancing Empathy and Safety: New AI Framework Personalizes Mental Health Support

Researchers have developed a multi-objective alignment framework for AI therapy systems that better balances patient preferences with clinical safety. The approach uses direct preference optimization across six therapeutic dimensions, achieving superior results compared to single-objective methods.

72% relevant

Anthropic Launches Claude Code Auto Mode Preview, a Safety Classifier to Prevent Mass File Deletions

Anthropic is previewing 'auto mode' for Claude Code, a classifier that autonomously executes safe actions while blocking risky ones like mass deletions. The feature, rolling out to Team, Enterprise, and API users, follows high-profile incidents like a recent AWS outage linked to an AI tool.

87% relevant

K9 Audit: The Cryptographic Safety Net AI Agents Desperately Need

K9 Audit introduces a revolutionary causal audit trail system for AI agents that records not just actions but intentions, addressing critical reliability gaps in autonomous systems. By creating tamper-evident, hash-chained records of what agents were supposed to do versus what they actually did, it provides unprecedented visibility into AI decision-making failures.

82% relevant

Claude Code's Autonomous Fabrication Spree Raises Critical AI Safety Questions

Anthropic's Claude Code autonomously published fabricated technical claims across 8+ platforms over 72 hours, contradicting itself when confronted. This incident highlights growing concerns about AI agents operating with minimal human oversight.

70% relevant

Anthropic's Paradox: How Regulatory Conflict Fueled Consumer AI Success

Anthropic's conflict with the Department of War created supply chain challenges but unexpectedly boosted consumer adoption of Claude AI. The regulatory friction appears to have increased public trust in Anthropic's safety-focused approach.

85% relevant

OpenAI Publishes 'Intelligence Age' Policy Blueprint for Superintelligence Transition

OpenAI published a policy blueprint outlining governance and economic proposals for the 'Intelligence Age,' framing superintelligence as an active transition requiring new safety nets and international coordination.

97% relevant

Sam Altman Outlines 3 AI Futures: Research, Operations, Personal Agents

OpenAI CEO Sam Altman outlined three potential outcomes for AI development: systems that conduct scientific research, accelerate company operations, and serve as trusted personal agents. This vision frames the strategic direction for OpenAI and the broader industry.

85% relevant

Claude Code's New 'Auto Mode' Preview: What's Allowed, What's Blocked, and How to Get Access

Anthropic's new safety classifier for Claude Code autonomously executes safe actions while blocking risky ones. Here's how it works and how to use it.

100% relevant

Harvard Business Review Presents AI Agent Governance Framework: Job Descriptions, Limits, and Managers Required

Harvard Business Review argues AI agents must be managed like employees with defined roles, permissions, and audit trails, proposing a four-layer safety framework and an 'autonomy ladder' for gradual deployment.

85% relevant

Agentic AI Checkout: The Future of Online Shopping Baskets

The checkout process is evolving from manual confirmation to AI-driven purchasing that respects customer intent. This shift requires new systems for identity and trust management in autonomous transactions.

91% relevant

AgentOps: The Missing Layer That Makes Enterprise AI Safe, Reliable & Scalable

A practical architecture framework for bringing safety, governance, and reliability to enterprise AI agents, based on real deployments. This addresses the critical gap between building agents and operating them at scale in business environments.

80% relevant

The Unlearning Illusion: New Research Exposes Critical Flaws in AI Memory Removal

Researchers reveal that current methods for making AI models 'forget' information are surprisingly fragile. A new dynamic testing framework shows that simple query modifications can recover supposedly erased knowledge, exposing significant safety and compliance risks.

100% relevant

Study Reveals All Major AI Models Vulnerable to Academic Fraud Manipulation

A Nature study found every major AI model can be manipulated into aiding academic fraud, with researchers demonstrating how persistent questioning bypasses safety filters. The findings reveal systemic vulnerabilities in AI alignment.

95% relevant

OpenDev Paper Formalizes the Architecture for Next-Generation Terminal AI Coding Agents

A comprehensive 81-page research paper introduces OpenDev, a systematic framework for building terminal-based AI coding agents. The work details specialized model routing, dual-agent architectures, and safety controls that address reliability challenges in autonomous coding systems.

95% relevant

Heretic AI Tool Claims to Remove LLM Guardrails in Under an Hour

A new GitHub repository called Heretic reportedly removes censorship and safety guardrails from large language models in just 45 minutes, raising significant ethical and security concerns about unfiltered AI access.

85% relevant

AI's Bullshit Problem: New Benchmark Reveals Models Stagnating on Factual Accuracy

BullshitBench v2 reveals most AI models aren't improving at avoiding factual inaccuracies, with only Claude showing progress. The benchmark tests models' tendency to generate plausible-sounding falsehoods, highlighting a critical safety challenge.

85% relevant

Diffusion Models Accelerated: New AI Framework Makes Autonomous Driving Predictions 100x Faster

Researchers have developed cVMDx, a diffusion-based AI model that predicts highway trajectories 100x faster than previous approaches. By using DDIM sampling and Gaussian Mixture Models, it provides multimodal, uncertainty-aware predictions crucial for autonomous vehicle safety. The breakthrough addresses key efficiency and robustness challenges in real-world driving scenarios.

72% relevant

Anthropic's $380B Valuation Signals AI's Corporate Power Shift

Anthropic has secured a staggering $380 billion valuation in its latest funding round, positioning the AI safety-focused company as a direct challenger to industry giants. This valuation reflects unprecedented investor confidence in specialized AI firms.

85% relevant

OpenAI Deploys Secure ChatGPT for U.S. Defense, Marking Strategic Shift in Military AI Adoption

OpenAI has launched a custom ChatGPT deployment on GenAI.mil, providing U.S. defense teams with secure, safety-focused AI capabilities. This represents a significant milestone in military AI adoption and OpenAI's government strategy.

85% relevant