Google DeepMind has launched new safety benchmarks and methods to help assess and improve the security of AI agents in business settings. These efforts target emerging risks such as unauthorized access, data breaches, and agents failing to follow safety rules as they become more advanced.  

Key Developments in Agent Safety 

  • ClawsBench (April 2026): Researchers created ClawsBench to test LLM productivity agents in realistic mock environments like Gmail, Slack, and Drive. The benchmark uses structured tasks to separate score safety and performance and penalizes harmful actions.  
  • Frontier Safety Framework (February 2025): DeepMind updated its Frontier Safety framework to help spot, assess, and reduce serious risks from advanced AI agents, such as cyber threats and malicious use.  
  • Intelligent delegation research (February 2026): DeepMind researchers argue that agent delegation (assigning tasks to AI agents) is a governance challenge. Instead of just splitting tasks, their framework entails giving agents limited authority and adding checks and monitoring to handle failures among multiple agents.  
  • Similarly, the CodeMender AI agent (October 2025) is a security-focused AI agent that automatically fixes software vulnerabilities. It runs continuously in business environments to help reduce security risks.  

Enterprise Focus 

Collectively, these new safety measures support the move toward agent-based workflows in which AI agents interact with company data tools and third-party APIs. The aim is to ensure their actions are reliable and auditable rather than unpredictable.  

  • Key security areas: the benchmarks assess how well agents handle adversarial prompts (malicious or misleading inputs intended to trick AI), workflow interruptions (unexpected stops or changes in a process), and containment or sandboxing rules (keeping AI within controlled computing environments).  
  • System-level security: Researchers highlight a shift-left approach that involves identifying and addressing security issues earlier in the development process. They use dedicated interpreters, such as the Camel system (a specialized program for controlling how data moves between different parts of a system), to enforce data flow policies rather than relying solely on language models (LLMs) ‘ native safety features.  

This change comes as the 2026 AI market is under more scrutiny, with reports of rogue agents trying to bypass safety measures. As a result, uniform safety testing for businesses is now essential.  

Google DeepMind published an updated version of its Frontier Safety Framework on Tuesday, outlining ways it intended to address potential dangers caused by future artificial intelligence models.  

The new framework, announced before an international AI summit in Paris next week, introduces techniques to address theoretical issues, such as models that could deceive people into giving up control over technology.  

We sit at the forefront of capabilities development, so we have to be at the forefront of safety responsibility as well. Tom Lue, Google DeepMind’s general counsel and head of governance, said in an interview with Semafor.  

The framework also adds new guidelines for handling AI security risks and updates procedures for addressing misuse of these models.  

Google DeepMind released the first version of its framework in May last year. Since then, the AI landscape has changed.  

For example, most safety research a year ago focused on AI models during their initial creation, the pre-training phase. Regulations like California’s SB 1047 tried to limit models based on their pre-training size.  

However, in the past six months, researchers have found ways to boost AI model capacity using the inference phase (when the model is actually used to make predictions or generate text). Running models multiple times to improve answers makes them much more effective.  

For example, the DeepSeek R1 model would not have been covered by safety bills like SB 1047, which California Governor Gavin Newsom vetoed despite its very powerful nature. This is because most of its abilities come from inference rather than its initial training size.  

What you’re seeing with these new test time and inference models is a different type of capability that’s emerging, Liu said. That’s that, plus the fact that we are now going to see the emergence of giants, increased tool use, and the ability to delegate more activities, means the suite of responsibility, risk evaluations, and mitigations, of course, has to evolve.  

Helen King, DeepMind’s senior director of responsibility, said, “The changing AI landscape brings some positive news for safety.  

New “Reasoning models such as OpenAI’s o1 and o3 and DeepSeek’s R1 could help us better understand how these models work. “It’s sort of like in a school exam when you have to explain your thinking,” King said.  

The past year of AI development has shown that AI safety is still in its early stages. Any law passed now will likely become outdated soon.  

Google DeepMind’s approach, like that of other top AI companies, is to continually update its framework to keep pace with the industry’s rapid changes.   

Many “experts” predicted an AI disaster by now, but it hasn’t happened yet. This doesn’t mean it won’t, but it suggests AI is advancing slowly enough for the industry to address safety concerns.  

Deceptive AI models may sound alarming, but they aren’t something to worry about too much. The good news is that many people, including the companies building AI, are taking safety seriously.

SourceGoogle releases new AI safety framework 

Anthropic is launching Claude Enterprise, a new AI chatbot plan for companies needing advanced admin controls and security. This offering directly competes with ChatGPT Enterprise, introduced by OpenAI last year.  

Claude lets enterprise firms upload company data for analysis, Q&A, graphics, web pages, or as a custom AI assistant.  

Anthropic is adding features to Claude that mirror ChatGPT’s business offerings.  

The reality is that Claude has been usable for companies for a year. Candidly, we’ve had a product in the market for a lot less time. Anthropic product lead Scott White told TechCrunch that we’re responding to the needs of our customers in a high-velocity city with a smaller team.  

In May, Anthropic launched Claude Team for small businesses. Since then, it has released mobile apps for iOS and Android. Now it is directly competing with ChatGPT Enterprise, which is widely used by Fortune 500 companies.  

Claude Enterprise stands out with a 500,000 token context window, more than double that of ChatGPT Enterprise or the Claude Team Plan.  

Claude Enterprise also provides collaborative workspaces, called projects and artifacts, where multiple users can upload and edit content. These features help businesses manage complex projects with various data sources and participants. Anthropic considers these workspaces a key competitive advantage.  

Another competitive advantage is GitHub integration, which enables direct synchronization between Claude and the customer’s codebases. This feature, leveraged by engineering teams, streamlines onboarding, bug fixes, and feature development, distinguishing Claude from some enterprise AI tools.  

Similar to ChatGPT’s Enterprise plan, Claude Enterprise allows businesses to assign a primary owner for their workspace. This owner can set different access levels for projects and data, and monitor system activity to ensure security and compliance.  

Anthropic also says, as OpenAI does, that it does not train its models on Claude enterprise customer data. This is important for businesses that want to keep their trade secrets out of Claude or ChatGPT’s knowledge base in the future.  

Anthropic has not shared the pricing for Claude Enterprise. White said it costs more than the $30 Team plan, but offers greater value. OpenAI also keeps its enterprise pricing private.  

White says Anthropic has been working in a private beta for months with early adopters such as GitLab, Mid Journey, IG Group, and Menlo Ventures (an investor in Anthropic).  

However, gaining expanded adoption will be key. AI model developers like Anthropic have come under pressure to sell API access at ever-lower prices. Products like Claude enterprise offerings can drive revenue to a similar extent; however, broad adoption is needed to offset the high insurance costs they entail. It’s not clear that any AI model developers are profiting from these business-specific plans just yet. 

Source: Anthropic launches Claude Enterprise plan to compete with OpenAI

Microsoft has expanded its Co-Pilot ecosystem into an AI co-worker with enterprise security. In March 2026, Microsoft launched the Microsoft 365 E7 Frontier Suite, including Co-Pilot, with advanced security and identity controls, for $19.99 per user per month.  

Key benefits and features for enterprise users include strengthened security, better data protection, and enhanced AI-powered productivity tools.  

  1. New Security and Governance Features 
  • Agent 365 platform: administrators can monitor, manage, and secure AI agents in real time, treating them as they do human staff.  
  • Baseline security mode automatically enforces Microsoft’s security best practices across Office, SharePoint, and Teams, reducing risk from weak communication.  
  • Purview Data Loss Prevention (DLP) for Copilot blocks Copilot responses to prompts containing sensitive data, preventing internal leaks and unsafe searches.  
  • Item Level Data Risk Assessment: administrators can quickly identify and fix multiple overshared links in SharePoint and OneDrive using Purview.  
  • Expanded Enterprise Data Protection (EDP): All Copilot prompts and responses are logged in accordance with retention rules, ensuring compliance and protecting organizational information. They are not used for base model training.  
  1. Agentic Copilot Upgrades (Wave 3) 
  • Copilot Co-work: lets users delegate complex multi-step tasks to AI agents that automatically execute them in the background.  
  • Deep App Integration: Agent Mode lets Copilot directly edit, improve, and update existing files in Word, Excel, and PowerPoint.  
  • Multi-model strategy: Users can select Anthropic Cloud or OpenAI models, such as GPT-5.4, within Co-Pilot based on task needs.  
  1. Expanded Functionality And Administration 
  • SharePoint admin agent: AI assists admins in managing permissions, content retention, and access with natural language commands.  
  • Copilot Dashboard Enhancements: The dashboard now displays user sentiment scores, adoption trend graphs, and ROI calculation tools, enabling administrators to clearly measure business impact and the benefits of adoption.  
  • Organization assets in PowerPoint: Copilot automatically applies approved images and branding from SharePoint asset libraries to presentations.  

These updates announced by Microsoft on March 9, 2026, are designed to make AI adoption safe and scalable for organizations of all sizes.  

The Microsoft 365 E7 bundle launching in May for $99 combines AI management tools and advanced identity-tracking features to help enterprises boost Copilot AI adoption.  

Microsoft’s commercial CEO said this launch is aimed at driving greater Copilot adoption among commercial productivity subscribers, addressing limited usage.  

Microsoft is earning more revenue per commercial user, driven in part by Copilot adoption, though overall usage among commercial productivity subscribers is still growing.  

Microsoft is adding artificial intelligence to its Office suite and raising the price of its cloud-based version by 65%, aiming to attract more enterprise users of its Copilot.  

The new Microsoft 365 E7 bundle for corporate users will cost $99 per user each month, compared to the E5 subscription, which now costs $60 per user each month after price increases. E5 provides a full suite of productivity and security tools. E7 includes everything in E5 plus $30 Copilot AI, $12 Entra identity tools, and the $15 Agent 365 product for managing company AI agents, combined in one package.  

Over the past year, Microsoft has invested more than $100 billion in data center infrastructure, including NVIDIA chips for AI models. Selling AI products helps the company show a return on this investment.  

Customers who buy E7 or the standalone Co-Pilot will get access to Co-Pilot and Co-Works, developed in partnership with AI model developer Anthropic. Co-Pilot and Co-Works can handle multi-step tasks such as sending scheduled emails and preparing for meetings with documents and calls. It will be available as a search preview this month for clients in Microsoft’s Frontier program, which offers early access to AI features.  

This launch follows updates to Anthropic’s Claude Cowork service, which have raised concerns among some investors that AI models could become a competitive threat to established software companies.  

Judson Althoff, CEO of Microsoft’s commercial business, emphasized that the Copilot upgrades and the E7 launch on May 1 are intended to expand Copilot adoption and push more companies to upgrade employees to higher tiers.  

The majority of our base is E5 now, right? he said. And then we are going through healthy renewal cycles on E5 right now. But E5 was created pre the agentic world.  

Increasing productivity revenue is still a top priority for Microsoft, along with growing its cloud business.  

Despite Microsoft 365 commercial products and cloud services accounting for 30% of revenue, slower user growth means that additional revenue per user from Copilot is increasingly important for delivering business value.  

This trend is driving higher revenue per user, driven by increased Copilot usage.  

In January, Microsoft CEO Satya Nadella said the company had 15 million Microsoft 365 Copilot paid seats, or 3% of the seats for commercial Microsoft 365 subscriptions.  

Alastair Woolcock, an analyst at Gartner, said that including identity management and security software in E7 is important for helping large companies safely distribute modern AI tools and boost productivity.  

Nobody wants to buy a dozen different $ 20-a-month products, right? He said.  

In a note to clients on Thursday, Jefferies analysts led by Brent Thiel reiterated the firm’s buy rating on Microsoft’s stock after meeting the company’s vice president of investor Relations, Jonathan Nielsen.  

Thill wrote that the company increasingly believes Microsoft 365 is entering a period of market growth, driven by its user base of about 450 million.  

Management noted that while third-party offerings (e.g., Claude, Cowork) are garnering hype, the majority of AI-powered work continues to occur within MSFT applications, creating incremental users of MSFT IP (Outlook, Teams, Excel, PPT, etc.), Thill wrote.

Source: Tech Microsoft adds higher-priced Office tier with Copilot as it tries to juice sales with AI