News: 1771549275

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

AI agents abound, unbound by rules or safety disclosures

(2026/02/20)


AI agents are becoming more common and more capable, without consensus or standards on how they should behave, say academic researchers.

So says MIT’s Computer Science & Artificial Intelligence Laboratory (CSAIL), which analyzed 30 AI agents for its [1]2025 AI Agent Index , which assesses machine learning models that can take action online through their access to software services.

AI agents may take the form of chat applications with tools (Manus AI, ChatGPT Agent, Claude Code), browser-based agents (Perplexity Comet, ChatGPT Atlas, ByteDance Agent TARS), or enterprise workflow agents (Microsoft Copilot Studio, ServiceNow Agent).

[2]

The paper accompanying the AI Agent Index observes that despite growing interest and investment in AI agents, "key aspects of their real-world development and deployment remain opaque, with little information made publicly available to researchers or policymakers."

[3]

[4]

The AI community frenzy around open source agent platform [5]OpenClaw , and its accompanying agent interaction network Moltbook – plus ongoing [6]frustration with AI-generated code submissions to open source projects – underscores the consequences of letting agents loose without behavioral rules.

In the paper, the authors note that the [7]tendency of AI agents to ignore the Robot Exclusion Protocol – which uses robots.txt files to signal no consent to scraping websites – suggests that established web protocols may no longer be sufficient to stop agents.

[8]

It's a timely topic. Anthropic, one of the main providers of AI agents, on Wednesday published its own [9]analysis of AI agent autonomy , focused more on how agents are used than the consequences of their use.

"AI agents are here, and already they're being deployed across contexts that vary widely in consequence, from email triage to [10]cyber espionage ," the company said. "Understanding this spectrum is critical for deploying AI safely, yet we know surprisingly little about how people actually use agents in the real world."

According to consultancy [11]McKinsey , AI agents have the potential to add $2.9 trillion to the US economy by 2030 – assuming the vast capital expenditures by OpenAI and other tech firms haven't derailed the hype train. We note that enterprises [12]aren't yet seeing [13]much of [14]a return on their AI investments . And researchers last year [15]found AI agents could only complete about a third of multi-step office tasks. But AI models have improved since then.

[16]AI agents can't teach themselves new tricks – only people can

[17]Google germinates Gemini 3.1 Pro in ongoing AI model race

[18]Android malware taps Gemini to navigate infected devices

[19]Don't believe the hyperscalers! AI can't cure the climate crisis

MIT CSAIL's 2025 AI Agent Index covers 30 AI agents. It is smaller than its 2024 predecessor, which looked at 67 agentic systems. The authors say the 2025 edition goes into greater depth, analyzing agents across six categories: legal, technical capabilities, autonomy & control, ecosystem interaction, evaluation, and safety. The AI Agent Index site makes this information available for every listed agent, each with 45 annotation fields.

According to the researchers, 24 of the 30 agents studied were released or received major feature updates during the 2024-2025 period. But the developers of agents talk more about product features than about safety practices.

[20]

"Of the 13 agents exhibiting frontier levels of autonomy, only four disclose any agentic safety evaluations (ChatGPT Agent, OpenAI Codex, Claude Code, Gemini 2.5 Computer Use)," according to the researchers.

Developers of 25 of the 30 agents covered provide no details about safety testing and 23 offer no third-party testing data.

To complicate matters, most agents rely on a handful of foundation models – the majority are harnesses or wrappers for models made by Anthropic, Google, and OpenAI, supported by scaffolding and orchestration layers.

The result is a series of dependencies that are difficult to evaluate because no single entity is responsible, the MIT boffins say.

Delaware-incorporated companies created 13 of the agents evaluated by the authors. Five come from China-incorporated organizations, and four come have non-US, non-China origins: specifically Germany (SAP, n8n), Norway (Opera), and Cayman Islands (Manus).

Among the five Chinese-incorporated agent makers, one has a published safety framework and one has a compliance standard.

For agents originating outside of China, 15 point to safety frameworks like Anthropic's Responsible Scaling Policy, OpenAI's Preparedness Framework, or Microsoft's Responsible AI Standard. The other ten lack safety framework documentation. Enterprise assurance standards are more common, with only five of 30 agents having no compliance standards documented.

Twenty-three of the evaluated agents are closed-source. Developers of seven agents open-sourced their agent framework or harness – Alibaba MobileAgent, Browser Use, ByteDance Agent TARS, Google Gemini CLI, n8n Agents, OpenAI Codex, and WRITER.

All told, the Index found agent makers reveal too little safety information, and that a handful of companies dominate the market. Other major findings include the difficulty of analyzing agents given their layers of dependencies, and that agents aren't necessarily welcome at every website.

The paper lists the following authors: Leon Staufer (University of Cambridge), Kevin Feng (University of Washington), Kevin Wei (Harvard Law School), Luke Bailey (Stanford University), Yawen Duan (Concordia AI), Mick Yang (University of Pennsylvania), A. Pinar Ozisik (MIT), Stephen Casper (MIT), and Noam Kolt (Hebrew University of Jerusalem). ®

Get our [21]Tech Resources



[1] https://aiagentindex.mit.edu/

[2] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=2&c=2aZfqcTTVGpasd3I8RghHWgAAAsM&t=ct%3Dns%26unitnum%3D2%26raptor%3Dcondor%26pos%3Dtop%26test%3D0

[3] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aZfqcTTVGpasd3I8RghHWgAAAsM&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[4] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aZfqcTTVGpasd3I8RghHWgAAAsM&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[5] https://www.theregister.com/2026/02/03/openclaw_security_problems/

[6] https://www.theregister.com/2026/02/12/ai_bot_developer_rejected_pull_request/

[7] https://www.theregister.com/2025/12/10/really_simple_licensing_spec_takes/

[8] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aZfqcTTVGpasd3I8RghHWgAAAsM&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0

[9] https://www.anthropic.com/research/measuring-agent-autonomy

[10] https://www.anthropic.com/news/disrupting-AI-espionage

[11] https://www.mckinsey.com/mgi/our-research/agents-robots-and-us-skill-partnerships-in-the-age-of-ai

[12] https://www.theregister.com/2026/01/28/ai_agents_redis/

[13] https://www.theregister.com/2025/10/01/gartner_ai_agents/

[14] https://www.theregister.com/2026/01/20/pwc_ai_ceo_survey/

[15] https://www.theregister.com/2025/06/29/ai_agents_fail_a_lot/

[16] https://www.theregister.com/2026/02/19/ai_agents_cant_teach_themselves/

[17] https://www.theregister.com/2026/02/19/google_germinates_gemini_31_pro/

[18] https://www.theregister.com/2026/02/19/genai_malware_android/

[19] https://www.theregister.com/2026/02/19/ai_climate_crisis_claims/

[20] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aZfqcTTVGpasd3I8RghHWgAAAsM&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0

[21] https://whitepapers.theregister.com/



Groo The Wanderer - A Canuck

I kind of pity the people who are trying to throw everything at AI.

Learning how to do something is what keeps the job fun and interesting.

You're going to be bored 'bot supervisors at best...

I pity you.

All this wool over my eyes and not a Llama in sight

Anonymous Coward

A bit surprised to see no Llama in this report, except for one mention of [1]PurpleLlama in HubSpot Breeze Agents' ' model protection layer '. Guess we'll have to wait for Avocado, Mango, and Wango Tango, to see what sort of [2]agentic guacamole disaster Meta's kitchen is cooking up ... (not that I'm looking forward to it or anything ;).

[1] https://github.com/meta-llama/PurpleLlama

[2] https://www.theregister.com/2023/07/13/robot_guacamole_chipotle/

If man is only a little lower than the angels, the angels should reform.
-- Mary Wilson Little