The rise of autonomous AI agents has brought with it a new kind of app store, public registries where developers upload reusable "skills" that extend what large language models can do. Need your agent to analyze stocks, process medical data, or automate workflows? There's a skill for that. But as with any open marketplace, openness comes with risk. And according to a new large-scale study, nearly 1 in 20 publicly available agent skills is designed to do something harmful.

Researchers at CISPA Helmholtz Center for Information Security conducted the first systematic investigation of harmful skills across two of the largest agent skill registries. Their findings reveal not only the scale of the problem, but a surprising vulnerability in how LLM-based agents handle these skills, one that dramatically undermines the safety guardrails you might assume are in place.

A Different Kind of Threat

Most existing research on agent security has focused on what you might call "trojan horse" threats: skills that look legitimate but secretly contain malicious code, prompt injections, or data-stealing payloads. In those scenarios, the user is the victim.

This study flips the threat model. Here, the user is the attacker. They deliberately install a skill whose entire advertised purpose violates usage policies, think tools for launching cyber attacks, scraping private data, generating fraud scams, or creating non-consensual sexual content. The agent becomes the weapon, and third parties become the victims.

The researchers call these harmful skills, and they found them everywhere.

The Numbers Are Sobering

After analyzing 98,440 skills across the two largest registries, ClawHub and Skills.Rest, the team identified 4,858 harmful skills, roughly 4.93% of the entire ecosystem. The breakdown is uneven:

The most common harmful categories form a depressingly familiar list of modern digital threats:

Category Count
Cyber Attacks 1,134
Privacy Violation 962
Fraud & Scams 926
Unsupervised Financial Advice 865
Platform Abuse 732

Together, these five categories account for 74% of all policy violations. Roughly a quarter of harmful skills violate multiple categories simultaneously, meaning a single skill might combine, say, privacy scraping with platform abuse and fraud.

Perhaps most troubling: harmful skills aren't languishing in obscurity. On ClawHub, the median harmful skill actually receives more downloads than the median benign skill (261 vs. 229). Users appear to value these tools. Production is also concentrated: on Skills.Rest, just 10% of builders produce nearly half of all harmful skills, with one prolific contributor publishing 82 harmful skills alone.

The "Skill-Reading Exploit": A Surprising Safety Gap

Here's where the study gets really interesting. The researchers built HarmfulSkillBench, a benchmark of 200 human-verified harmful skills across 20 categories, and tested six major LLMs (GPT-4o, GPT-5.4-Mini, Gemini 3 Flash, Qwen3-235B, Kimi K2.5, and DeepSeek V3.2) under four conditions.

The results exposed a fundamental weakness in how agents handle skills.

When a user asks a harmful question directly, without any skill installed, models refuse most of the time. The average harm score across models is just 0.27, with a refusal rate of roughly 60%. Good news: safety training works.

But when that same harmful task is wrapped inside an installed skill and the agent is asked to execute it, refusal rates drop sharply. The harm score climbs to 0.47.

When the user doesn't even state the harmful intent explicitly, when they simply ask the agent to "read the skill and create an execution plan for its intended purpose," the harm score rockets to 0.76. The average refusal rate collapses to just 9.75%.

The researchers call this the skill-reading exploit. It reveals that safety training done on user-query prompts doesn't transfer to semantically identical content delivered through a pre-installed skill. The skill acts as a kind of laundering mechanism for harmful intent.

The Tier 2 Problem: Silent Compliance in High-Risk Domains

The study also examined "high-risk" skills, those operating in professional domains like legal advice, medical advice, insurance underwriting, financial planning, employment screening, and academic assessment. Established AI usage policies require two safeguards for these domains: human-in-the-loop (HiTL) review and AI disclosure (AID).

The findings here are perhaps even more concerning than the skill-reading exploit. On Tier 2 high-risk skills:

When explicitly instructed to include HiTL review, models obeyed roughly 97% of the time. But when explicitly instructed to disclose AI involvement, compliance dropped to 41-74%. Meanwhile, when told not to disclose AI involvement, models complied 99% of the time.

This asymmetry reveals a structural bias toward non-disclosure. Models seem reluctant to volunteer that AI is behind high-stakes decisions, like insurance underwriting or candidate screening, unless explicitly told to do so. Under one tested condition, all six models produced end-to-end autonomous insurance underwriting plans with neither safeguard when instructed to omit them.

Why Existing Defenses Fall Short

Both major registries claim some form of security review:

But these mechanisms operate at the security layer, not the policy layer. They're designed to catch malicious code, embedded malware, prompt injections, known attack signatures. They don't assess whether a skill's advertised functionality itself violates usage policies.

A skill called "phishing-campaign-launcher" that honestly describes its purpose in clean, well-structured markdown will pass every security scan. The harm isn't hidden. It's declared openly.

What Organizations Deploying AI Agents Should Take Away

For any organization building products on top of agent frameworks, this research offers several practical warnings:

01

Skill sources matter.

If your agents can install skills from public registries, assume that roughly 5% of what's available is designed to violate usage policies. Curated, vetted skill sets are not optional.

02

Pre-installed skills expand the attack surface in unexpected ways.

Even if your users don't intend harm, a harmful skill sitting in the agent's tool context can be triggered by innocuous-seeming requests. The safety training protecting your LLM was not designed for this input channel.

03

Safeguards must be defaults, not opt-ins.

For high-risk decision domains, "the user can request human review" is not adequate protection. Human-in-the-loop review and AI disclosure need to be baked into the agent's behavior regardless of user instructions.

04

Registry-level moderation needs to evolve.

Security scanning for malware is necessary but insufficient. Registries should add policy-compliance analysis, publisher identity verification, and domain-expert review for sensitive categories like weapons, elections, and professional advice.

05

Alignment training needs to catch up with agent architectures.

Current safety alignment treats user queries as the primary input channel. But in agent systems, skill specifications are just as consequential, and currently much less scrutinized by safety measures.

The Bigger Picture

Agent skill ecosystems are young. ClawHub and Skills.Rest together host nearly 100,000 skills, and that number is growing rapidly. As agents become more capable and more autonomous, the attack surface represented by these registries will only expand.

The researchers responsibly disclosed their findings to affected registries, some of which removed flagged skills. They've also released HarmfulSkillBench under a gated-access protocol to support further research, while redacting specific attack details to prevent misuse.

But the core challenge remains: in the race to build useful AI agents, the infrastructure for sharing agent capabilities has outpaced the infrastructure for ensuring those capabilities are safe. The result is a marketplace where 5% of what's on the shelves is designed to cause harm, and where the safety mechanisms built into the AI itself can be bypassed simply by presenting a harmful task as a file to be read rather than a question to be answered.

For anyone building, deploying, or relying on AI agents, that's a gap that demands attention now, not after the next high-profile incident.

Need help assessing AI agent risk in your organization? Get in touch with our team.