🎧 Listen to this article

What Is OpenC" cover: image: "" alt: “AI Agents News” hidden: true

The promise of autonomous AI agents has collided with a harsh reality: the very openness that makes them powerful also makes them dangerously vulnerable to code injection attacks.


What Is OpenClaw?

OpenClaw (formerly Clawdbot/Moltbot) is an open-source AI agent that took the developer world by storm—amassing over 100,000 GitHub stars in just days. It connects to messaging apps (WhatsApp, Telegram, Discord, Slack) and can autonomously:

  • Execute shell commands
  • Access files and databases
  • Browse the web
  • Manage user data
  • Install “skills” from a marketplace called ClawHub

Sounds powerful? It is. But that power comes with a price.


The Architecture Problem: Data vs. Instructions

The fundamental flaw in AI agents like OpenClaw is that they blur the line between data and instructions.

In traditional software, code and data are separate. A database stores information; a program processes it. But LLM-powered agents interpret natural language as executable commands. This creates inherent injection vulnerabilities that traditional security tools cannot detect.

Three attack vectors dominate:

Attack TypeHow It Works
Direct Prompt InjectionMalicious instructions embedded in content the agent processes (emails, documents, webpages)
Indirect Prompt InjectionHidden malicious instructions in external data sources (RAG databases, API responses, files)
Code Injection via SkillsMalicious packages in ClawHub execute arbitrary code with full system privileges

Critical Vulnerabilities: The CVEs You Need to Know

Security researchers have disclosed multiple high-severity vulnerabilities in OpenClaw:

CVE-2026-25253 (CVSS 8.8) — “ClawJacked”

One-click Remote Code Execution via WebSocket hijacking. A malicious website could steal authentication tokens and take full control of the agent—even on localhost-bound instances.

CVE-2026-24763 & CVE-2026-25157 (CVSS 7.8-8.8)

Command injection vulnerabilities allowing arbitrary shell command execution through crafted input.

The Default Configuration Nightmare

  • Authentication disabled by default — Gateway openly accessible on installation
  • No WebSocket origin validation — Any website can connect to your local agent
  • No rate limiting — Brute-force attacks possible without detection
  • Plaintext credential storage — API keys and tokens stored unencrypted
  • mDNS broadcast leaks — Configuration parameters exposed across local networks

The ClawHavoc Supply Chain Attack

Perhaps the most alarming finding is the systematic poisoning of the skills marketplace:

  • 341+ malicious skills discovered on ClawHub out of ~2,857 total (12% of the ecosystem)
  • 1,184 confirmed malicious skills by March 2026
  • Skills disguised as legitimate crypto tools, trading bots, and productivity apps
  • Payloads included Atomic macOS Stealer (AMOS), reverse shells, and credential harvesters
  • 91% of malicious skills combined prompt injection with traditional malware

The Social Engineering Factor

Attackers didn’t just exploit technical vulnerabilities—they convinced users to run “setup commands” that installed malware. The trust users placed in the ClawHub marketplace became the attack vector.


This Isn’t Just an OpenClaw Problem

The AI agent security crisis extends across the entire ecosystem:

Agent/ToolVulnerability
Claude CodeCVE-2025-59536 — Configuration injection allowing arbitrary shell commands
Cursor, Cline, GitHub CopilotPrompt injection in CI/CD workflows
MCP ClientsCommand injection via OAuth flows

Snyk’s “ToxicSkills” Research Findings:

  • 36.8% of all agent skills have at least one security flaw
  • 13.4% contain critical-level issues
  • 76 confirmed malicious payloads for credential theft and backdoor installation

Why Traditional Security Fails Against AI Agents

AI agents bypass conventional security controls in three critical ways:

1. Perimeter Defenses Are Useless

Prompt injection operates at the semantic layer, not the network or application layer. Firewalls and intrusion detection systems can’t distinguish between legitimate instructions and malicious ones when both use natural language.

2. Behavior Appears Legitimate

Malicious actions use the agent’s own APIs and permissions. When an AI agent exfiltrates data, it looks like normal operation—not like a breach.

3. No Code Signing for Skills

Anyone can publish to ClawHub. Verification is minimal. Compare this to mobile app stores with rigorous review processes—the difference is stark.

4. Shadow AI Adoption

22% of enterprises report unauthorized OpenClaw usage by employees. These tools are being adopted faster than security teams can assess them.


Real-World Exploit Scenarios

Security researchers have demonstrated:

  • 5-minute private key extraction via email prompt injection
  • Email deletion by tricking the agent with misleading instructions
  • Full admin interface access through exposed reverse proxies
  • Data exfiltration via link previews in messaging apps
  • Cross-tenant data access in multi-user environments

The Bottom Line

“An AI agent that interprets natural language, has built-in tooling for code execution and file access, and looks like legitimate developer software to endpoint detection tools is a potent post-exploitation asset.”

— Yuval Zacharia, Security Researcher

The “open” nature of these agents—their ability to execute arbitrary code, install community extensions without verification, and interpret natural language as instructions—creates an attack surface that traditional security tools cannot adequately protect.


What Should You Do?

If You’re Using AI Agents:

  1. Enable authentication immediately — Don’t run with default settings
  2. Isolate the environment — Run in a container or VM with limited privileges
  3. Audit all skills — Only install from trusted sources; review code when possible
  4. Monitor outbound connections — Use network segmentation to detect exfiltration
  5. Disable unnecessary capabilities — If you don’t need shell access, turn it off

If You’re Building AI Agents:

  1. Implement sandboxing — Skills should never run with host system privileges
  2. Require code signing — Verify publisher identity before execution
  3. Add semantic filtering — Detect and block suspicious prompt patterns
  4. Enable audit logging — Track all actions for forensic analysis
  5. Default-deny permissions — Require explicit user approval for sensitive operations

The Path Forward

Until robust sandboxing, mandatory authentication, and supply chain verification become standard in AI agent architectures, these tools remain inherently unsafe for production use. The convenience of autonomous AI comes with security tradeoffs that many organizations aren’t prepared to handle.

The question isn’t whether AI agents will be attacked—it’s whether you’ll detect it when they are.


Sources & Further Reading

  • Snyk ToxicSkills Research (February 2026)
  • Koi Security ClawHavoc Report
  • Check Point Research on Claude Code Vulnerabilities
  • Oasis Security OpenClaw Analysis
  • CNCERT China Advisory (March 2026)
  • Microsoft Security Blog: “Agent Runtime Risk”
  • HiddenLayer AI Threat Landscape Report 2025

What are your thoughts on AI agent security? Have you encountered these vulnerabilities in your work? Share your experiences in the comments.

#AIsecurity #Cybersecurity #OpenClaw #PromptInjection #CodeInjection #SupplyChainSecurity #AIagents #DevSecOps


Disclaimer: This blog post was automatically generated using AI technology based on news summaries. The information provided is for general informational purposes only and should not be considered as professional advice or an official statement. Facts and events mentioned have not been independently verified. Readers should conduct their own research before making any decisions based on this content. We do not guarantee the accuracy, completeness, or reliability of the information presented.