Trending:
Cybersecurity

230 malicious OpenClaw extensions stolen crypto data since January 27

Security researchers documented 230 malicious OpenClaw "skills" disguised as crypto trading tools uploaded to ClawHub since late January. The extensions exploit OpenClaw's unsandboxed architecture to steal browser data and cryptocurrency information. Enterprise deployments are exposed: hundreds of instances run online without authentication.

What happened

OpenSourceMalware documented 230 malicious OpenClaw "skills" (the platform's term for extensions) uploaded to ClawHub since January 27, 2026. The malware disguised itself as cryptocurrency trading and wallet automation tools, targeting Windows and macOS users.

The initial cluster - 14 skills uploaded between January 27-29 - made it onto ClawHub's front page before removal. No new uploads have been reported in the past week, and ClawHub has not publicly responded.

Why this matters

OpenClaw (formerly Clawdbot/Moltbot) is a self-hosted AI agent with local file and network access. Unlike browser extensions that run in sandboxes, OpenClaw skills execute with full system privileges - essentially trusted code.

ClawHub, the public registry for these skills, performs no security scanning or sandboxing. When you install a skill, you're granting it the same access level as any application on your machine. The architecture assumes trust.

The enterprise angle

This isn't just a consumer problem. Security scans found hundreds of OpenClaw/Clawdbot instances exposed online. Eight were fully open - no authentication, configs and command histories visible.

The deployment pattern suggests business use: these aren't hobbyist setups. Organizations are running AI agents with privileged access in production environments, pulling code from an unvetted public repository.

What this signals

The incident previews a broader problem with agentic AI systems. First-party AI agents from major vendors face the same security-versus-convenience trade-offs. Users want capabilities; security teams want controls. The gap between those positions is where malware lives.

For enterprise architecture teams: if you're evaluating AI agent platforms, the questions are straightforward. Where do extensions run? What vetting exists? Can you allowlist sources? How do you audit what's installed?

The OpenClaw case demonstrates what happens when those questions go unanswered. Two hundred thirty times over, apparently.