DEV Community

AgentGraph
AgentGraph

Posted on • Edited on • Originally published at agentgraph.co

We Scanned 231 OpenClaw Skills for Security Vulnerabilities — Here's What We Found

AI agents are running third-party code on your machine. Last week, Anthropic announced extra charges for OpenClaw support in Claude Code, drawing fresh attention to the ecosystem. We wanted to answer a straightforward question: how safe are the most popular OpenClaw skills?

We first published results from 25 repos. We have now expanded the scan to 231 repositories out of 2,007 discovered — nearly a 10x increase in coverage — and the picture has gotten worse.

Why Independent Trust Verification Matters Now

Anthropic just temporarily banned OpenClaw's creator from accessing Claude (TechCrunch, April 10). Whether you agree with their decision or not, it highlights a structural gap: platform trust is revocable. There's no independent way to verify whether an AI agent or tool is safe to use.

That's why we built agentgraph.co/check — a free, instant safety checker for any AI agent, MCP server, or skill. Paste a URL, get a letter grade. The result is a cryptographically signed attestation that you can verify yourself. No platform controls the score.

Methodology

We used AgentGraph's open-source security scanner to analyze 231 OpenClaw skill repositories from GitHub (out of 2,007 discovered). The scanner inspects source code for:

  • Hardcoded secrets (API keys, tokens, passwords in source)
  • Unsafe execution (subprocess calls, eval/exec, shell=True)
  • File system access (reads/writes outside expected boundaries)
  • Data exfiltration patterns (outbound network calls to unexpected destinations)
  • Code obfuscation (base64-encoded payloads, dynamic imports)

It also detects positive signals: authentication checks, input validation, rate limiting, and CORS configuration. Each repo receives a trust score from 0 to 100.

Results Summary

All 231 repositories scanned successfully. The aggregate numbers:

Metric Value
Repos discovered 2,007
Repos scanned 231
Total findings 14,350
Critical 98
High 6,192
Medium 8,045
Repos with critical findings 20 (9%)
Average trust score 57.0 / 100 (Grade C)
Repos scoring F (0-20) 74 (32%)

Findings by category: file system access accounted for 8,239, unsafe execution patterns for 5,871, data exfiltration patterns for 146, hardcoded secrets for 58, dependency vulnerabilities for 29, and code obfuscation for 7.

Score Distribution

Score Range Grade Repos Percentage
81 - 100 A / A+ 118 51%
61 - 80 B / B+
41 - 60 C
21 - 40 D
0 - 20 F 74 32%

The distribution remains bimodal. More than half of repos score A or above, but over a quarter score F. Repos tend to be either clean or deeply problematic, with almost nothing in the middle. There is no gentle gradient between "secure" and "insecure" — it is one or the other.

Notable Findings

openclaw/clawhub (official skill registry)
Score: 0/100. 2 critical, 228 high, 75 medium findings across 200 files. This is the registry that indexes skills for the broader ecosystem.

adversa-ai/secureclaw (OWASP security plugin)
Score: 0/100. 21 critical, 66 high, 177 medium findings. A security-focused plugin that itself has significant findings. The scanner flagged a high density of unsafe execution patterns and file system access.

openclaw/openclaw (main framework)
Score: 0/100. 1 critical, 14 high, 4 medium findings. The core framework that other skills build on.

FreedomIntelligence/OpenClaw-Medical-Skills (medical AI)
Score: 0/100. 1 critical, 30 high, 12 medium findings. Medical AI skills with critical findings deserve particular scrutiny given their potential deployment context.

Not all skills are problematic. tuya/tuya-openclaw-skills scored 95/100, and several others came in at 90/100. The clean repos demonstrate that writing secure OpenClaw skills is entirely achievable — it is just not the norm across the board.

What This Means

When Claude Code or any AI assistant runs a third-party tool, it executes that tool's code with whatever permissions the host process has. If that code contains unsafe exec patterns, broad file system access, or exfiltration vectors, the attack surface is your machine — your files, your environment variables, your credentials.

The finding categories tell the story: 5,871 unsafe execution patterns means eval, exec, subprocess, and shell=True calls scattered across these codebases. 8,239 file system access findings means code reaching into the filesystem in ways that may not be bounded. 146 data exfiltration patterns and 58 hardcoded secrets round out the picture.

Anthropic's decision to gate OpenClaw behind additional pricing starts to make more sense in this context. The cost is not just computational — it is risk.

New: PyPI Packages and Trust Gateway

Since the initial scan, we have shipped three PyPI packages:

  • agentgraph-trust (v0.3.1) — the MCP server for scanning tools directly from Claude Code or any MCP-compatible client
  • agentgraph-agt — the AgentGraph Trust CLI for CI pipelines and local use
  • open-agent-trust — a lightweight library for embedding trust checks into any Python agent framework

We have also built a trust gateway — an enforcement layer that sits between your agent runtime and third-party tools. Instead of scanning after the fact, the gateway intercepts tool invocations at runtime and makes enforcement decisions based on the tool's trust score: allow, throttle, require user confirmation, or block entirely. The trust tiers (detailed below) drive these decisions automatically.

The gateway turns scan results into policy. A tool scoring 0/100 does not just get a warning — it gets denied execution unless the user explicitly overrides.

Check Your Own Tools

We built an MCP server that lets you check any agent or tool directly from Claude Code.

Install:

pip install agentgraph-trust
Enter fullscreen mode Exit fullscreen mode

Add to your Claude Code MCP config:

{
  "mcpServers": {
    "agentgraph-trust": {
      "command": "agentgraph-trust",
      "env": {
        "AGENTGRAPH_URL": "https://agentgraph.co"
      }
    }
  }
}
Enter fullscreen mode Exit fullscreen mode

Then ask Claude: "Check the security of [agent name]"

It returns a signed attestation with findings, trust score, and boolean safety checks. The attestation is cryptographically signed (Ed25519, JWS per RFC 7515) and verifiable against our public JWKS at https://agentgraph.co/.well-known/jwks.json.

Public API — Trust-Tiered Rate Limiting

We also built a free public API that any framework can use to check tools before execution. No authentication required.

GET https://agentgraph.co/api/v1/public/scan/{owner}/{repo}
Enter fullscreen mode Exit fullscreen mode

The API returns a trust tier with recommended rate limits:

Tier Score Rate Limit Token Budget User Confirm
verified 96-100 unlimited unlimited No
trusted 81-95 60/min 8K No
standard 51-80 30/min 4K No
minimal 31-50 15/min 2K Yes
restricted 11-30 5/min 1K Yes
blocked 0-10 denied denied N/A

Every response includes a signed JWS attestation. Framework authors can use the trust tier to throttle tool execution — spend less compute on risky tools, let clean tools run freely.

This is the foundation for a trust gateway: instead of binary accept/deny, graduated throttling based on verified security posture.

You can also embed a trust badge in your README:

![Trust Score](https://agentgraph.co/api/v1/public/scan/{owner}/{repo}/badge)
Enter fullscreen mode Exit fullscreen mode

Full Data

The scanner and full results are open source:

Try It Now

agentgraph.co/check — Paste any GitHub repo URL, MCP server name, or agent package and get an instant letter grade. No signup, no API key, no cost. The result is a signed attestation you can independently verify.

7 PyPI packages available now:

Package Purpose
agentgraph-trust MCP server — scan tools from Claude Code or any MCP client
agentgraph-agt CLI for CI pipelines and local scanning
open-agent-trust Lightweight library for embedding trust checks in any Python agent
agentgraph-scanner Core scanning engine
agentgraph-attestation Cryptographic attestation signing and verification
agentgraph-gateway Trust gateway enforcement layer
agentgraph-badges Trust badge generation for READMEs

GitHub Action — Add trust scanning to any CI pipeline. Runs on every PR, blocks merges that introduce tools below your trust threshold. Drop it into your workflow in two lines:

- uses: agentgraph-co/agentgraph-trust-action@v1
  with:
    fail-below: 50
Enter fullscreen mode Exit fullscreen mode

The agent ecosystem needs trust infrastructure. We are building it at agentgraph.co.

Top comments (0)