March 5, 2026·5 min read·Cortex Team

Your AI Agent Is a Security Nightmare (And Here's How to Fix It)

ai-securitymcpagent-safetyproduction-deploymentopenclaw

Free skill included with this post

Download on GitHub →

Your AI Agent Is a Security Nightmare (And Here's How to Fix It)

Last week, a developer posted a tool to Hacker News that struck a nerve. Within hours, it had hundreds of upvotes and dozens of comments from people saying the same thing: "I didn't realize how exposed my AI agents were until I saw this."

The tool was Kvlar, an open-source firewall for AI agent tool calls. It sits between your agent and its MCP (Model Context Protocol) servers, evaluating every tool call against security policies before execution.

The problem it solves? Most AI agents in production have zero security boundaries.

The Problem

When you connect Claude or another AI agent to your systems via MCP, you're giving it the ability to:

  • Execute arbitrary shell commands
  • Query and modify production databases
  • Push code to repositories
  • Send messages on your behalf
  • Access internal APIs and services

Most developers set up these integrations, test that they work, and call it a day. But they never stop to ask: What happens if the agent decides to run rm -rf /? Or drops a production table? Or exfiltrates data?

As one HN commenter noted: "AI agents using MCP can execute database queries, push code, send Slack messages, and run shell commands — with no security boundary. Claude Desktop has basic approve/deny, but it's per-call with no persistent rules, no automation, and no audit trail."

The recent BrokenClaw security research series exposed exactly how bad this can get — including remote code execution vulnerabilities through email processing.

Real-World Scenarios That Keep Security Teams Awake

The Database Disaster: An agent with Postgres access runs a DROP TABLE command after misunderstanding a user request. No backup exists. Your production database is gone.

The Repository Wipe: A GitHub-connected agent force-pushes an empty branch to main, rewriting your entire git history. Weeks of work vanish.

The Silent Exfiltration: An agent with shell access pipes sensitive data through curl to an external server. It looks like a legitimate API call in the logs.

The Privilege Escalation: An agent runs sudo commands on a server, installing persistent backdoors that persist after the agent session ends.

These aren't theoretical. They're the natural consequence of giving AI systems broad tool access without guardrails.

The Skill: MCP Security Audit

We built the mcp-security-audit skill to help you identify and fix these vulnerabilities before they become incidents.

What it does:

  1. Scans your MCP configurations for high-risk permissions and dangerous defaults
  2. Identifies overprivileged tools that could cause data loss or system compromise
  3. Generates security policies in Kvlar-compatible YAML format
  4. Provides actionable recommendations for hardening your deployment

Download free: github.com/thenatechambers/openclaw-skills-repo/tree/main/skills/mcp-security-audit

How to Use It

Step 1: Install the Skill

cp -r skills/mcp-security-audit ~/.openclaw/skills/
pip install -r ~/.openclaw/skills/mcp-security-audit/requirements.txt

Step 2: Run the Audit

# Audit all MCP servers in your OpenClaw config
python ~/.openclaw/skills/mcp-security-audit/audit.py --action audit

You'll get a report like this:

MCP SECURITY AUDIT REPORT

Overall Risk Score: 72/100 - HIGH RISK

Server: postgres
   Command: npx
   Risk Score: 45/100
   Issues Found:
   Database write access enabled
      Gate DELETE/DROP/UPDATE operations

Server: shell
   Command: bash
   Risk Score: 85/100
   Issues Found:
   Shell access detected - can execute arbitrary commands
      Restrict to allowedCommands or use deny-by-default policy

Step 3: Generate Security Policies

# Generate policies for high-risk servers
python ~/.openclaw/skills/mcp-security-audit/audit.py \
  --action generate-policy \
  --servers postgres,shell \
  --output policy.yaml

This creates a YAML policy file:

policies:
  - name: postgres-safety
    server: postgres
    defaultAction: gate
    rules:
      - pattern: "SELECT *"
        action: allow
        reason: "Read-only operations"
      - pattern: "DROP *"
        action: deny
        reason: "Prevent schema destruction"

  - name: shell-restrictions
    server: shell
    defaultAction: deny
    allowedCommands:
      - "ls *"
      - "cat *"
      - "grep *"
    blockedPatterns:
      - pattern: "rm -rf *"
        reason: "Recursive deletion"
      - pattern: "sudo *"
        reason: "Privilege escalation"

Step 4: Deploy with Kvlar (Optional but Recommended)

Kvlar is a production-ready MCP firewall that enforces these policies:

kvlar init --config policy.yaml
kvlar wrap

Now every tool call is checked against your policies before execution. Dangerous operations are blocked. Sensitive operations require approval. You have an audit trail.

The Recommendation

If you're running AI agents with MCP access to production systems, you need security policies. Not tomorrow. Today.

Start with these rules:

  1. Default-deny for destructive operations - Block DROP, DELETE, rm -rf, sudo, and force-push by default
  2. Gate sensitive write operations - Require human approval for UPDATE, INSERT, git push, and API mutations
  3. Restrict shell access - If you need shell tools, whitelist specific safe commands only
  4. Audit weekly - MCP configurations drift over time. Run this audit regularly
  5. Segment by environment - Different policies for dev, staging, and production

Why This Matters for Cortex Users

Cortex is designed for production AI agent deployments. Production means security is non-negotiable.

This skill gives you:

  • Visibility into what your agents can actually do
  • Control through enforceable security policies
  • Confidence that a misunderstood prompt won't become a production incident

The teams winning with AI agents aren't just building fast — they're building safely. Security isn't friction; it's what lets you move faster without breaking things.


Want to deploy your own secure AI agent? Sign up for Cortex →

Have questions about agent security? Reply to this post or reach out on GitHub.

Get new posts + free skills in your inbox

One email per post. Unsubscribe anytime.

Want an AI agent that runs skills like these automatically?

Cortex deploys your own AI agent in 10 minutes. No DevOps required.

Start free trial →