top of page

Your AI agent could be your next security breach and you might not even know it.

  • gs9074
  • Aug 5
  • 2 min read

Updated: Aug 19

AIs are increasingly doing more for you with new agent based features shipped faster than ever.


But behind every “smart assistant” or “agent-powered” workflow lies a hidden risk few teams are securing.


If your AI agent can act on your behalf—query data, call APIs, or trigger workflows—it’s relying on something most teams overlook: a tool server.


And if that server is misconfigured or compromised, your agent becomes a liability.


What’s really happening when an agent takes action?

Modern AI agents (like OpenAI Assistants, LangChain agents, or custom GPTs) don’t just generate text—they choose tools, send inputs, and act.

To do this, they need to know what tools exist, what each one does, and which ones they’re allowed to use.

This tool metadata is typically served via a system called MCP: Model Context Protocol.

You might not have built an MCP server explicitly but if you’re letting agents call functions or external APIs, you’re already depending on one.



So where’s the risk?



If an MCP server:


  • Returns unsafe or unverified tools

  • Ranks malicious tools higher

  • Forgets to remove deprecated or admin-level tools

  • Is managed by a third party you don’t control



…then your agent is one prompt away from doing something it shouldn’t.


This includes:


  • Leaking customer data

  • Making external calls to malicious servers

  • Executing poisoned functions

  • Bypassing user-level constraints


GitHub’s advice—only part of the picture

GitHub recently published guidance on building MCP servers. It covers:

• ✅ Authentication (OAuth, keys)

• ✅ Deployment architecture (containerised tools)

• ✅ Metrics and alerts

But it doesn’t cover:

• ❌ Tool poisoning

• ❌ Preference manipulation

• ❌ Rogue third-party tools

• ❌ Scanning for agent-safe behaviour

It’s like securing your front door but leaving your garage open—because you forgot your agent had the key.

What should you do right now?


You don’t need a massive overhaul.

But you do need these minimums in place:


  • Limit which tools your agents can access

  • Scan tool metadata for unsafe patterns (use MCPSafetyScanner or equivalent)

  • Log and monitor all agent-tool interactions

  • Avoid using third-party MCP servers you haven’t vetted

  • Treat agent tool access like you would access to production APIs


Your agents are smart.

But they’re only as safe as the tools they’re told to use.


At Bagh Co, we help fast-moving teams ship AI features without opening up security holes.

We review your architecture, secure your agent-tool interface, and help you avoid hidden liabilities.


→ Book a 30-minute agent safety review

No pitch. No fluff. Just clarity before things go sideways.

 
 
 

Comments


Bagh Co Logo

Bagh Co Ltd

  • LinkedIn
  • X
  • Threads

©2025 by Bagh Co Ltd.

bottom of page