When enterprises want to deploy a chat bot, the calculus is straightforward. When they want to deploy an agent — something that can write files, send emails, execute code, provision resources — the calculus collapses. The agent can do things. There is nothing between the agent's decision and the consequences of that decision.
Microsoft published something on GitHub Tuesday that attempts to fill that gap. The Microsoft Agent Governance Toolkit is an open-source runtime enforcement layer for AI agents: it sits between the agent framework and the operating system, intercepts every tool call, and enforces policy before execution. Not after. Before.
The GitHub repository contains five language SDKs — Python (pip), TypeScript (npm), .NET (NuGet), Rust (crates.io), and Go — a combined 9,500-plus tests, and coverage for all 10 entries in the OWASP Agentic Top 10, the security framework that maps how autonomous AI systems can be compromised. Policy evaluation runs at sub-millisecond latency, according to the project's architecture documentation. The license is MIT.
"This is not a model safety or prompt guardrails tool," the README states flatly. "It governs agent actions — tool calls, resource access, inter-agent communication — at the application layer."
The distinction matters. Content moderation checks what an LLM says. Runtime governance checks what an agent does. A guardrail can tell an agent not to send an email to the wrong recipient. The Agent Governance Toolkit can block the email at the operating system level before it leaves the outbox, regardless of what the model's instructions said.
What OWASP Agentic Top 10 coverage looks like in practice
The OWASP Agentic Top 10 — finalized over the past year by a cross-industry working group — catalogues risks specific to AI agents: prompt injection through tool definitions, tool poisoning in registries, uncontrolled resource consumption, trust boundary violations, and similar failure modes. The standard has been visible in conference talks and threat models for months. Concrete implementations have been slower.
The Agent Governance Toolkit maps each of the 10 entries — labeled ASI-01 through ASI-10 — to specific enforcement mechanisms. ASI-01, prompt injection via tool definition, is addressed by the MCP Security Scanner, a component that scans Model Context Protocol tool definitions for signs of tool poisoning, typosquatting, hidden instructions, and rug-pull attacks. The MCP registry ecosystem has grown rapidly as a distribution mechanism for agentic AI tools; it has also become an attack surface, since any tool definition can contain instructions that an agent will execute without explicit user review.
The scanner is a CLI utility, callable as agentmesh trust report or via direct Python import. It is not a research prototype. It shipped.
Zero-trust identity for agents
The identity model uses Ed25519 cryptographic credentials with SPIFFE/SVID support — the same standards used for service mesh identity in cloud-native infrastructure — and produces a trust score from 0 to 1,000 for every agent and tool interaction. An agent executing within a governed environment carries credentials; those credentials are checked against policy at every boundary crossing.
Execution sandboxing uses four-tier privilege rings, saga orchestration for distributed rollback if an action sequence fails partway through, and an explicit kill switch for termination control. The combination means an enterprise can define policy — "this agent can read from this database but cannot write to it" — and enforce that policy structurally, not instructionally.
"Governs what agents do, not just what they say," is how the project puts it.
Agent SRE is also in the package: service level objectives, error budgets, replay debugging, circuit breakers, and chaos engineering tooling — the operational surface that production deployments need and that early-stage agent frameworks typically defer.
Works with anything, owns nothing
The README lists compatibility with AWS Bedrock, Google ADK, Azure AI, LangChain, CrewAI, AutoGen, OpenAI Agents, and LlamaIndex. The phrase "and more" appears at the end. Installing the Python package is pip install agent-governance-toolkit[full]. The TypeScript SDK is npm install @agentmesh/sdk. No Azure account required. No vendor lock-in.
This is an unusual posture for Microsoft. The company's historical pattern with open-source tooling — where it does engage open-source — tends toward/Azure-specific integration or permissive-but-not-MIT licensing. MIT means the code can be forked, embedded, and commercialized without Microsoft's name attached. The governance layer is now infrastructure that competitors can build on or around.
The practical implication for enterprise buyers is concrete: compliance teams have had nothing to point at when boards asked how agentic AI risks were being managed. The OWASP Top 10 coverage gives them a framework. The toolkit gives them an implementation. The MIT license means the implementation does not require Microsoft to function.
Agentic AI deployment at scale requires solving the enforcement problem before the autonomy problem. The autonomy problem has attracted more attention — agents that can do more things, faster, with less human oversight. The enforcement problem has been the thing enterprises cite when they say they are "piloting" but not shipping. This toolkit is a direct attempt to close that gap.
The public preview status is worth noting: Microsoft describes the packages as production-quality but subject to breaking changes before general availability. Enterprises adopting it for critical infrastructure should track the GitHub releases closely.