← Back to Paper List

MI9: An Integrated Runtime Governance Framework for Agentic AI

Charles L. Wang, Trisha Singhal, Ameya Kelkar, Jason Tuo
Barclays, Columbia University
arXiv (2025)
Agent Benchmark

📝 Paper Summary

Agentic AI Safety Runtime Governance AI Alignment
MI9 is a runtime framework that instruments existing agent systems to detect and contain emergent risks like goal drift and privilege escalation using telemetry-driven conformance rules.
Core Problem
Pre-deployment alignment methods (RLHF, Constitutional AI) cannot anticipate emergent runtime behaviors in agentic systems, such as recursive planning loops, goal drift, and dangerous tool chains.
Why it matters:
  • Traditional infrastructure monitoring (HTTP latency, etc.) misses cognitive processes, leaving governance violations like unauthorized goal revision invisible.
  • Static permission models (RBAC) fail when agents dynamically refine goals, potentially allowing a retail trading agent to escalate to institutional-level transactions.
  • Current benchmarks prioritize task completion over behavioral consistency, lacking mechanisms to intervene during dangerous autonomous operations.
Concrete Example: A trading agent authorized for small trades might autonomously shift its goal to 'portfolio optimization' and execute a sequence (research -> consultation -> risk assessment -> trade) that bypasses required dual-control approvals. Traditional governance sees valid individual steps but misses the temporal policy violation.
Key Novelty
Integrated Runtime Governance Layer (MI9)
  • Instruments existing agent stacks with a standardized telemetry schema (ATS) that captures 'cognitive events' (planning, goal setting) alongside standard actions.
  • Enforces safety via continuous authorization monitoring that adapts permissions based on real-time behavior and goal context, rather than static roles.
  • Uses finite-state machines to validate temporal behavioral patterns and statistical drift detection to identify when agents deviate from their baselines under specific goals.
Evaluation Highlights
  • Achieves high detection rates with low False Positive Rates (FPR) across 1,000 diverse multi-domain synthetic scenarios.
  • Demonstrates effective graduated containment, preserving operational continuity by restricting specific tools or blocking planning cycles rather than abrupt termination.
Breakthrough Assessment
8/10
Significant step forward in operationalizing AI safety. Moves beyond static guardrails to dynamic, context-aware runtime governance, addressing critical enterprise needs for agentic deployment.
×