Do we have the illusion of autonomy? This may come off as a critical analysis of governance gaps in the model context protocol era. When OpenAI’s CEO Sam Altman announced that AI agents would soon handle your driver’s license renewal, the promise was simple: convenience without complexity. But beneath this glossy narrative lies a more sobering reality—one where the architecture of memory itself becomes a mechanism of control, and where the protocols that enable agent autonomy simultaneously entrench new forms of digital dependency.
The Model Context Protocol (MCP) has been heralded as the “USB-C for AI”—a universal standard that will democratize access and enable seamless interoperability. But this framing obscures a fundamental question: who controls the connective tissue? While policy discussions have focused on visible harms like privacy breaches and security vulnerabilities, they’ve largely ignored the structural dynamics that determine how power accumulates within these systems.
This analysis goes deeper than existing frameworks to examine three critical blind spots in current AI agent governance discourse: the architectural determinism embedded in MCP’s design, the false dichotomy between privacy and functionality, and the geopolitical dimensions of protocol standardization. By scrutinizing what the current conversation misses, we can chart a path toward governance that addresses not just symptoms, but root causes.
The Architecture of Orchestration: More Than Just a Protocol
What MCP Actually Does (And Doesn’t Do)
Most analyses treat MCP as neutral infrastructure—a simple coordination layer that routes information between services. This is dangerously incomplete. MCP is not just a protocol; it’s an architectural paradigm that fundamentally restructures how data flows, who controls access, and where power concentrates.
At the technical level, MCP operates through a client-server architecture where:
- Clients (AI agents like Claude or ChatGPT) initiate requests
- Servers (external tools and services) respond with capabilities and data
- Resources (files, databases, APIs) become dynamically discoverable
- Prompts (pre-defined templates) guide agent-tool interactions
- Tools (executable functions) enable agents to perform actions
But here’s what existing analyses miss: this architecture creates a fundamental asymmetry in data visibility and control. The orchestration layer—the component that manages these interactions—becomes a privileged vantage point from which all data flows are visible, even if individual services cannot see beyond their own interactions.
Consider a concrete example: You ask your AI agent to “schedule a doctor’s appointment that works with my calendar and insurance.” The agent must:
- Query your calendar API for availability
- Access your insurance provider’s database for network doctors
- Cross-reference medical specialty databases
- Integrate transportation routing from mapping services
- Store the resulting appointment details across multiple systems
Each individual service sees only its slice of this transaction. But the orchestration layer sees everything. It knows your health condition (inferred from specialty), your location patterns (from calendar and maps), your financial constraints (from insurance data), and your behavioral patterns (from how you prioritize appointments).
This is not a bug—it’s the inherent structure of MCP-based orchestration. And current governance frameworks do not adequately address it.
The Hidden Layer: Orchestrator Privilege
The concept of “orchestrator privilege” deserves deeper examination because it represents a novel form of information asymmetry in computing. Unlike traditional client-server models where data moves between discrete endpoints, or even cloud computing where data is centralized but compartmentalized, MCP-based systems create a transient omniscience—a moment where disparate data streams converge within a single computational context.
This has three critical implications that existing analyses overlook:
First: Context collapse without consent. Traditional privacy frameworks assume users can grant granular permissions to specific services for specific purposes. But when an orchestrator dynamically assembles context from multiple sources, the resulting data aggregate was never explicitly authorized. You consented to share your calendar with the scheduling app and your insurance info with the healthcare portal, but you never consented to having these cross-referenced to infer your health status and schedule patterns.
Second: Inference amplification. Because orchestrators have access to multiple data streams simultaneously, they can make inferences that no individual service could. A calendar app knows you have frequent appointments; a pharmacy app knows you refill certain prescriptions; a mapping app knows you visit medical facilities. Separately, these reveal little. Together, they construct a detailed health profile. The orchestrator is uniquely positioned to perform this synthesis, creating emergent sensitive information from nominally non-sensitive inputs.
Third: Accountability diffusion. When something goes wrong—a misused data point, an erroneous inference, a privacy breach—who is responsible? The orchestrator can claim it was simply routing data between services. Individual services can claim they operated within their permissions. The result is a responsibility gap where harm occurs but no single entity is clearly accountable.
The False Promise of Stateless Protocols
MCP’s technical specification emphasizes that it is “stateless” and “ephemeral”—that it doesn’t store data, merely routes it. This framing is technically accurate but pragmatically misleading. The question is not whether the protocol itself retains state, but whether the systems built atop it create persistent dependencies that functionally recreate vendor lock-in despite technical interoperability.
Consider the “context flywheel” dynamic. While MCP enables agents to technically switch between providers, the accumulated context—preferences, historical interactions, learned behavioral patterns—becomes a proprietary asset. Even if OpenAI and Anthropic both support MCP, moving from ChatGPT to Claude means starting from scratch. Your agent forgets who you are.
This creates what economists call a “switching cost” that has nothing to do with technical compatibility and everything to do with informational capital. The more context your agent accumulates, the more valuable it becomes, and the higher the cost of migration. MCP’s interoperability does not solve this problem—it may actually exacerbate it by making data collection more seamless, thereby accelerating context accumulation.
Privacy: Beyond Consent Theater
The Fallacy of Informed Consent in Orchestrated Systems
Current privacy governance relies heavily on informed consent as its foundational mechanism. The EU’s GDPR, California’s CCPA, and most emerging AI regulations assume that users can meaningfully understand and authorize data uses. But MCP-based agents expose this assumption as fiction.
The problem is not simply that consent forms are too complex or that users don’t read terms of service. The problem is structural: in orchestrated systems, the data use cannot be specified in advance because it emerges dynamically from user interactions.
When you ask your agent to “plan a weekend trip within budget,” the specific services it will query, the data it will access, and the inferences it will make depend on contextual factors that cannot be enumerated at consent time. Will it check your calendar? Almost certainly. Your banking transactions to infer budget? Probably. Your social media to understand preferences? Maybe. Your health records to ensure accommodations? Potentially.
The set of possible data flows is combinatorially large. Pre-specifying them all would result in consent requests so expansive as to be meaningless. And yet, allowing blanket authorization defeats the purpose of consent entirely.
This is not a problem that can be solved through better UX or clearer privacy policies. It is an architectural incompatibility between the dynamism required for effective agent orchestration and the specificity required for meaningful consent.
Dynamic Privacy: A New Framework
What’s needed is a shift from static consent to dynamic authorization—a system where permissions are granted contextually, evaluated in real-time, and revocable at any moment. This requires several technical and governance innovations that current frameworks don’t adequately address:
Contextual Permission Tokens: Instead of granting broad access to services, users should issue short-lived, purpose-specific tokens that encode exactly what data can be accessed for what purpose. For example: “Access my calendar for availability from 9am-5pm on weekdays for medical appointments only, valid for the next 30 minutes.”
Real-Time Consent Engines: Before each data access, the orchestrator should query a user-controlled consent engine that evaluates whether the access aligns with current permissions. This is computationally expensive but technically feasible using modern cryptographic protocols like homomorphic encryption and secure multi-party computation.
Revocable Memory: Users should be able to retroactively revoke not just future access, but past inferences. If an agent inferred your health status from calendar and pharmacy data, you should be able to delete that inference even after it was made. This requires treating inferences as first-class data objects with their own lifecycle management.
Audit Trails as a Right: Users should have access to detailed logs of every data access, every inference made, and every action taken by their agents. This is not just a matter of transparency—it’s a precondition for accountability. Without audit trails, users cannot verify whether agents respected their permissions or identify when breaches occurred.
These technical mechanisms are possible today but not required by current regulations. That gap represents a significant governance failure.
Cross-Border Data Flows: The Jurisdictional Maze
One of the most underdiscussed challenges in AI agent governance is the collision between protocol-based data flows and jurisdiction-based privacy laws. MCP enables agents to seamlessly query services hosted in different countries, but privacy regulations are territorially bounded.
Consider an American user asking their agent to book a European hotel. The agent might:
- Query a US-based travel site (subject to CCPA)
- Access hotel databases in the EU (subject to GDPR)
- Process payments through an Asian payment processor (subject to varying local laws)
- Store confirmation details on a cloud service with servers globally distributed
Which law governs this transaction? Under GDPR, EU data protection rules apply whenever EU residents’ data is processed, regardless of where the company is located. But the user is American, the orchestrator is operated by a US company, and the individual services may have different privacy policies.
Current frameworks address this through data transfer mechanisms like Standard Contractual Clauses (SCCs) and Privacy Shield agreements. But these were designed for relatively static data relationships—a US company contracting with an EU processor. They don’t map well onto dynamic, orchestrated data flows where services are assembled on-the-fly based on user requests.
Moreover, the compliance burden is increasingly untenable. A small developer building an MCP-compatible tool may suddenly find their service being invoked by agents processing data subject to dozens of different regulatory regimes. The liability exposure could be enormous, creating a chilling effect on innovation and favoring large incumbents who can afford legal teams to navigate this complexity.
The governance gap here is not just about harmonizing privacy laws—it’s about developing jurisdictional frameworks that can accommodate dynamic, protocol-mediated data flows. This likely requires international coordination on a scale not yet attempted in technology policy.
Security: The Attack Surface You Can’t See
Prompt Injection: The SQL Injection of the AI Era
Security researchers have identified prompt injection attacks—where malicious instructions embedded in data cause an agent to behave in unintended ways—as a critical threat to AI systems. But most analyses treat this as a model-level vulnerability that can be fixed through better training or input sanitization.
This misunderstands the nature of the threat. Prompt injection is not merely a technical bug—it’s an inevitable consequence of the way LLM-based agents process information.
Language models are trained to follow instructions embedded in text. This is a feature, not a bug. But it means they cannot reliably distinguish between instructions from trusted sources (the user) and instructions from untrusted sources (external data). When an agent reads an email, a website, or a database entry, it processes that content in the same cognitive context as the user’s prompt.
This creates an intrinsic vulnerability: any external data that an agent processes could potentially contain adversarial instructions. And in MCP-based systems, agents routinely ingest data from dozens of external sources—emails, calendars, web scraping results, API responses. Each is a potential vector for prompt injection.
The 2024 “Echoleak” incident demonstrated this vividly. A carefully crafted email caused an AI agent to expose private information from previous conversations. The attack worked because the agent treated the email content and the user’s chat history as part of the same context window. There was no cryptographic or access-control failure—the agent was simply doing what it was designed to do: following instructions it found in data.
Current defenses against prompt injection are inadequate. Techniques like instruction segregation (using special tokens to demarcate trusted instructions) and output filtering (scanning agent responses for leaked information) are partial mitigations at best. They don’t address the fundamental problem: that language models lack a built-in concept of trust boundaries.
What’s needed is a context isolation architecture where:
- User instructions and external data are processed in cryptographically separated contexts
- Information flow between contexts is mediated by explicit policies
- Agents can be formally verified to respect these boundaries
This is an active research area, but it remains largely absent from policy discussions. Until we have robust technical solutions, prompt injection represents an existential risk for any high-stakes deployment of AI agents.
The Supply Chain Problem: MCP as a Single Point of Failure
Most security analyses focus on individual components—the model, the data, the application. But MCP introduces a new systemic risk: the orchestration protocol itself becomes a critical dependency that, if compromised, can affect all agents that rely on it.
Consider the software supply chain attacks that have plagued open-source ecosystems. In 2021, a vulnerability in Log4j—a widely used Java logging library—affected thousands of applications worldwide. The impact was so severe because Log4j was a foundational dependency that many developers didn’t even realize they were using.
MCP is poised to become the Log4j of AI agents. As it becomes the de facto standard for agent-tool integration, more and more systems will depend on its security properties. But MCP is still in early development. Its specification is evolving, implementations vary, and security audits are limited.
What happens when a vulnerability is discovered in MCP itself? Coordinating patches across the entire ecosystem of agents, orchestrators, and services could take months. And because many implementations are closed-source, some vendors may not even disclose that they’re affected.
Moreover, the decentralized nature of MCP deployment makes it difficult to enforce security standards. Unlike a centralized API where a platform can mandate security requirements, MCP allows any developer to build a compliant server. There’s no gatekeeper ensuring that every MCP-compatible service follows security best practices.
The governance implications are profound. We need:
- Formal security standards for MCP implementations, backed by certification programs
- Vulnerability disclosure frameworks that coordinate across the AI agent ecosystem
- Supply chain transparency tools that let users see which MCP servers their agents are connecting to
- Incident response protocols that can rapidly disseminate patches when vulnerabilities are discovered
These exist in other critical infrastructure domains (e.g., financial systems, power grids) but are largely absent in AI agent governance.
Authentication Without Identity: The KYA Problem
Current MCP implementations lack standardized mechanisms for authenticating agents or establishing trust relationships between services. This is not an oversight—it’s a deliberate design choice to keep the protocol simple and flexible. But it creates serious security vulnerabilities.
When a service receives a request via MCP, it often cannot verify:
- Whether the request actually came from the claimed agent
- Whether the agent is authorized to act on behalf of the user
- Whether the orchestrator properly validated permissions
This “authentication gap” enables several classes of attacks:
Agent Impersonation: A malicious actor could create a fake agent that claims to be Claude or ChatGPT, gaining unauthorized access to services that trust those brands.
Privilege Escalation: An agent with limited permissions could exploit the authentication gap to access services it shouldn’t be able to reach.
Man-in-the-Middle Attacks: An attacker could intercept MCP communications and inject their own requests, appearing to originate from legitimate agents.
The proposed solution—Know-Your-Agent (KYA) requirements analogous to financial KYC rules—is a step in the right direction. But it raises difficult questions:
- Who issues agent credentials?
- How are they verified across jurisdictions?
- What happens when an agent is compromised?
- How do we balance identity verification with privacy?
These questions have no easy answers, but they must be addressed before MCP-based agents can be safely deployed in high-stakes contexts like healthcare, finance, or government services.
Power and Competition: The Moat You Can’t See
Context as Capital: The New Lock-In
Perhaps the most significant blind spot in current AI agent governance is the failure to recognize context accumulation as a form of market power. While technical interoperability receives significant attention, the economics of context portability remain largely unexplored.
Here’s why this matters: In traditional software markets, vendor lock-in occurs through proprietary data formats, incompatible APIs, or network effects. You stay with Microsoft Office because your files are in .docx format; you stay with iPhone because your apps and data are in Apple’s ecosystem.
AI agents introduce a new form of lock-in: accumulated context. The more your agent knows about you—your preferences, your communication style, your decision-making patterns, your relationships—the more valuable it becomes. This context is not a static dataset; it’s a dynamic model of you that improves with every interaction.
Critically, this context is not interoperable even when the protocol is. MCP standardizes how agents connect to services, but it doesn’t specify how context is represented, stored, or transferred. As a result:
- Moving from one agent provider to another means losing all your accumulated context
- Each provider uses proprietary formats for storing and representing user models
- Context becomes a form of informational capital that accrues to the platform, not the user
This is a subtle but powerful form of lock-in because it doesn’t prevent migration technically—you can switch agents—but it makes migration prohibitively costly from a user experience perspective. Starting over with a new agent means months of re-training, correcting mistakes, and rebuilding preferences.
The Concentration Dynamics of Protocol Control
While MCP is theoretically an open protocol, its development and governance are heavily concentrated. Anthropic (Claude’s creator) designed MCP and maintains the reference implementation. OpenAI, Google, and Microsoft have adopted it, but they also exert significant influence over its evolution.
This creates a protocol oligopoly where a handful of powerful actors effectively determine how agent orchestration works for the entire ecosystem. The parallels to early internet protocols are instructive:
In the 1990s, HTTP and HTML were developed through relatively open processes (W3C, IETF) with broad participation from academics, nonprofits, and companies. This openness helped prevent any single entity from dominating web standards.
In contrast, MCP’s development has been dominated by commercial AI companies with strong incentives to shape the protocol in ways that benefit their business models. There’s no independent standards body, no formal public comment process, and limited participation from civil society or academic researchers.
The risk is that MCP will encode the priorities of large AI companies rather than users or society broadly. We’re already seeing this in design choices that prioritize functionality over privacy, convenience over security, and growth over equity.
For example:
- MCP makes it easy to add new data connections but provides limited mechanisms for data minimization
- The protocol emphasizes agent capabilities but not user control
- Integration with commercial services is prioritized over integration with user-owned infrastructure
These aren’t necessarily malicious choices—they reflect the incentives and perspectives of the companies building the protocol. But they have profound implications for how power is distributed in the AI agent ecosystem.
Data Portability: Necessary But Not Sufficient
Many governance proposals call for data portability requirements—mandating that users be able to export their data and switch providers without losing functionality. This is a good start, but it’s insufficient for AI agents because:
- Context is not just data—it’s a learned model of user behavior that may not be representable in standard formats
- Portability without standardization is meaningless—if every provider uses different context representations, exported data won’t work with new agents
- Porting context could amplify privacy risks—transferring a rich user model between providers expands the attack surface and introduces new breach vectors
What’s needed is a multi-layered portability framework that includes:
Data Portability (Layer 1): Users can export raw data (messages, documents, preferences) in standardized formats. This is the easiest layer and should be table stakes.
Context Portability (Layer 2): Users can export learned models and behavioral patterns. This requires standardized representations of agent memory, preference models, and interaction history. Organizations like the Partnership on AI or IEEE could develop these standards, but currently no such effort exists.
Capability Portability (Layer 3): Users can migrate not just data and context, but also custom tools, integrations, and workflows they’ve built. This is the hardest layer because it requires interoperability not just of protocols but of entire agent ecosystems.
Current governance proposals focus almost exclusively on Layer 1. But without Layers 2 and 3, data portability doesn’t meaningfully reduce lock-in for AI agents.
The Antitrust Implications Nobody’s Discussing
Traditional antitrust enforcement focuses on observable market behaviors: pricing, market share, exclusive dealing, predatory conduct. But the competitive dynamics of AI agents operate through more subtle mechanisms that existing frameworks struggle to capture:
Privileged Integration: Platform providers can give their own agents preferential access to data, faster response times, or deeper integration with services. This is difficult to detect and harder to prove as anticompetitive, but it creates structural advantages that competitors cannot match.
Context Barriers: As discussed, accumulated context creates switching costs. A dominant agent provider can leverage this to maintain market position even if competitors offer superior technology. This is anticompetitive but doesn’t violate traditional antitrust standards because no explicit exclusionary conduct occurs.
Protocol Capture: If a small number of companies effectively control MCP’s evolution, they can steer the protocol in ways that disadvantage competitors. This could include adding features that require resources only large players can provide, or creating technical barriers that favor established ecosystems.
Data Feedback Loops: Companies that operate both popular agents and popular services (e.g., Google Search + Google AI, Microsoft 365 + Copilot) can create closed loops where agent use generates data that improves services, which attracts more users, which generates more data. These feedback loops are self-reinforcing and difficult to break even with interoperability mandates.
These dynamics suggest that AI agent markets may naturally tend toward concentration even in the presence of technical interoperability. Preventing this requires governance mechanisms that go beyond traditional antitrust:
- Structural separation requirements preventing companies from both operating agents and controlling critical services
- Non-discrimination obligations mandating equal access to APIs and data for all agents
- Context portability mandates with technical standards enforced by regulators
- Algorithmic transparency requirements allowing auditors to verify that platforms aren’t favoring their own agents
These interventions would be controversial and face significant industry pushback. But without them, MCP’s openness may be illusory—a technical veneer over structural concentration.
International Dimensions: The Governance Fragmentation Problem
The US-EU Regulatory Divergence
While the original brief notes that the EU and China are advancing AI governance frameworks, it understates the significance of regulatory fragmentation for agent systems. The transatlantic divide over AI regulation represents not just different policy choices, but fundamentally incompatible visions of technology governance:
The EU’s Risk-Based Approach (AI Act):
- Categorizes AI systems by risk level (unacceptable, high, limited, minimal)
- Imposes heavy compliance burdens on “high-risk” systems
- Emphasizes ex-ante regulation (requirements before deployment)
- Prioritizes fundamental rights and democratic values
The US’s Innovation-First Approach:
- Relies primarily on existing sectoral regulations (finance, healthcare, etc.)
- Emphasizes voluntary standards and industry self-governance
- Favors ex-post enforcement (addressing harms after they occur)
- Prioritizes economic competitiveness and technological leadership
For AI agents, this divergence creates serious practical challenges:
Scenario: A US-based agent provider operates in Europe
Under the AI Act, their agent orchestration system likely qualifies as “high-risk” because it:
- Makes consequential decisions affecting users’ access to services
- Processes sensitive personal data across multiple contexts
- Operates with significant autonomy and limited human oversight
This triggers requirements for:
- Risk assessment and mitigation documentation
- Dataset governance and quality standards
- Technical documentation and logging
- Human oversight mechanisms
- Transparency and information to users
- Cybersecurity and robustness testing
Many of these requirements are vague or technically challenging to implement for orchestrated systems. How do you document “dataset governance” when your agent dynamically queries hundreds of external databases? How do you provide “human oversight” for interactions that happen in milliseconds across distributed services?
The compliance burden is significant enough that some US companies may simply choose not to operate in Europe, or to offer degraded functionality that avoids triggering AI Act requirements. This creates a fragmented global market where agent capabilities differ by jurisdiction.
China’s Algorithmic Regulation: A Third Model
China’s approach to AI governance introduces yet another model—one that emphasizes state control over algorithmic systems through mechanisms that are largely foreign to Western governance frameworks:
- Algorithm Registration: Companies must register major algorithms with regulators and disclose their training data, model architecture, and decision-making logic
- Content Control: AI systems must incorporate censorship mechanisms to prevent generation of politically sensitive content
- Data Localization: Personal data and important data must be stored domestically and cannot be transferred abroad without approval
- Socialist Values: AI systems must align with “socialist core values” and not undermine social stability
For agent systems operating globally, Chinese requirements create fundamental tensions:
An agent that works in the US or EU may be technically incompatible with Chinese requirements. For example:
- Transparency requirements in the EU may conflict with China’s expectation that certain algorithmic details remain confidential to prevent circumvention of content controls
- Data portability mandates in Western jurisdictions may conflict with China’s data localization requirements
- Freedom of expression principles in the West may conflict with China’s content restrictions
The result is likely to be regime-specific agent variants—systems that behave fundamentally differently depending on where they operate. This undermines the promise of universal AI agents that can seamlessly assist users globally.
The Race to Set Global Norms
Despite regulatory divergence, there’s a fierce competition to establish norms that will shape global AI governance:
The EU’s Regulatory Export Strategy:
- The “Brussels Effect”: EU regulations become de facto global standards because companies design for the strictest regime
- AI Act as template: Other jurisdictions (UK, Canada, Brazil) are considering similar frameworks
- Digital sovereignty: Positioning Europe as the ethical alternative to US tech dominance and Chinese state control
The US’s Standards-Based Strategy:
- NIST AI Risk Management Framework as voluntary best practices
- Industry leadership through OpenAI, Anthropic, Google setting technical standards
- Alliance building with democracies to create alternative governance bloc to China
China’s Belt-and-Road Approach:
- Exporting AI systems and governance models to developing countries
- Using economic leverage to promote Chinese technical standards
- Positioning algorithmic governance as compatible with diverse political systems
For AI agents specifically, this geopolitical competition plays out in battles over:
- Which protocols become standard (MCP vs. Chinese alternatives)
- What security requirements apply (end-to-end encryption vs. state access)
- How identity works (anonymous vs. real-name registration)
- Who can operate agents (open market vs. licensed providers)
The stakes are high because these technical choices encode political values. An agent ecosystem built on MCP with Western security models looks fundamentally different from one built on protocols designed for state supervision.
Governance Arbitrage and the Race to the Bottom
The existence of multiple regulatory regimes creates opportunities for governance arbitrage—structuring operations to minimize regulatory compliance costs. Companies can:
- Locate orchestration infrastructure in lenient jurisdictions
- Route data flows through privacy havens
- Segment users by geography to avoid triggering high-risk classifications
- Use technical obfuscation to make cross-border data flows difficult to audit
This is not merely theoretical. We’ve seen similar dynamics in:
- Tax havens: Corporations routing profits through low-tax jurisdictions
- Data havens: Companies locating servers in countries with weak privacy laws
- Content moderation: Platforms applying different speech rules by country
For AI agents, governance arbitrage is even easier because:
- Services are purely digital with minimal physical infrastructure
- Data flows are encrypted and difficult to inspect
- Orchestration can be split across multiple jurisdictions
- Users often don’t know or care where their agent is operated from
Without international coordination, the risk is a race to the bottom where agents are operated from whichever jurisdiction imposes the fewest restrictions. This would undermine all national governance efforts, as even strict regulations become unenforceable if agents can simply relocate.
What’s needed is a new international framework for AI governance—something analogous to the Basel Accords for banking or ICAO standards for aviation. This would establish:
- Minimum safety and security standards for agent systems
- Mutual recognition of compliance certifications across borders
- Coordinated enforcement mechanisms for cross-border violations
- Technical standards for audit and transparency that work globally
Achieving this requires political will that currently doesn’t exist. But without it, the promise of safe, trustworthy AI agents remains elusive.
Critical Analysis: What the Current Discourse Gets Wrong
Blind Spot #1: The Assumption of Technological Neutrality
The most fundamental error in current AI agent governance discourse is the assumption that protocols like MCP are neutral infrastructure analogous to roads or electrical grids. This framing appears throughout the literature: MCP as “USB-C for AI,” as “plumbing,” as “connective tissue.”
This is a dangerous category error. Unlike physical infrastructure, digital protocols encode choices about information flow, access control, and power distribution. These choices are not neutral—they reflect the values and incentives of their designers.
Consider what MCP prioritizes:
- Capability over control: The protocol makes it easy to add new agent capabilities but harder to restrict them
- Convenience over privacy: Data flows are optimized for seamless user experience, not data minimization
- Interoperability over security: Broad compatibility is prioritized over hardened security boundaries
These aren’t inevitable technical requirements—they’re design decisions that could have been made differently. An alternative protocol could prioritize privacy by default, security over convenience, and user control over agent capability.
The myth of neutrality is dangerous because it frames governance as simply managing the effects of inevitable technology, rather than shaping the technology itself. It suggests we must accept MCP’s architecture and merely mitigate its harms, rather than demanding different architectural choices.
Effective governance requires rejecting technological neutrality and recognizing that protocol design is a political choice that distributes power. Regulators should not merely regulate uses of MCP—they should shape MCP itself through technical requirements that encode public interest values.
Blind Spot #2: The Individual Rights Paradigm
Current privacy governance operates almost exclusively through individual rights frameworks: consent, access, correction, deletion. Users are given rights to control their personal data, and privacy is understood as individuals protecting themselves from institutional surveillance.
This paradigm is increasingly inadequate for AI agents because the harms are often collective and systemic rather than individual and particularized:
Inference Amplification: When agents make inferences by combining data from many users, the resulting insights affect entire populations. My decision to share health data with my agent may enable inferences about others with similar profiles, even if they never shared their data.
Market Concentration: When context accumulation creates lock-in, the harm is not to individual users (who may love their agent) but to the competitive structure of the market. This is a collective harm that individual rights can’t address.
Epistemic Pollution: When agents increasingly mediate our access to information, they shape our understanding of reality. If they’re trained on biased data or designed with particular worldviews, the harm is to the epistemic commons—our collective ability to reason about the world. Individual users can’t opt out of this harm.
Democratic Capture: When agents become infrastructure for civic life (voting, organizing, accessing government services), their design affects democratic functioning. The harm is to democratic legitimacy itself, not to individuals.
These collective harms require governance mechanisms beyond individual rights:
- Impact assessments that evaluate effects on populations, markets, and institutions
- Structural regulations that constrain system design, not just individual uses
- Collective decision-making where affected communities have voice in technology design
- Public interest oversight that represents societal values, not just user preferences
The individual rights paradigm isn’t wrong—it’s insufficient. Effective AI agent governance requires supplementing it with collective governance mechanisms that address systemic harms.
Blind Spot #3: The Competence Assumption
Much of the policy discourse assumes that governance institutions are capable of understanding and overseeing AI agent systems. Regulations are proposed, enforcement is promised, and compliance is expected.
This is wishful thinking. The reality is that:
Regulators Lack Technical Capacity: Most government agencies don’t employ engineers who understand modern AI systems, much less the complexities of agent orchestration. They rely on industry self-reporting and lack the technical tools to verify compliance.
The Technology Evolves Faster Than Regulation: By the time regulations are written, debated, and implemented, the technology has moved on. MCP barely existed two years ago; now it’s being treated as established infrastructure. Regulation operates on 5-10 year timescales; AI on 6-12 month timescales.
The Problem Space Is Genuinely Difficult: Even experts disagree about fundamental questions: What is an “agent” vs. a “tool”? When does orchestration become “autonomous”? What constitutes “meaningful” consent? These aren’t political questions with clear answers—they’re conceptual puzzles that require ongoing research.
Compliance Is Performative: Companies have become expert at regulatory compliance theater—checking boxes, issuing reports, and creating impressive-looking governance structures that don’t actually constrain harmful behavior. Regulators lack the capacity to distinguish genuine compliance from performance.
Acknowledging these limitations doesn’t mean giving up on governance. It means pursuing strategies that account for limited state capacity:
Liability Regimes: Instead of prescriptive rules that require enforcement, create liability frameworks where harms trigger meaningful consequences. This leverages private litigation to supplement limited regulatory capacity.
Mandatory Disclosure: Require publication of technical documentation, audit logs, and impact assessments. This creates transparency that empowers researchers, journalists, and civil society to identify problems even when regulators can’t.
Bounty Programs: Offer rewards for discovering security vulnerabilities or governance failures. This crowdsources enforcement and incentivizes independent scrutiny.
Regulatory Sandboxes: Create safe spaces for experimentation where new agent systems can be tested under supervision before broad deployment. This gives regulators hands-on learning opportunities.
Multistakeholder Governance: Include technical experts, civil society, and affected communities in oversight bodies. This supplements limited government capacity with distributed expertise.
The key insight is that effective governance doesn’t require regulators to be as technically sophisticated as developers—it requires governance mechanisms that work even with limited institutional capacity.
A Path Forward: Governance That Addresses Root Causes
Principle 1: Architectural Accountability
Rather than focusing solely on use cases and applications, governance must engage with protocol design as a site of political intervention. This means:
Technical Standards With Teeth: Not voluntary guidelines but mandatory requirements for MCP implementations:
- Cryptographic logging of all data accesses with user-accessible audit trails
- Formal verification that agents respect specified permission boundaries
- Standardized formats for context representation to enable portability
- Rate limiting and access controls at the protocol level
- Mandatory security certifications before agents can access sensitive services
Participatory Protocol Design: Opening MCP’s evolution to broader stakeholder input:
- Public comment periods on proposed protocol changes
- Representation for civil society and academics in governance bodies
- Impact assessments before major architectural decisions
- Open-source reference implementations that anyone can audit
Accountability by Design: Building oversight mechanisms into the protocol itself:
- Agents must expose why they made particular choices (explainability)
- Services must log what data they provided to agents (provenance)
- Users must be able to challenge agent decisions (contestability)
- Orchestrators must demonstrate compliance with permissions (accountability)
This represents a fundamental shift from regulating agents to regulating the infrastructure that enables agents. It’s more difficult politically but more effective technically.
Principle 2: Context as a Regulated Asset
Recognition that context is capital requiring special governance:
Context Portability Standards: Regulatory mandates for standardized context representations:
- User behavior models must be exportable in machine-readable formats
- Preference systems must use standardized ontologies
- Interaction histories must include full provenance of inferences
- Third-party services must accept imported context without discrimination
Context Minimization Requirements: Limits on what context can be collected and retained:
- Agents can only remember data necessary for explicitly requested functions
- Sensitive context (health, finances, beliefs) subject to strict time limits
- Users must be able to selectively delete context categories
- Context cannot be used for purposes beyond those consented to
Context Auditing: Independent verification that context handling complies with regulations:
- Accredited auditors with technical capability to assess implementations
- Regular audits required for high-risk agent systems
- Public reporting of audit findings (with privacy protections)
- Enforcement actions for failures to protect context
Anti-Lock-In Provisions: Preventing context accumulation from creating monopolies:
- Interoperability requirements for context formats
- Prohibitions on exclusive context-service relationships
- Mandated open APIs for context export and import
- Antitrust scrutiny of acquisitions that consolidate context
This treats context as a form of essential infrastructure that, like railroads or telecommunications networks, requires special regulation to prevent monopolization.
Principle 3: Collective Governance Mechanisms
Moving beyond individual rights to collective decision-making about agent systems:
Participatory Technology Assessment: Structured processes for public input on AI agent deployments:
- Community forums where affected populations can voice concerns
- Expert panels that translate technical details into policy implications
- Citizen juries that deliberate on whether deployments serve public interest
- Binding recommendations that developers must address or publicly justify ignoring
Sector-Specific Governance Councils: Domain experts overseeing agent use in critical areas:
- Healthcare: physicians, ethicists, patient advocates overseeing medical agents
- Education: teachers, researchers, students governing educational agents
- Finance: regulators, consumer advocates, economists overseeing financial agents
- Public services: civil servants, advocates, citizens governing government agents
Algorithmic Impact Assessments: Comprehensive evaluation of agent systems before deployment:
- Technical assessment: security, privacy, robustness
- Social assessment: equity, accessibility, discrimination
- Economic assessment: market effects, employment impacts, power concentration
- Democratic assessment: effects on information, organizing, civic participation
- Results must be published, and deployment prohibited if harms outweigh benefits
Ongoing Monitoring and Adjustment: Recognition that governance must be dynamic:
- Continuous auditing of deployed systems for emergent harms
- Rapid response mechanisms to address problems as they arise
- Regular reassessment of whether systems still serve public interest
- Sunset provisions requiring re-authorization after specified periods
These mechanisms shift power from developers and platforms to affected communities and the public interest. They recognize that those who bear risks from technology should have voice in how it’s designed and deployed.
Principle 4: International Coordination
Creating governance structures that transcend national boundaries:
Multilateral Standards Bodies: International organizations developing binding standards:
- Technical standards for agent security, privacy, and safety
- Common definitions and classifications of agent capabilities and risks
- Mutual recognition of compliance certifications across borders
- Harmonized enforcement mechanisms for cross-border violations
Data Governance Treaties: International agreements on data flows and sovereignty:
- Permitted uses of data in agent orchestration
- Requirements for cross-border data transfers
- Rights of data subjects regardless of jurisdiction
- Mechanisms for dispute resolution when laws conflict
Research Collaboration: Sharing knowledge about agent risks and mitigation:
- Joint research programs on agent safety and security
- Common repositories of vulnerability disclosures
- International incident response coordination
- Collaborative development of best practices
Regime Diversity Management: Acknowledging that full harmonization is impossible:
- Technical standards that accommodate different governance approaches
- Interoperability frameworks that work across regime types
- Clear labeling so users know which governance regime applies
- Prohibitions on governance arbitrage to avoid regulations
This is admittedly ambitious—international technology governance has a poor track record. But the alternative is regulatory fragmentation that makes effective governance impossible.
From Inevitability to Intentionality
The conventional narrative treats AI agents as an inevitable evolution of computing—a natural next step from calculators to personal computers to smartphones to autonomous assistants. This framing is both descriptively false and normatively pernicious.
It’s descriptively false because technology trajectories are shaped by choices, not natural laws. The way agents work today—MCP’s architecture, context accumulation dynamics, orchestration patterns—reflects decisions made by particular actors with particular incentives. Different choices would have produced different systems.
It’s normatively pernicious because inevitability narratives disempower democratic governance. If agents are inevitable, the question becomes “how do we adapt?” rather than “what kind of agents do we want?” It shifts debate from fundamental questions about power and purpose to technical questions about risk mitigation.
Effective AI agent governance requires rejecting inevitability and embracing intentionality—the recognition that we can and must collectively choose what these systems become. This doesn’t mean halting innovation or imposing rigid controls. It means ensuring that agent development serves human flourishing rather than narrow commercial interests.
The moment is urgent because architectural choices made now will be difficult to reverse later. Once MCP becomes entrenched as the de facto standard, once billions of users accumulate context with specific providers, once markets concentrate around dominant platforms, the opportunity for meaningful intervention shrinks dramatically.
We’ve seen this pattern before—with social media, with mobile ecosystems, with cloud computing. In each case, early governance failures created path dependencies that proved extremely difficult to correct. We let platforms become “too big to fail” and then found ourselves negotiating with them from positions of weakness.
AI agents present an opportunity to do better. The technology is early enough that fundamental choices remain open. The governance mechanisms discussed here—architectural accountability, context regulation, collective oversight, international coordination—are still possible to implement. But the window is closing.
The question is not whether AI agents will become infrastructure for modern life—they likely will. The question is whose infrastructure they’ll be. Will they serve users or platforms? Democratic values or surveillance capitalism? Human autonomy or algorithmic control?
The answer depends on choices we make now—not as inevitable adaptations to technological progress, but as intentional expressions of the future we want to build.
References and Further Reading
Technical Background:
- Anthropic. “Model Context Protocol.” https://www.anthropic.com/news/model-context-protocol
- OpenAI. “Introducing ChatGPT Agent.” https://openai.com/index/introducing-chatgpt-agent/
- Partnership on AI. “Preparing for AI Agent Governance.” https://partnershiponai.org/resource/preparing-for-ai-agent-governance/
Privacy and Security:
- Whittaker, M. “AI Agents Are Coming for Your Privacy.” The Economist, 2025.
- Echoleak Incident Analysis. Feedly CVE-2025-32711.
- Future Society. “Ahead of the Curve: Governing AI Agents Under the EU AI Act.” 2025.
Power and Competition:
- AI Frontiers. “Open Protocols Prevent AI Monopolies.” https://ai-frontiers.org/articles/open-protocols-prevent-ai-monopolies
- Tech Policy Press. “Whom Does AI Agentiality Serve?” https://www.techpolicy.press/new-perspectives-on-ai-agentiality-and-democracy-whom-does-it-serve/
International Governance:
- European Commission. “EU AI Act.” Official Journal, 2024.
- DigiChina. “China’s Pioneering Draft Algorithm Regulations.” Stanford, 2025.
- NIST. “AI Risk Management Framework.” https://www.nist.gov/itl/ai-risk-management-framework
Critical Perspectives:
- Knight First Amendment Institute. “AI Agents and Democratic Resilience.” Columbia University, 2025.
- Center for Democracy & Technology. “AI Systems and Personalization.” https://cdt.org/insights/
- Data Transfer Initiative. “Path Forward for AI Portability.” 2025.