Skip to content

MCP Server Data Retention Policies Compared: Which Platforms Keep Your Data? (2026)

Compare MCP server data retention policies across Merge, Composio, StackOne, and Truto. Learn which platforms store your customers' data at rest and which offer true zero-retention architecture.

Nachi Raman Nachi Raman · · 14 min read
MCP Server Data Retention Policies Compared: Which Platforms Keep Your Data? (2026)

If you are a product manager or engineering leader evaluating platforms to host MCP servers for your AI agents, the most critical question is not about tool variety or connection speed. It is about data retention. If your AI agent connects to Salesforce, BambooHR, or Jira through an MCP server, the data flowing through that server is governed by that server provider's data retention policy - not yours.

Every record your agent reads or writes passes through a third-party layer. If the infrastructure routing that request stores a copy of those contacts on its own disks, they become a sub-processor of your customers' highly regulated enterprise data. Your SOC 2 scope immediately expands, your GDPR obligations multiply, and enterprise InfoSec teams will flag your application during procurement.

OpenAI explicitly states in its data controls documentation: "MCP servers (used with the remote MCP server tool) are third-party services, and data sent to an MCP server is subject to their data retention policies." The major LLM providers wash their hands of what happens after the tool call is dispatched. Vendor selection is entirely on your shoulders.

The timing matters. Gartner predicts that 50% of all enterprise cybersecurity incident response efforts will focus on incidents involving custom-built AI-driven applications by 2028. According to Gartner VP Analyst Christopher Mixter, "AI is evolving quickly, yet many tools - especially custom-built AI applications - are being deployed before they're fully tested. These systems are complex, dynamic and difficult to secure over time."

The compliance exposure is direct: Gartner further projects that manual AI compliance processes will expose 75% of regulated organizations to fines exceeding 5% of global revenue by the end of 2027. If you're building AI agents that touch employee records, customer PII, or financial data, the platform you route that data through becomes a sub-processor in your compliance chain.

The Hidden Compliance Cost of MCP Servers

The Model Context Protocol (MCP) solves a massive integration bottleneck. Instead of writing custom API connectors for every LLM framework, you deploy a single MCP server that exposes your SaaS product's capabilities to Claude, ChatGPT, Cursor, and custom LangChain agents.

However, connecting an AI reasoning engine to enterprise systems like HRIS, CRM, and accounting platforms introduces severe security vectors. When an AI agent triggers a tool call - for example, list_hubspot_contacts - that request travels from the LLM client, through the MCP server, to the upstream API, and back.

If the platform hosting your MCP server caches that payload to normalize the data or log the transaction, it is writing your customer's PII to a database you do not control. You are now responsible for auditing that vendor's encryption standards, data lifecycle policies, and breach notification protocols. For B2B SaaS companies selling to the mid-market and enterprise, this single architectural detail can stall a six-figure deal for months.

Here is the question your InfoSec team will ask during procurement review: does this MCP server platform store our customers' data? If the answer is "yes" - even if that data is encrypted - you've just added a new sub-processor to your SOC 2 scope, triggered GDPR data processing agreement requirements, and created a data residency question you'll need to answer for every enterprise customer.

How Traditional Integration Platforms Handle Data (And Why It Breaks AI Compliance)

To understand why many integration platforms store data, you have to look at their original architecture. Most integration platforms were designed long before AI agents existed. Legacy unified APIs and embedded iPaaS platforms were designed for asynchronous ETL workflows, not real-time AI agent execution. Their architecture assumes that storing and caching third-party data is the right approach - it makes normalization easier, reduces upstream API calls, and simplifies pagination.

The Sync-and-Cache Architecture

Traditional unified APIs like Merge operate on a sync-and-cache model. To provide a standardized common data model across 50 different CRMs, these platforms cannot simply proxy requests on the fly. Instead, they periodically poll the upstream APIs (Salesforce, Workday, HubSpot, etc.), download the raw data, transform it into their standardized schema, and store it in their own massive multi-tenant databases. When you query their API, you are reading from their database, not the source system.

This architecture was practical when the use case was a backend integration feeding a dashboard. But when an AI agent pulls employee salary data through this layer, that salary data now sits in a third-party database you don't control.

Agent Execution Layers

Newer platforms built specifically for AI agents, such as Composio, position themselves as agent-first execution and authentication layers. They offer pre-built tool packs that agents can call directly. However, managing stateful agent execution, logging complex multi-step workflows, and handling authentication often requires persistent storage.

Warning

The "Encrypted at Rest" Trap When evaluating integration vendors for AI agents, do not confuse "encrypted at rest" with "secure architecture." When a vendor says data is "encrypted at rest," they're confirming two things: (1) they store your data, and (2) they encrypt it while stored. Encryption is a baseline requirement for storing data. For strict enterprise compliance, the requirement is zero data retention - meaning the vendor should not store the payload at rest in the first place.

Under GDPR's storage limitation principle (Article 5(1)(e)), personal data must be kept no longer than necessary for the purposes for which it is processed. Organizations that retain personal data beyond its legitimate retention period are in direct violation - regardless of how securely the data is stored. Encryption doesn't exempt you from the storage limitation principle. It doesn't reduce your obligation to define retention periods, respond to deletion requests, or justify why a sub-processor holds copies of your customers' HR records.

Why this matters for AI agent architectures

AI agents are different from traditional integrations in a critical way: they make unpredictable, context-driven API calls. A traditional integration syncs the same data on a schedule. An AI agent might pull one employee record today, list all open deals tomorrow, and create a Jira ticket next week - all based on a conversation with a human user. If the MCP server platform stores every payload that passes through it, you're accumulating a growing, unpredictable dataset of your customers' most sensitive information in someone else's infrastructure.

MCP Server Data Retention Policies Compared (2026)

Let us look at the specific data retention postures of the prominent platforms used to host MCP servers and AI agent tools today, based on their own public documentation. For a broader look at how these platforms handle execution and rate limits, see our comparison of StackOne, Composio, and Truto.

Platform Default Data Storage Architecture Key Evidence
Merge Yes - stores data at rest Sync-and-cache "Data is encrypted at rest and in transit" (security page)
Composio Yes - stores data at rest Agent execution layer "Encryption at rest" (integration docs)
StackOne No - pass-through by default Real-time proxy "Never stores your customers' data by default"
Truto No - stateless pass-through In-memory processing Strict zero data retention by design

1. Merge: Sync-and-Cache by Default

Merge's security page states that "data is encrypted at rest and in transit, and PII is protected with an additional layer of application encryption." Their help documentation confirms: "Merge encrypts all data at rest and in-transit! All our data is stored in AWS, and is encrypted using the AES-256 encryption algorithm."

To Merge's credit, they explain the rationale: "Ultimately, every 3rd party API is unique. By storing customer data, Merge can obfuscate most of these differences behind our API." They also offer Merge Destinations as a premium add-on: "a streaming-only option that still provides all of Merge's syncing and normalization benefits, without data ever resting on Merge infrastructure."

The default behavior matters here. If your security team reviews Merge's standard offering, they'll see a platform that stores third-party data in databases - with encryption, yes, but with storage. The zero-storage option exists, but it's a premium feature, not the default posture. Passing sensitive enterprise data through their default layer expands your compliance scope.

2. Composio: Data Stored as Part of Managed Execution

Composio positions itself as an agent-first execution and authentication layer. Their documentation describes a production-ready platform as needing "a secure credential vault with encryption at rest, automated token lifecycle management (refresh, rotation, and revocation), detailed audit logs for every action, and verifiable compliance with standards like SOC 2 and GDPR."

The phrase "encryption at rest" inherently confirms that data is persisted - you don't encrypt data at rest if there's no data at rest. Similar to Merge, passing sensitive enterprise data through their execution layer creates a compliance footprint for any application that routes traffic through them.

3. StackOne: Real-Time Proxy, No Storage by Default

StackOne markets itself as "real-time by design" and claims it "never stores your customers' data by default. SOC 2 Type II certified, GDPR and HIPAA compliant. StackOne proxies every request in real-time." Their architecture page further states: "No customer data is persisted by default - nothing to breach, nothing to leak."

A nuance worth noting: StackOne offers a "no-storage-by-default" architecture, but for synthetic events, "tokens and logs may still be stored depending on config." This means the zero-storage posture has configuration-dependent exceptions that you should clarify during procurement. Because they act as a pass-through by default, they do not retain the sensitive data your agents process, simplifying security reviews.

4. Truto: Stateless Pass-Through Architecture

Truto takes a zero data retention approach by design, not as a premium tier. The platform operates as a stateless pass-through proxy with dynamic tool generation. Truto processes API payloads entirely in-memory and never writes customer data to disk.

This extends to Truto's MCP server implementation. When an AI agent calls a tool - say, list_all_hub_spot_contacts - the request flows through Truto's MCP endpoint, hits the HubSpot API with the correct OAuth credentials, transforms the response according to the tool's schema, and returns the result to the agent. At no point does the contact data persist in Truto's infrastructure. The only data Truto stores is authentication credentials (OAuth tokens) and configuration metadata - not your customers' business data. Your SOC 2 and GDPR boundaries remain exactly where they are today.

The Pass-Through Architecture: Zero Data Retention in Practice

Building a true zero data retention MCP server requires a completely different engineering approach. You cannot rely on cron jobs, background workers, or staging databases. Every operation must happen in real-time, in memory, during the lifecycle of a single HTTP request.

What does a zero-retention MCP server actually look like at the architectural level? Here is how a stateless pass-through architecture operates when an AI agent makes an MCP tool call.

sequenceDiagram
    participant Agent as AI Agent<br>(Claude, ChatGPT)
    participant MCP as MCP Server<br>(Pass-Through)
    participant SaaS as Upstream API<br>(Salesforce, Workday)

    Agent->>MCP: JSON-RPC tools/call (list_employees)
    MCP->>MCP: Validate token, load config<br>(in-memory)
    MCP->>SaaS: GET /api/employees<br>(Inject OAuth token)
    SaaS-->>MCP: Raw JSON response
    MCP->>MCP: JSONata Transformation<br>(in-memory, no disk write)
    MCP-->>Agent: Formatted MCP result payload
    Note over MCP: Zero data persisted.<br>Memory buffer freed after response.

The key properties of this architecture include:

In-Memory Payload Transformation

When a tool like create_a_jira_issue is invoked, the platform receives the flat JSON-RPC arguments. Instead of writing this request to a queue backed by a database, a pass-through system splits the arguments into query and body parameters purely in memory.

Platforms like Truto use JSONata - a lightweight query and transformation language - to map the agent's standardized input into the specific schema required by the provider. The request is dispatched to the upstream API, the response is caught in a memory buffer, transformed back into the desired format, and streamed directly to the client. Once the HTTP connection closes, the memory is freed. Nothing is written to a durable database.

Dynamic Tool Generation

Another major difference is how MCP tools are actually created. Platforms that store data often rely on pre-built, hardcoded tool packs stored in a database.

Pass-through platforms generate tools dynamically. In Truto, when an MCP client sends a tools/list request, the platform reads the integration's OpenAPI documentation and configuration schemas on the fly. It generates descriptive, snake_case tool names, extracts JSON Schema definitions for query and body parameters, and injects pagination instructions directly into the tool descriptions.

No tools are cached. If an environment-level override updates a tool description to give the LLM better context, that change is reflected instantly on the next request. This documentation-driven approach ensures the MCP server acts purely as a real-time reflection of the underlying API.

Transparent Rate Limit Handling for AI Agents

One of the most misunderstood aspects of integration architecture is how to handle rate limits (HTTP 429 errors). Traditional integration platforms often attempt to "help" by absorbing rate limit errors. They catch the 429, hold the connection open, and apply exponential backoff and retries under the hood.

For standard ETL jobs, this is fine. For AI agents, this is an architectural disaster.

LLMs require deterministic, fast responses. If your infrastructure holds a connection open for 45 seconds while it retries a Salesforce endpoint, the LLM client will likely time out. Worse, the AI agent loses all context about why the tool failed. It cannot reason about the delay because the infrastructure hid the error. No request queuing means no temporary storage of request payloads waiting for retry.

Truto takes a radically different approach: it does not retry, throttle, or apply backoff on rate limit errors. When an upstream API returns a 429, Truto passes that error directly back to the caller immediately. However, it normalizes the chaotic, provider-specific rate limit headers into standardized IETF RateLimit headers.

HTTP/1.1 429 Too Many Requests
ratelimit-limit: 100
ratelimit-remaining: 0
ratelimit-reset: 60

By passing these standardized headers back directly, you give the AI agent's reasoning engine full visibility. The agent can read the ratelimit-reset value and intelligently decide to either sleep for 60 seconds or switch to a different task entirely. The caller remains in total control of the retry logic, and no requests are arbitrarily queued in a third-party database.

No Sync Jobs or Cache Layers

There is no background process pulling data from upstream APIs into a local database. Every request is a live, real-time call to the provider. Responses aren't cached between requests. If the agent asks for the same employee list twice, it hits the upstream API twice.

This design has trade-offs. You give up the ability to serve stale data when an upstream API is down. You give up the performance benefit of a warm cache. Every request adds the latency of a round-trip to the upstream provider. For many enterprise compliance teams, those trade-offs are acceptable - even preferable - because they eliminate an entire class of data-at-rest risk.

Why Your Enterprise Customers Demand Zero Data Retention

Selling AI features to enterprise customers is difficult. Selling AI features that require duplicating their sensitive data into an unverified startup's database is nearly impossible.

If you're building AI agent integrations for a product that sells to enterprise, your customers' InfoSec teams will evaluate your sub-processors during procurement. Here's what that looks like in practice:

The security questionnaire. Every enterprise deal over a certain ACV triggers a vendor security review. The questionnaire will ask: "Does your product transmit customer data to third parties? Do those third parties store data at rest? What are their retention policies?" If your architecture relies on a sync-and-cache unified API, you have to explain why a copy of their employee directory or customer list resides on a third-party server just so your AI agent can read it. You need to document it, justify it, and potentially negotiate a Data Processing Agreement (DPA) between your customer, you, and the MCP server vendor.

SOC 2 scope creep. Under SOC 2, the right scope "includes all systems that store, process, or transmit customer data." A sub-processor that stores your customers' data at rest falls squarely within that scope. Your auditor will want to see their SOC 2 report, their data handling procedures, and evidence that your DPA is enforceable.

GDPR sub-processor obligations. Under GDPR, you're liable as the data controller for any processing activities carried out by external service providers. "This is why it's critical - and legally required - to have comprehensive contracts in place before any data processing by third parties begins." Every sub-processor that stores EU personal data extends your data mapping obligations and your exposure to Article 17 deletion requests.

The deal-killer scenario. An enterprise prospect's CISO asks during the final stages of procurement: "Your AI agent reads our employee data from BambooHR. Where does that data go?" If the answer involves a third-party MCP server that caches payloads in a US-based cloud, and the prospect's employees are in the EU, you've just triggered a data residency conversation that can stall or kill the deal entirely.

Adopting a stateless, zero-data-retention architecture bypasses this friction entirely. You can confidently state that your MCP server infrastructure acts only as a secure, in-memory proxy. It routes the request, injects the OAuth tokens, maps the schema, and forgets the data the millisecond the transaction completes. There's no retention policy to negotiate because there's nothing to retain. The security questionnaire answer becomes: "Data passes through the MCP server in-memory and is discarded after the response is sent." That's a much shorter conversation with an InfoSec team.

Evaluation Checklist: What to Ask Your MCP Server Vendor

Before you sign with any MCP server platform, get clear answers to these questions:

  1. Does your platform store API response payloads at rest? "Encrypted at rest" is not the same as "not stored." Push past the marketing language.
  2. What data do you persist? Even zero-retention platforms store some metadata - OAuth tokens, configuration, API logs. Understand exactly what's stored and where.
  3. Can I point to a documented zero-retention architecture for my customers' security reviews? You need something your sales team can hand to a prospect's CISO.
  4. How do you handle rate limits? If the platform queues or retries requests, request payloads may be temporarily stored. Understand the mechanism.
  5. Is zero data retention the default, or a premium add-on? A feature that exists but isn't enabled by default means your team needs to remember to configure it - and mistakes happen.
  6. What happens during an outage? If the upstream API is down, does the platform serve stale cached data? If so, that cache is stored data.

Pick the Architecture That Matches Your Compliance Requirements

The right MCP server platform depends on your compliance posture. If you're selling to SMBs with minimal security review requirements, a sync-and-cache architecture might be fine - the normalization benefits outweigh the compliance overhead. But if you're selling to enterprise customers in regulated industries - healthcare, finance, HR tech - a pass-through, zero-retention architecture isn't a nice-to-have. It is a strategic requirement for moving upmarket.

The best MCP server platforms for enterprise are converging on this reality. StackOne and Truto both default to zero data storage. Merge offers it as a premium feature. The direction is clear: enterprise buyers want their sensitive data to pass through, not pile up.

Truto's approach - stateless pass-through, dynamic tool generation, and transparent rate limit headers instead of hidden retry queues - is built for teams that need to clear enterprise security reviews without adding compliance overhead. If your AI agents need to connect to SaaS data without expanding your SOC 2 boundary, that's the architecture to evaluate.

Frequently Asked Questions

Do MCP servers store my data?
It depends on the platform. Some MCP server providers (like Merge) sync and cache data at rest by default. Others (like Truto and StackOne) use pass-through architectures that process data in-memory without persisting it. OpenAI explicitly states that data sent to third-party MCP servers is governed by those servers' retention policies.
What does encrypted at rest mean for integration platforms?
Encrypted at rest means data is stored and encrypted while stored. It does not mean data isn't retained. Under GDPR and SOC 2, the fact that data is stored at all creates compliance obligations - encryption reduces breach impact but doesn't eliminate retention, sub-processor, or deletion obligations.
How does a zero data retention architecture work?
Zero data retention platforms operate as real-time proxies. They receive a request, map the schema in memory using tools like JSONata, fetch data from the upstream API, and stream the response directly to the client without writing to a persistent database.
Why is absorbing rate limits bad for AI agents?
If an infrastructure layer absorbs a 429 error and silently retries with backoff, the LLM client will likely time out. Agents need rate limit errors returned immediately with reset headers so they can reason about the delay and switch tasks.
Why do enterprise customers care about MCP server data retention?
Enterprise InfoSec teams evaluate sub-processors during procurement. An MCP server that stores data at rest expands your SOC 2 audit scope, triggers GDPR Data Processing Agreement requirements, and creates data residency questions that can stall or kill deals.

More from our Blog