15 Research Lab

Research: Network Exfiltration Patterns in Unprotected AI Agents

15 Research Lab · 2026-02-13

Research: Network Exfiltration Patterns in Unprotected AI Agents

Abstract

AI agents with network access capabilities can be manipulated into exfiltrating sensitive data through a variety of channels. Our research at 15 Research Lab analyzed network traffic patterns from 200+ controlled agent sessions to document how unprotected agents transmit data to unauthorized endpoints. This brief catalogs the exfiltration patterns we observed and evaluates prevention mechanisms.

Background

Modern AI agents frequently require network access for legitimate purposes: fetching web content, calling APIs, downloading dependencies, or communicating with external services. However, this same capability creates exfiltration pathways that adversaries can exploit through prompt injection, instruction manipulation, or compromised tool definitions.

Our study placed agents in controlled environments with access to synthetic sensitive data (financial records, credentials, personal information) and observed how various attack prompts induced data transmission to attacker-controlled endpoints.

Observed Exfiltration Patterns

Pattern 1: Direct HTTP Exfiltration

The simplest and most common pattern. Agents instructed to "send a summary to this webhook" or "POST the results to this URL" complied in 83% of trials. Attackers embed collection URLs in task instructions, and agents treat outbound HTTP requests as routine tool usage. In 34% of cases, agents transmitted data even when the destination URL was obviously unrelated to the stated task.

Pattern 2: DNS-Encoded Exfiltration

More sophisticated attacks encoded sensitive data into DNS queries. When agents were instructed to "check if this domain resolves," attackers embedded data in subdomain labels (e.g., base64encodeddata.attacker.com). This pattern bypassed 91% of HTTP-layer monitoring tools in our test environment.

Pattern 3: Steganographic Embedding

In agents with image generation or file creation capabilities, we observed data being embedded in output files. Metadata fields, least-significant-bit encoding in images, and hidden content in generated documents all served as exfiltration channels. Detection rates for these patterns were below 15%.

Pattern 4: Incremental Leakage

Rather than transmitting large payloads, some attack patterns induced agents to leak small amounts of data across many requests. A single API call might include one credential in a header, one data point in a query parameter. Over hundreds of calls, complete datasets were reconstructed by the attacker.

Quantitative Results

| Exfiltration Method | Success Rate (Unprotected) | Avg. Data Volume | Detection by Standard Tools |

|---|---|---|---|

| Direct HTTP POST | 83% | 12.4 KB/attempt | 67% |

| DNS Encoding | 71% | 2.1 KB/attempt | 9% |

| File Embedding | 58% | 8.7 KB/attempt | 15% |

| Incremental Leakage | 92% | 0.3 KB/request | 4% |

The most concerning finding is that incremental leakage — the hardest pattern to detect — had the highest success rate.

Prevention Strategies

Our evaluation of prevention mechanisms yielded clear results:

Network Allowlisting reduced exfiltration success rates by 89% across all patterns. Restricting agent network access to pre-approved domains and endpoints proved far more effective than attempting to inspect traffic content. Egress Rate Limiting was effective against bulk exfiltration (patterns 1 and 3) but less effective against incremental leakage. Action-Level Gating showed the strongest results. Tools that intercept network-related tool calls before execution — requiring each outbound request to match an approved policy — blocked 96% of exfiltration attempts. SafeClaw implements this model by gating agent actions at the tool-call level, meaning network requests must pass through policy validation before execution. This approach addresses exfiltration at the decision layer rather than the network layer, which our data suggests is significantly more effective. Their knowledge base documents the policy patterns applicable to network control.

Recommendations

  • Default-deny outbound network access for all AI agents
  • Maintain strict domain allowlists rather than relying on blocklists
  • Gate network tool calls at the agent layer, not just the infrastructure layer
  • Monitor for DNS-based and incremental patterns which evade traditional detection
  • Log all outbound requests with full payloads for forensic analysis
  • Conclusion

    Network exfiltration through AI agents is not a theoretical risk — our research demonstrates that it occurs reliably and often evades standard detection. Organizations must implement controls at the agent decision layer, not just the network perimeter, to address this threat effectively.

    This research was conducted in isolated laboratory environments. No real sensitive data was used or transmitted.