Logan Kelly
Most PII tools scrub logs after exposure. That's not enough for agentic systems. Here's what pre-execution PII enforcement actually requires.

In early 2026, one developer shipped a local privacy firewall on Hacker News with a simple explanation: they'd "recently caught myself almost pasting a block of logs containing AWS keys into Claude." The solution was a local interceptor that scanned data before it reached any cloud model. Around the same time, OpenAI launched Privacy Filter — an open-weight model for detecting and redacting PII in text that developers can run locally or integrate into their pipelines. Two products, the same instinct — something is reaching the model that shouldn't.
Both reflect the same architectural assumption: PII will get to the AI system, and the job is to catch it in transit or label it afterward. For single-turn LLM interactions with bounded scope, this is a reasonable posture. For agentic systems, it is not.
Why Detection Works for Chatbots and Fails for Agents
A single-turn LLM interaction is bounded in scope. A user sends a message. The model processes it. A response returns. Even if the model sees a social security number it shouldn't, the exposure is contained — one model call, one context, one response.
Agentic systems do not work this way. An agent that receives PII in its context window does not just respond to it. It reasons over it across multiple steps. It writes the output of that reasoning to databases. It passes data to external API calls. It may spawn subagents that inherit the full context. It generates artifacts — summaries, reports, embeddings — that persist long after the original interaction ends.
By the time a detection system flags a PII exposure in an agentic pipeline, the data may have already propagated through a dozen downstream tool calls, been embedded in a vector store, or reached a partner API. Detection identifies the event. It does not contain the downstream consequences.
The practitioner who built the local privacy firewall understood this intuitively: the only safe moment to catch sensitive data is before it leaves your environment. Detection after that point is forensics, not prevention.
The Observability Gap: What Your Trace Scrubber Misses
The standard enterprise response to PII in agentic systems has converged on trace and log redaction. Arize's documentation, for example, provides guidance on building a custom span processor that regex-strips PII from OTEL spans before export — a straightforward and sensible control. SS&C Blue Prism's AI gateway scans text inputs for sensitive patterns before they reach an LLM. OpenAI's Privacy Filter operates at the model layer.
These controls have real value. They reduce the risk of PII persisting in retained traces, logs, and observability platforms. But the architectural assumption embedded in all of them is the same: the data will reach the model or the trace system, and we will clean it up on the way out.
The cleanup controls have two failure modes that matter for agents specifically.
First, cleanup is sequential. In a multi-step agentic workflow, an agent may take a tool action — writing to a database, calling an external API, posting data to a webhook — before any cleanup layer has seen the full trace context. By the time a span processor scrubs the OTEL export, the tool call has already fired. The PII has already left.
Second, cleanup misses context propagation. Subagents that inherit parent context windows inherit whatever PII the parent was working with. A scrubber operating on exported spans doesn't prevent that inheritance — it just means the logs don't show it.
Regulatory Pressure Is on the Processing Event, Not the Logs
The framing regulators apply to PII in AI systems is not "what did you retain?" It is "what did you process, and was it necessary?"
GDPR Article 5(1)(c) requires that personal data be "adequate, relevant and limited to what is necessary in relation to the purposes for which they are processed." This is data minimization. An agent that receives a full customer record to answer a billing question has processed more personal data than the task required — regardless of whether that data was scrubbed from the subsequent trace.
The European Data Protection Board's Coordinated Enforcement Framework for 2026 is running a coordinated action on GDPR transparency and information obligations (Articles 12–14), involving 25 EU data protection authorities. Under Articles 12–14, controllers must be able to explain to data subjects what personal data they are processing and why — an obligation that applies directly to AI systems handling personal data. Trace scrubbing is a compliance mitigation for retention risk. It does not satisfy the transparency and data minimization obligations that regulators are actively assessing.
The Colorado AI Act (SB 24-205), effective June 30, 2026, similarly focuses on what automated systems do with personal data — data subject rights include the right to appeal adverse automated decisions, which creates audit requirements that trace cleanup cannot satisfy if the decisions were made using unnecessarily broad data inputs.
CCPA's automated decision-making transparency regulations took effect January 1, 2026 (with compliance deadlines for most businesses by January 1, 2027), requiring organizations to disclose how AI systems use personal data and provide consumers the right to opt out of automated decision-making. That explanation becomes harder if PII flowed freely into agent context at every step.
The architectural direction these regulations are pointing toward is pre-execution control: enforce what data enters agent context in the first place, rather than cleaning up after the fact.
The Prevention Architecture: Signal, Domain, and Pre-Execution Policy
A PII prevention architecture for agentic systems has two components that work together.
The first is a controlled data interface — the principle behind Waxell's Signal/Domain pattern. Instead of surfacing full records to agents, the interface exposes only the fields an agent needs for the task it is performing. A billing inquiry agent receives billing identifiers. A scheduling agent receives calendar data. Neither receives the customer's medical history, unrelated PII, or fields outside the task scope, because those fields were never made available through the interface.
This is distinct from field-level encryption or database access controls. It is a data interface design pattern that shapes what flows into agent context at all. An agent cannot reason over, propagate, or expose data it never received.
The second component is a pre-execution policy layer. Waxell Runtime's 26 policy categories include data handling policies that evaluate and enforce constraints before agent actions execute — not during observability export, not after the fact. A policy that prevents agents from passing data matching configured PII patterns to external tool calls applies across every agent in the system, without requiring each agent implementation to encode that logic independently.
Together, these controls change the posture from "detect and scrub PII as it flows" to "prevent PII from flowing in the first place."
How Waxell Handles This
Waxell Runtime enforces data handling policies at the governance plane — above individual agent implementations — so that PII constraints don't have to be re-implemented by every team that ships an agent. Policies run at the pre-execution stage, which means a rule blocking agents from passing PII-matching data to external tool calls evaluates and blocks the call before it fires.
The Signal/Domain interface pattern lets teams define exactly what data flows into agent context. The constraint lives at the interface level — "this agent receives customer IDs, not customer records" — and is enforced structurally rather than relying on model self-restriction or post-hoc cleanup.
For teams deploying vendor or third-party agents — systems built outside the organization's codebase — Waxell Connect provides the same governance layer without requiring SDK integration or code changes. External agents receive the same data handling policies and interface constraints as internally built ones. PII risk in agentic systems extends to every agent that touches organizational data, not just the ones your team wrote.
Waxell Observe initializes with two lines of code and auto-instruments 157+ libraries, giving teams immediate visibility into what is flowing through their agent systems — the necessary starting point for understanding where PII controls need to be applied before a policy layer is configured.
Waxell's safety model is built on the premise that governance controls should sit above agents, not inside them. PII enforcement is a clear case where that separation matters: a control inside the agent depends on the agent behaving correctly. A control above the agent enforces regardless of agent behavior.
FAQ
What's the difference between PII detection and PII prevention in AI agents?
PII detection identifies sensitive data after it has entered a system — in a trace, in agent output, or in a log — and redacts or scrubs it. PII prevention controls what data enters agent context in the first place. For agentic systems where data propagates across tool calls, subagents, and external APIs before any detection system sees the full picture, prevention is the more reliable primary control. Detection remains valuable as a backstop.
Can't trace redaction tools keep me GDPR-compliant?
Trace redaction reduces retained exposure, but it doesn't address the processing event. GDPR's data minimization principle applies to what an agent processes, not just what it stores. An agent that processes a full customer record to answer a billing question has processed more personal data than necessary — and trace cleanup afterward doesn't change that. Regulators assess the processing decision, not the log hygiene.
What is the Signal/Domain pattern for agentic systems?
The Signal/Domain pattern is a data interface design that defines what flows into an agent's context window. Rather than exposing full records, the interface surfaces only the fields required for the task at hand. This prevents agents from receiving data they shouldn't reason over, without requiring changes to the agent model itself. The enforcement is structural — at the interface — rather than behavioral.
Does GDPR's data minimization principle apply to agent context windows?
Yes. GDPR Article 5(1)(c) requires personal data to be "adequate, relevant and limited to what is necessary" for the processing purpose. This applies to what an agent receives in its context, not just to what is retained afterward. An agent processing unnecessary PII — even if logs are subsequently scrubbed — has still processed data in excess of what the task required.
How does multi-agent context propagation create PII risk?
In multi-agent architectures, subagents typically inherit context from parent agents. If the parent agent's context contains PII, that data propagates to every subagent in the chain, potentially crossing security boundaries in the process. A trace redactor operating on exported spans doesn't prevent this inheritance. Pre-execution policy enforcement at the governance layer — applying consistently across parent and child agents — is the control that addresses propagation risk directly.
What PII-related regulatory deadlines matter most for enterprise AI deployments in 2026?
The EU AI Act begins broad enforcement August 2, 2026, with transparency and data governance requirements for AI systems covered under Annex III. The Colorado AI Act (SB 24-205) takes effect June 30, 2026, covering automated decision-making and data subject rights including the right to appeal adverse decisions. CCPA's automated decision-making transparency regulations took effect January 1, 2026, with compliance deadlines for most businesses by January 1, 2027. Across all three frameworks, the focus is on what AI systems do with personal data — not merely on retention and cleanup.
Sources
Show HN: Local Privacy Firewall — blocks PII and secrets before ChatGPT sees them — https://news.ycombinator.com/item?id=46206591
OpenAI, Introducing OpenAI Privacy Filter (April 22, 2026) — https://openai.com/index/introducing-openai-privacy-filter/
OpenAI Privacy Filter — Hacker News discussion — https://news.ycombinator.com/item?id=47870901
Arize AX Docs: Redact sensitive data from traces — https://arize.com/docs/ax/observe/tracing/configure/redact-sensitive-data-from-traces
SS&C Blue Prism: AI Gateway for PII Sanitization — https://www.blueprism.com/resources/blog/ai-gateway-pii-sanitization/
GDPR Article 5(1)(c), data minimization — https://gdpr-info.eu/art-5-gdpr/
EDPB, CEF 2026: EDPB launches coordinated enforcement action on transparency and information obligations under the GDPR (March 19, 2026) — https://www.edpb.europa.eu/news/news/2026/cef-2026-edpb-launches-coordinated-enforcement-action-transparency-and-information_en
EU Artificial Intelligence Act, Annex III: High-Risk AI Systems Referred to in Article 6(2) — https://artificialintelligenceact.eu/annex/3/
Colorado AI Act SB 24-205, effective June 30, 2026 — https://leg.colorado.gov/bills/sb24-205
California Privacy Protection Agency, CCPA Updates, Automated Decisionmaking Technology (ADMT), and Related Regulations — https://cppa.ca.gov/regulations/ccpa_updates.html
Agentic Governance, Explained




