<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:media="http://search.yahoo.com/mrss/" xmlns:dc="http://purl.org/dc/elements/1.1/"><channel><title>Data Protection | Awesome Agents</title><link>https://awesomeagents.ai/tags/data-protection/</link><description>Your guide to AI models, agents, and the future of intelligence. Reviews, leaderboards, news, and tools - all in one place.</description><language>en-us</language><managingEditor>contact@awesomeagents.ai (Awesome Agents)</managingEditor><lastBuildDate>Wed, 22 Apr 2026 23:28:57 +0200</lastBuildDate><atom:link href="https://awesomeagents.ai/tags/data-protection/index.xml" rel="self" type="application/rss+xml"/><image><url>https://awesomeagents.ai/images/logo.png</url><title>Awesome Agents</title><link>https://awesomeagents.ai/</link></image><item><title>OpenAI Open-Sources Privacy Filter: 96% F1 PII Masker</title><link>https://awesomeagents.ai/news/openai-privacy-filter-on-device-pii/</link><pubDate>Wed, 22 Apr 2026 23:28:57 +0200</pubDate><guid>https://awesomeagents.ai/news/openai-privacy-filter-on-device-pii/</guid><description>&lt;p>OpenAI released Privacy Filter today under Apache 2.0, and the thing worth noting isn't the name. It is the shape. A 1.5-billion-parameter total, 50-million-active Mixture of Experts with a 128K context window, shipped as a bidirectional token classifier with WebGPU support via Transformers.js. In practice, that means an enterprise can run the full masking pass on text before it leaves a browser tab and hits any OpenAI endpoint, Azure tenant, or third-party API. That is a deliberate architectural position, and it answers a question the industry has been asking for two years.&lt;/p></description><content:encoded xmlns:content="http://purl.org/rss/1.0/modules/content/"><![CDATA[<p>OpenAI released Privacy Filter today under Apache 2.0, and the thing worth noting isn't the name. It is the shape. A 1.5-billion-parameter total, 50-million-active Mixture of Experts with a 128K context window, shipped as a bidirectional token classifier with WebGPU support via Transformers.js. In practice, that means an enterprise can run the full masking pass on text before it leaves a browser tab and hits any OpenAI endpoint, Azure tenant, or third-party API. That is a deliberate architectural position, and it answers a question the industry has been asking for two years.</p>
<h2 id="the-landscape-before-and-after">The Landscape, Before And After</h2>
<table>
  <thead>
      <tr>
          <th>Approach</th>
          <th>Example</th>
          <th>Context-aware</th>
          <th>Runs on-device</th>
          <th>License</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>Regex + pattern</td>
          <td>Microsoft Presidio</td>
          <td>No</td>
          <td>Yes</td>
          <td>MIT</td>
      </tr>
      <tr>
          <td>BERT token-classifier</td>
          <td><code>dslim/bert-base-NER</code></td>
          <td>Partial</td>
          <td>Yes</td>
          <td>MIT</td>
      </tr>
      <tr>
          <td>Hosted API call</td>
          <td>AWS Comprehend PII</td>
          <td>Yes</td>
          <td>No</td>
          <td>Paid service</td>
      </tr>
      <tr>
          <td>Hosted large model</td>
          <td>GPT-4 via API</td>
          <td>Yes</td>
          <td>No</td>
          <td>Paid service</td>
      </tr>
      <tr>
          <td><strong>Privacy Filter</strong></td>
          <td><code>openai/privacy-filter</code></td>
          <td>Yes</td>
          <td>Yes (browser)</td>
          <td><strong>Apache 2.0</strong></td>
      </tr>
  </tbody>
</table>
<p>The row that didn't previously exist is the one that matters. Regex tools are fast and local but confuse addresses with product codes. Large hosted models understand context but require sending the raw text off-device, which is the exact thing a privacy filter is meant to prevent. Privacy Filter is the first release this year that sits in the contextual-and-local cell of the matrix with a license you can ship in enterprise software.</p>
<div class="news-tldr">
<p><strong>Privacy Filter at a glance</strong></p>
<ul>
<li>1.5B total / 50M active (128 experts, top-4 routing)</li>
<li>128K context, non-autoregressive single-forward-pass inference</li>
<li>8 PII categories across 33 BIOES output classes</li>
<li>96% F1 on PII-Masking-300k (97.43 on the corrected set)</li>
<li>Apache 2.0, available on <a href="https://huggingface.co/openai/privacy-filter">HuggingFace</a> and <a href="https://github.com/openai/privacy-filter">GitHub</a></li>
<li>WebGPU support via Transformers.js - runs in the browser</li>
</ul>
</div>
<h2 id="how-it-actually-works">How It Actually Works</h2>
<p>The model card calls it a &quot;bidirectional token classifier,&quot; which understates the structural choices.</p>
<h3 id="the-architecture-choice">The Architecture Choice</h3>
<p>Eight transformer encoder blocks, pre-norm, d_model 640. Attention is grouped-query with rotary positional embeddings: 14 query heads, 2 KV heads, giving a group size of seven queries per KV head. The feedforward layers are sparse Mixture of Experts, 128 experts with top-4 routing per token. Every token sees four of the 128 experts, and the total active parameter count lands at 50M. The rest sits dormant in the checkpoint, which is why a 1.5B-weight download runs at 50M-weight latency.</p>
<p><img src="/images/news/openai-privacy-filter-on-device-pii-chip.jpg" alt="Circuit board close-up">
<em>The 50M-active design is the load-bearing decision. A standard dense 1.5B PII tagger would saturate a laptop GPU under batch load; a 50M-active MoE with top-4 routing runs at interactive latency on a WebGPU target in a Chromium tab. The architecture was built to deploy, not to impress on a leaderboard.</em>
<small>Source: unsplash.com</small></p>
<h3 id="the-output-format">The Output Format</h3>
<p>This isn't a generative model. For each input token the model predicts one of 33 output classes: a background <code>O</code> plus 8 PII categories in BIOES span encoding (Begin / Inside / End / Single / Outside). Decoding uses a constrained Viterbi procedure with linear-chain transition scoring, which means the span boundaries are globally consistent rather than inferred greedily. The precision-recall tradeoff is tunable at inference time through the Viterbi parameters without retraining.</p>
<p>The 8 categories are specific, and their specificity is a design statement: <code>account_number</code>, <code>private_address</code>, <code>private_email</code>, <code>private_person</code>, <code>private_phone</code>, <code>private_url</code>, <code>private_date</code>, <code>secret</code>. The last bucket, <code>secret</code>, absorbs passwords, API keys, and tokens - the category an enterprise compliance team cares about most in 2026, given the rate at which these leak through agent logs.</p>
<h3 id="the-browser-target">The Browser Target</h3>
<p>WebGPU via the Transformers.js runtime is the shipping target, not an afterthought. A three-line <code>pipeline(&quot;token-classification&quot;, &quot;openai/privacy-filter&quot;)</code> call runs the full masking inference locally with no outbound network request to OpenAI infrastructure. For an enterprise that wants to paste a customer service transcript into a summarisation tool, the filter redacts names, phone numbers, and API keys before the transcript ever leaves the device.</p>
<h2 id="what-it-does-not-tell-you">What It Does Not Tell You</h2>
<p>Three things the announcement glosses over.</p>
<h3 id="the-eight-categories-are-the-perimeter">The Eight Categories Are The Perimeter</h3>
<p>This isn't a universal PII detector. It's a detector for eight specific categories of data, trained on English-language corpora with &quot;selected multilingual robustness evaluation.&quot; It doesn't tag medical PHI as a separate class, it doesn't tag financial account information beyond generic <code>account_number</code>, and it doesn't tag biometric identifiers. An enterprise targeting HIPAA or PCI-DSS compliance isn't done when this model passes.</p>
<h3 id="the-4-that-slips">The 4% That Slips</h3>
<p>OpenAI itself reports a 96% F1 on the benchmark, which means four out of every hundred labelled spans in the test set are either missed or mis-categorised. In adversarial or out-of-distribution text - dialects the training set underrepresents, proper names written in non-Latin scripts, phone numbers in formats the tokenizer did not see often - the miss rate is higher. OpenAI's own model card states the filter &quot;may miss unusual identifiers&quot; and &quot;can over-redact short sentences.&quot; A 96% filter processing a billion-request-a-day support queue leaks 40 million spans. That is a floor, not a worst case.</p>
<h3 id="the-compliance-question">The Compliance Question</h3>
<p><img src="/images/news/openai-privacy-filter-on-device-pii-keyboard.jpg" alt="Keyboard close-up in dark light">
<em>OpenAI's own framing in the model card is worth quoting verbatim: the filter &quot;is not an anonymization tool, a compliance certification, or a substitute for policy review.&quot; The company is explicitly declining to position this as a regulatory checkbox. Enterprises deploying it still need legal review of the residual risk.</em>
<small>Source: unsplash.com</small></p>
<p>The model is positioned as a pre-processing layer, not a privacy guarantee. GDPR, HIPAA, and state-level privacy laws impose obligations that are process-level, not model-level: data retention schedules, subject access requests, consent flows, lawful basis for processing. Privacy Filter makes the pre-processing step better. It does nothing for the seven other steps.</p>
<h2 id="the-release-itself">The Release Itself</h2>
<p>The strategic note worth marking is that OpenAI released this under Apache 2.0 rather than under the open-weight-but-restricted license pattern the company has tended to use for its other open releases. No commercial use restriction, no redistribution clause, no delayed release for non-US customers. The closest precedent inside OpenAI is Whisper, released under MIT in 2022 and still the reference speech-to-text baseline for the industry. Privacy Filter is positioned to do the same work in text sanitization: ubiquitous, forkable, hard to displace.</p>
<p>What makes this release different from OpenAI's <a href="/news/openai-codex-enterprise-consulting-partners/">enterprise-Codex push</a> last week is that Privacy Filter does not require an OpenAI endpoint to deliver value. Run the classifier locally, keep the text on-premise, send nothing back to San Francisco. That is a model OpenAI is giving away to customers who have explicitly asked for exactly that. It is, in the most literal sense, a trust-building release.</p>
<hr>
<p>The model is small, permissively licensed, architecturally appropriate for the job, and honestly documented about what it can't do. The open question is whether enterprises actually wire it into their pipelines or whether it becomes a reference implementation that sits in a GitHub repository while the same customers keep pasting raw transcripts into chat. OpenAI has done its part. The next move is the integrators'.</p>
<p><strong>Sources:</strong></p>
<ul>
<li><a href="https://huggingface.co/openai/privacy-filter">OpenAI Privacy Filter model card on HuggingFace</a></li>
<li><a href="https://github.com/openai/privacy-filter">openai/privacy-filter repository on GitHub</a></li>
<li><a href="https://cdn.openai.com/pdf/c66281ed-b638-456a-8ce1-97e9f5264a90/OpenAI-Privacy-Filter-Model-Card.pdf">OpenAI Privacy Filter Model Card PDF</a> - OpenAI, 22 April 2026</li>
<li><a href="https://news.bloomberglaw.com/privacy-and-data-security/openai-releases-privacy-filter-model-to-redact-sensitive-data">OpenAI Releases Privacy Filter Model to Redact Sensitive Data</a> - Bloomberg Law, 22 April 2026</li>
<li><a href="https://www.tradingview.com/news/reuters.com,2026:newsml_FWN415196:0-openai-releases-openai-privacy-filter/">OpenAI Releases OpenAI Privacy Filter (Reuters wire)</a> - Reuters via TradingView, 22 April 2026</li>
<li><a href="https://decrypt.co/365139/openai-privacy-filter-open-source-pii-masking-model">OpenAI Just Open-Sourced a Tool That Scrubs Your Secrets Before ChatGPT Ever Sees Them</a> - Decrypt, 22 April 2026</li>
</ul>
]]></content:encoded><dc:creator>Elena Marchetti</dc:creator><category>News</category><media:content url="https://awesomeagents.ai/images/news/openai-privacy-filter-on-device-pii_hu_3190a2e2cf9807b4.jpg" medium="image" width="1200" height="675"/><media:thumbnail url="https://awesomeagents.ai/images/news/openai-privacy-filter-on-device-pii_hu_3190a2e2cf9807b4.jpg" width="1200" height="675"/></item></channel></rss>