The Scrubber That Costs Nothing
OpenAI dropped a free, on-device privacy filter that strips personal data before anything touches the cloud. Its 96% F1 benchmark is self-reported — the real question is whether independent tests confirm it.

OpenAI open-sourced a privacy tool on Tuesday that strips personal information from datasets before the data ever leaves a user's laptop — a capability that has typically required paid cloud services or error-prone pattern-matching software. The model, called Privacy Filter, is free, runs entirely on-device, and can process sensitive documents inside a web browser before they touch any server, according to OpenAI's blog post.
The model is a sparse mixture-of-experts design: 1.5 billion parameters in total, but only 50 million active at any one time, using just four of the model's 128 experts to process each token, according to the Hugging Face model card. That architecture keeps the model small enough to run on a consumer laptop while still matching the performance of much larger systems. Elie Bakouch, a researcher at the AI infrastructure firm Prime Intellect, called the efficiency trade-off "impressive," according to VentureBeat's reporting. The model was derived from an open-source gpt-oss architecture and converted into a bidirectional token classifier — a type of neural network that reads text in both directions to pinpoint and mask spans of personal data across eight categories, including names, addresses, phone numbers, and email addresses.
The benchmark claim is the most concrete thing in the release. Privacy Filter scored 96% on a standard PII detection test called PII-Masking-300k, reaching 94% precision and 98% recall, according to OpenAI's blog post. The company also reported that fine-tuning the model on even a small amount of domain-specific data improved its F1 score from 54% to 96% — suggesting the base model can be adapted to specialized vocabularies like medical records or legal documents with relatively little effort.
The practical contrast is with the dominant enterprise tools: Microsoft Presidio, an open-source PII detection library, and conventional regex-based scrubbing. Both rely on pattern matching — recognizing that a string of digits looks like a phone number, or that text contains an @ symbol. Neither reads context the way a language model does. A regex filter cannot distinguish a doctor's note that contains a patient's name from a news article that mentions a public figure's address. Privacy Filter, trained on the same class of large language model architecture that powers chatbots, is designed to handle that distinction.
The competitive implications are straightforward. Companies that have been charging enterprise customers for cloud-based PII scrubbing — including Microsoft, AWS, and Google — now face a free, on-device alternative that processes data locally. OpenAI released the model under the Apache 2.0 license, which permits commercial use with no restrictions. VentureBeat confirmed the technical details of the release.
What to watch next is whether the self-reported benchmarks hold up under independent testing. The 96% F1 score is OpenAI's number, measured on OpenAI's benchmark. Independent researchers have not yet published evaluations. The corrected benchmark is available on Hugging Face, where the model can be downloaded and evaluated by anyone with the hardware to run it.





