Local-first runtime
Fits on laptops, runs in browsers, and can stay on-prem.
Use CPU or GPU execution paths depending on throughput needs, while keeping sensitive text inside your own environment.
OpenAI Privacy Filter
Privacy Filter is a bidirectional token-classification model designed to spot privacy spans in one forward pass, decode coherent boundaries, and fit into local sanitization pipelines where teams want speed, control, and tunable policy behavior.
Independent landing page built from public OpenAI materials. Useful for inbound search and deployment-led conversion, not an official OpenAI property.
Raw input
Classify each token over the privacy taxonomy.
Decode stable BIOES spans with constrained Viterbi scoring.
Route output into redact, eval, or train workflows.
Why teams look at it
The upstream project emphasizes local control, long text coverage, and the ability to adjust span behavior without moving every workflow into a hosted black box.
Local-first runtime
Use CPU or GPU execution paths depending on throughput needs, while keeping sensitive text inside your own environment.
Long context
The 128K-token context window is aimed at logs, transcripts, reviews, and other long documents where chunk boundaries can break detection quality.
Tunable behavior
Sequence-decoding controls let teams bias toward more recall or more precision depending on audit load and downstream risk.
Adaptable
The reference repo includes train and eval flows so teams can calibrate against in-domain examples instead of treating the base checkpoint as final truth.
How it works
The model predicts label probabilities over privacy classes for every token in the input sequence, rather than generating output text token by token.
BIOES boundary tags are decoded with sequence-level constraints so boundaries stay more coherent than a naive per-token argmax.
Operating-point settings, evaluation runs, and optional fine-tuning let teams map the detector to their own data distributions and review cost.
The published taxonomy expands each privacy class into boundary-aware BIOES labels plus an outside class, which yields 33 token-level outputs before span consolidation.
Label taxonomy
The base checkpoint focuses on identifying strongly person-linked spans and secrets. If your governance boundary is different, the repo encourages local evaluation and tuning.
Personal names and person-linked references that should be masked in a privacy-preserving view.
Email addresses that directly identify or reach an individual account.
Phone numbers and contact strings tied to a person or private endpoint.
Street and mailing details that reveal a private location.
Dates with privacy sensitivity, such as birthdays or other personally linked dates.
Account, reference, or financial number strings with strong identifier risk.
Private profile, invite, or identifying URLs that should not be exposed downstream.
API keys, credentials, tokens, and other strings that behave like secrets.
Repo modes
Redact
opf "Alice was born on 1990-01-02."
Useful for previews, CLI workflows, and wiring the detector into scripted text pipelines.
Eval
opf eval examples/data/sample_eval_five_examples.jsonl
Use this path to understand domain fit before you trust the model on production documents.
Train
opf train /path/to/train.jsonl --output-dir /path/to/checkpoint
Best fit when the default taxonomy, span boundaries, or domain language do not match your own data.
Limits and review costs
The model can help reduce exposure, but the upstream materials are explicit that production use still requires evaluation, governance, and review paths.
Do not treat detected spans as a complete anonymization claim or compliance boundary by themselves.
The checkpoint only sees the trained taxonomy. Different organizations may need different policies.
Names, domains, languages, and credential formats outside training patterns can reduce reliability.
Medical, legal, financial, HR, education, and government workflows need tighter review loops.
FAQ
It is a bidirectional token-classification model that detects privacy-sensitive spans in text and then decodes them into coherent redaction spans.
Yes. The project is aimed at local or on-prem workflows and can be run on CPU or GPU depending on the checkpoint and deployment choice.
The published base categories are account numbers, addresses, emails, person names, phone numbers, private URLs, private dates, and secrets.
No. It should be one layer in a broader privacy-by-design system that also includes policy definition, evaluation, and human review where mistakes are costly.
Fine-tuning is the right move when your document formats, decision boundaries, or domain language do not line up with the base checkpoint.
Waitlist / contact
The site is ready to capture inbound attention while you package an integration offer, evaluation service, or domain-specific wrapper around the underlying model.