The open-weight model, dubbed OpenAI Privacy Filter, is part of what the company calls its broader effort to give developers tools aimed at safely building AI and making privacy and security protections easier to include from the start.
Wednesday’s release is the latest announcement from tech giants about new tools that promise to better protect data in an AI-era.
OpenAI Privacy Filter seeks to solve a different problem: helping redact sensitive information from unstructured text, like an online forum. It can spot names, dates, account or credit card numbers, and email addresses, according to the company. Users can also fine-tune the model to their own needs and privacy policies, and catch other types of information, OpenAI said.
“We think a strong ecosystem is one where more builders have usable tools and clear guidance and the ability to improve protections in their own environments,” Charles de Bourcy, an OpenAI privacy engineer working on Privacy Filter, told Bloomberg Law. “And this spans across both privacy and also security. So we wanted to give developers practical tools that they can run, inspect and improve on their own environments to improve privacy protections.”
To build its latest model, OpenAI said it first defined the types of information Privacy Filter should detect, including contact details and passwords. It then converted an already-trained language model for this new purpose, and trained it on a mix of publicly available and synthetic data.
The model is small enough that it can run locally, the company said, meaning unredacted data stays on devices. The AI giant said it uses a fine-tuned version of Privacy Filter for its own work, including data minimization.
Still, the company in a blog post said its model isn’t “an anonymization tool, a compliance certification, or a substitute for policy review in high-stakes settings,” but rather “one component in a broader privacy-by-design system.”
OpenAI noted that, like all models, Privacy Filter can make mistakes and miss uncommon identifiers. Users in heavily regulated industries, like the legal, medical, and financial sectors, should still rely on humans for review, it said.
“Part of what we look forward to is receiving feedback from the community,” de Bourcy said. “We are open to being surprised by which types of companies use it because I really think it is a technology that can be applied very broadly.”
To contact the reporter on this story:
To contact the editors responsible for this story:
Learn more about Bloomberg Law or Log In to keep reading:
See Breaking News in Context
Bloomberg Law provides trusted coverage of current events enhanced with legal analysis.
Already a subscriber?
Log in to keep reading or access research tools and resources.
