AI News
12 Oct 2025
Read 14 min
OpenAI chat log preservation order 2025 explained
OpenAI chat log preservation order 2025 ends sweeping retention, restoring user privacy protections.
What the OpenAI chat log preservation order 2025 actually says
The original hold: keep everything
In May 2025, the court told OpenAI to save all chat logs. This was part of discovery in the Times’ copyright lawsuit. The idea was simple: do not delete anything that might show how ChatGPT behaves with news content or how model outputs may relate to the Times’ claims.OpenAI pushed back on scope and privacy
OpenAI appealed. The company argued the order was too broad and risky. Keeping every output forever could expose sensitive user data. It could also disrupt normal operations and data hygiene. The company said the hold was an overreach for what the case actually needs.The October 9 decision: a narrower path
On October 9, Judge Ona T. Wang ended the broad hold. OpenAI no longer has to “preserve and segregate all output log data that would otherwise be deleted on a going forward basis.” This means OpenAI can resume normal deletions and retention rules.Effective date and key exceptions
The court set a clear cutoff. As of September 26, OpenAI does not need to preserve new logs under the old, sweeping order. But there are two big exceptions: – Logs and data already preserved under the earlier order remain accessible. – OpenAI must keep data linked to ChatGPT accounts that the New York Times has flagged. The Times can also expand the list of flagged accounts as it reviews preserved records. That lets the plaintiff continue to build its case without forcing OpenAI to stockpile all new logs.What remains preserved
Evidence captured under the earlier order stays. This includes chat logs that existed and were saved up to the cutoff, and any material tied to flagged accounts. The court struck a balance: it protects relevant evidence while limiting fresh, bulk collection of user conversations.Why this matters: privacy, product operations, and legal discovery
User privacy risk goes down
A blanket hold on every output log can sweep in sensitive user data. It can include personal details typed into prompts, private business plans, or proprietary code. Narrowing the order reduces that risk. It limits the number of new user conversations stored just for litigation.OpenAI gets operational relief
Saving everything costs money and slows systems. It makes routine deletion and minimization harder. The new order lets OpenAI return to standard retention practices. That means better data hygiene, lower storage costs, and fewer chances for accidental exposure of old logs.Discovery stays on target
Courts want relevant evidence, not everything under the sun. This order keeps preserved data that might matter to the case and focuses on accounts the Times flags. It still gives the plaintiff a path to request more, but it ties new preservation to concrete needs.Signals for future AI lawsuits
This decision shows courts are willing to limit broad data holds in AI cases. Judges may ask plaintiffs to define narrower scopes rooted in specific users, time windows, or content types. That approach aims to protect privacy while keeping discovery fair.Who gains and who loses from the change
Flagged accounts: what that means and how expansion could work
What counts as “flagged”
The Times can identify specific ChatGPT accounts it believes are relevant. Once an account is flagged, OpenAI must preserve data linked to it. The order does not describe public criteria. It simply allows the plaintiff to mark accounts for preservation.Expansion over time
As the Times reviews existing, preserved logs, it can add more accounts to the list. This lets the plaintiff follow leads. It also avoids a blanket hold on all users. The court can still step in if scope grows too wide or too vague.Practical takeaways for companies using generative AI
Set clear retention policies
Write simple rules for how long you keep prompts, outputs, and logs. Limit retention by default. Delete data you do not need for security, quality, or legal reasons. Document your practices so you can defend them in court if needed.Be ready for targeted litigation holds
When a dispute starts, you may need to pause deletion for certain data. Aim for holds that are specific to users, dates, and topics. Broad holds raise cost and risk. Work with counsel to make holds precise and justified.Separate training data from user logs
Keep a clear boundary between model training corpora and user chat logs. Document sources and licenses. This helps you answer questions about copyright without exposing private user conversations.Minimize sensitive inputs
Teach teams not to paste secrets into prompts. Use data-loss prevention tools. If you offer an AI product, provide controls for redaction and enterprise-grade privacy settings. Less sensitive input means less sensitive output to preserve.Audit and explain
Maintain records that show what data you collect, why you collect it, and how long you keep it. Be ready to explain retention to courts and customers. Transparent logs and policies build trust and reduce legal friction.Timeline at a glance
What everyday users can do now
Check your settings
Review how your AI tool handles chat history and data controls. If you can turn off history or opt out of certain uses, decide what works for you. Smaller data footprints lower risk.Think before you paste
Do not share personal identifiers, contracts, credentials, or private source code unless you are comfortable with retention policies and security controls. Keep sensitive work in secure, approved environments.Use enterprise features where possible
If your company offers an enterprise plan with stronger privacy, use it. These plans often include data isolation and shorter retention windows designed for compliance.Follow case updates
The court’s order can shape how long your chats live in legal storage if your account is ever flagged. Staying informed helps you make better choices about what you share with AI tools.The bigger picture for AI training data and copyright
Core dispute remains
This order does not decide the copyright question. The Times says OpenAI used its content without proper compensation. OpenAI disputes the claims. The court will weigh evidence, including preserved logs, to decide if and how outputs relate to the Times’ works.Discovery without overreach
The new order shows judges can cut back overly broad holds while protecting key evidence. That balance respects user privacy and reduces unnecessary data stockpiles, yet it still supports a fair process.Precedent for future cases
Other AI cases will likely cite this approach: keep preserved material that matters, allow targeted flags, and avoid perpetual, universal retention. It is a model for handling data-rich products where discovery pressures can collide with privacy.Open questions to watch
FAQ
Contents