OpenAI slams court order to save all ChatGPT logs, including deleted chats (arstechnica.com)
1101 points by ColinWright 2 days ago
sinuhe69 a day ago
Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
Billion people use the internet daily. If any organization suspects some people use the Internet for illicit purposes eventually against their interests, would the court order the ISP to log all activities of all people? Would Google be ordered to save the search of all its customers because some might use it for bad things? And once we start, where will we stop? Crimes could happen in the past or in the future, will the court order the ISP and Google to retain the logs for 10 years, 20 years? Why not 100 years? Who should bear the cost for such outrageous demands?
The consequences of such orders are of enormous impact the puny judge can not even begin to comprehend. Privacy right is an integral part of the freedom of speech, a core human right. If you don’t have private thoughts, private information, anybody can be incriminated against them using these past information. We will cease to exist as individuals and I argue we will cease to exist as human as well.
capnrefsmmat a day ago
Courts have always had the power to compel parties to a current case to preserve evidence. (For example, this was an issue in the Google monopoly case, since Google employees were using chats set to erase after 24 hours.) That becomes an issue in the discovery phase, well after the defendant has an opportunity to file a motion to dismiss. So a case with no specific allegation of wrongdoing would already be dismissed.
The power does not extend to any of your hypotheticals, which are not about active cases. Courts do not accept cases on the grounds that some bad thing might happen in the future; the plaintiff must show some concrete harm has already occurred. The only thing different here is how much potential evidence OpenAI has been asked to retain.
dragonwriter a day ago
> Courts have always had the power to compel parties to a current case to preserve evidence.
Not just that, even without a specific court order parties to existing or reasonably anticipated litigation have a legal obligation that attaches immediately to preserve evidence. Courts tend to issue orders when a party presents reason to believe another party is out of compliance with that automatic obligation, or when there is a dispute over the extent of the obligation. (In this case, both factors seem to be in play.)
btown 19 hours ago
golol a day ago
So if Amazon sues Google, claiming that it is being disadvantaged in search rankings, a court should be able to force Google to log all search activity, even when users delete it?
cogman10 a day ago
dragonwriter a day ago
monetus 20 hours ago
saddist0 a day ago
lcnPylGDnU4H9OF a day ago
So then the courts need to find who is setting their chats do be deleted and order them to stop. Or find specific infringing chatters and order OpenAI to preserve these specified users’ logs. OpenAI is doing the responsible thing here.
capnrefsmmat a day ago
dragonwriter a day ago
IAmBroom 21 hours ago
MeIam a day ago
Time does not need user logs to prove such a thing if it was true. Times can show that it is possible so they can show how their own users can access the text. Why would they need other user's data?
KaiserPro a day ago
mandevil a day ago
dragonwriter a day ago
mrtksn a day ago
>Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
Probably because they bothered to pursue such a thing and hundreds of millions people did not.
How do you conclusively know if someone's content generating machine infringe with your rights? By saving all of its input/output for investigation.
It's ridiculous, sure but is it less ridiculous than AI companies claiming that the copyrights shouldn't apply to them because it will be bad for their business?
IMHO those are just growth pain. Back in the day people used to believe that the law don't apply on them because they did it on the internet and they were mostly right because the laws were made for another age. Eventually the laws both for criminal stuff and copyright caught up. Will be the same for AI, now we are in the wild west age of AI.
TimPC a day ago
AI companies aren't seriously arguing that copyright shouldn't apply to them because "it's bad for business". The main argument is that they qualify for fair use because their work is transformative which is one of the major criteria for fair use. Fair use is the same doctrine that allows a school to play a movie for educational purposes without acquiring a license for the public performance of that movie. The original works don't have model weights and can't answer questions or interact with a user so the output is substantially different from the input.
c256 a day ago
no_wizard a day ago
mrtksn a day ago
mandevil a day ago
rodgerd 19 hours ago
freejazz 20 hours ago
AStonesThrow a day ago
shkkmo a day ago
> It's ridiculous, sure but is it less ridiculous than AI companies claiming that the copyrights shouldn't apply to them because it will be bad for their business?
Since that wasn't ever a real argument, your strawman is indeed ridiculous.
The argument is that requiring people to have a special license to process text with an algorithm is a dramatic expansion of the power of copyright law. Expansions of copyright law will inherently advantage large corporate users over individuals as we see already happening here.
New York Times thinks that they have the right to spy on the entire world to see if anyone might be trying to read articles for free.
That is the problem with copyright. That is why copyright power needs to be dramatically curtailed, not dramatically expanded.
dogman144 a day ago
You raise good points but the target of your support feels misplaced. Want private ai? You must self-host and inspect if it’s phoning home. No way around it in my view.
Otherwise, you are picking your data privacy champions as the exact same companies, people and investors that sold us social media, and did something quite untoward with the data they got. Fool me twice, fool me three times… where is the line?
In other words - OAI has to save logs now? Candidly they probably were already, or it’s foolish not to assume that.
jrm4 a day ago
Love the spirit of what you say and I practice it myself, literally.
But also, no - Just self-host or it's all your fault is never ever a sufficient answer to the problem.
It's exactly the same as when Exxon says "what are you doing to lower your own carbon footprint?" It's shifting the burden unfairly; companies like OpenAI put themselves out there and thus must ALWAYS be held to task.
dogman144 a day ago
naming_the_user a day ago
lovich a day ago
fluidcruft a day ago
A pretty clear distinction is that all ISPs in the world are not currently involved in a lawsuit with New York Times and are not accused of deleting evidence. What OpenAI is accused of is significantly different from merely agnostically routing packets between A and B. OpenAI is not raising astronomical funds because they operate as an ISP.
tailspin2019 a day ago
> Privacy right is an integral part of the freedom of speech
I completely agree with you, but as a ChatGPT user I have to admit my fault in this too.
I have always been annoyed by what I saw as shameless breaches of copyright of thousands of authors (and other individuals) in the training of these LLMs, and I've been wary of the data security/confidentiality of these tools from the start too - and not for no reason. Yet I find ChatGPT et al so utterly compelling and useful, that I poured my personal data[0] into these tools anyway.
I've always felt conflicted about this, but the utility just about outweighed my privacy and copyright concerns. So as angry as I am about this situation, I also have to accept some of the blame too. I knew this (or other leaks or unsanctioned use of my data) was possible down the line.
But it's a wake up call. I've done nothing with these tools which is even slightly nefarious, but I am today deleting all my historical data (not just from ChatGPT[1] but other hosted AI tools) and will completely reassess my approach of using them - likely with an acceleration of my plans to move to using local models as much as I can.
[0] I do heavily redact my data that goes into hosted LLMs, but there's still more private data in there about me than I'd like.
[1] Which I know is very much a "after the horse has bolted" situation...
CamperBob2 a day ago
Keeping in mind that the purpose of IP law is to promote human progress, it's hard to see how legacy copyright interests should win a fight with AI training and development.
100 years from now, nobody will GAF about the New York Times.
stackskipton a day ago
DannyBee a day ago
Lawyer here
First - in the US, privacy is not a constitutional right. It should be, but it's not. You are protected against government searches, but that's about it. You can claim it's a core human right or whatever, but that doesn't make it true, and it's a fairly reductionist argument anyway. It has, fwiw, also historically not been seen as a core right for thousands of years. So i think it's a harder argument to make than you think despite the EU coming around on this. Again, I firmly believe it should be a core right, but asserting that it is doesn't make that true.
Second, if you want the realistic answer - this judge is probably overworked and trying to clear a bunch of simple motions off their docket. I think you probably don't realize how many motions they probably deal with on a daily basis. Imagine trying to get through 145 code reviews a day or something like that. In this case, this isn't the trial, it's discovery. Not even discovery quite yet, if i read the docket right. Preservation orders of this kind are incredibly common in discovery, and it's not exactly high stakes most of the time. Most of the discovery motions are just parties being a pain in the ass to each other deliberately. This normally isn't even a thing that is heard in front of a judge directly, the judge is usually deciding on the filed papers.
So i'm sure the judge looked at it for a few minutes, thought it made sense at the time, and approved it. I doubt they spent hours thinking hard about the consequences.
OpenAI has asked to be heard in person on the motion, i'm sure the judge will grant it, listen to what they have to say, and determine they probably fucked it up, and fix it. That is what most judges do in this situation.
zerocrates 20 hours ago
Even in the "protected against government searches" sense from the 4th Amendment, that right hardly exists when dealing with data you send to a company like OpenAI thanks to the third-party doctrine.
pama a day ago
Thanks. As an EU citizen am I exempt from this order? How does the judge or the NYTimes or OpenAI know that I am an EU citizen?
ElevenLathe a day ago
mananaysiempre a day ago
adgjlsfhk1 a day ago
tiahura a day ago
While the Constitution does not explicitly enumerate a "right to privacy," the Supreme Court has consistently recognized substantive privacy rights through Due Process Clause jurisprudence, establishing constitutional protection for intimate personal decisions in Griswold v. Connecticut (1965), Lawrence v. Texas (2003), and Obergefell v. Hodges (2015).
ComposedPattern a day ago
> It has, fwiw, also historically not been seen as a core right for thousands of years.
Nothing has been seen as a core right for thousands of years, as the concept of human rights is only a few hundred years old.
HardCodedBias a day ago
"First - in the US, privacy is not a constitutional right"
What? The supreme court disagreed with you in Griswold v. Connecticut (1965) and Roe v. Wade (1973).
While one could argue that they were vastly stretching the meaning of words in these decisions the point stands that at this time privacy is a constitutional right in the USA.
DannyBee a day ago
krapp a day ago
shkkmo a day ago
> It has, fwiw, also historically not been seen as a core right for thousands of years. So i think it's a harder argument to make than you think despite the EU coming around on this.
This doesn't seem true. I'd assume you know more about this than I do though so can you explain this in more detail? The concept of privacy is definitely more than thousands of years old. The concept of a "human right", is arguably much newer. Do you have particular evidence that a right to privacy is a harder argument to make that other human rights?
While the language differs, the right to privacy is enshrined more or less explicitly in many constitutions, including 11 USA states. It isn't just a "european" thing.
static_motion a day ago
dragonwriter a day ago
> Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
Because the law favors preservation of evidence for an active case above most other interests. It's not a matter of arbitrary preference by the particular court.
piombisallow a day ago
Regardless of the details of this specific case, the courts are not democratic and do not decide based on the interest of the parties or how many they are, they decide based on the law.
brookst a day ago
This is not true even in the slightest.
The law is not a deterministic computer program. It’s a complex body of overlapping work and the courts are specifically chartered to use judgement. That’s why briefs from two parties in a dispute will often cite different laws and precedents.
For instance, Winter v. NRDC specifically says that courts must consider whether an injunction is in the public interest.
piombisallow a day ago
oersted a day ago
I completely agree with you. But perhaps we should be more worried that OpenAI or Google can retain all this data and do pretty much what they want with it in the first place, without a judge getting into the picture.
fireflash38 a day ago
In your arguments for privacy, do you consider privacy from OpenAI?
rvnx a day ago
Cut a joke about ethics and OpenAI
maest a day ago
ethersteeds a day ago
humpty-d a day ago
I fail to see how saving all logs advances that cause
hshdhdhj4444 a day ago
wat10000 16 hours ago
ChatGPT isn’t like an ISP here. They are being credibly accused of basing their entire business on illegal activity. It’s more like if The Pirate Bay was being sued. The alleged infringement is all they do, and requiring them to preserve records of their users is pretty reasonable.
baobun a day ago
It's a honeypot from the beginning y'all
trod1234 a day ago
It doesn't, it favors longstanding caselaw and laws already on the books.
There is a longstanding precedent with regards to business document retention, and chat logs have been part of that for years if not decades. The article tries to make this sound like this is something new, but if you look at the e-retention guidelines in various cases over the years this is all pretty standard.
For a business to continue operating, they must preserve business documents and related ESI upon an appropriate legal hold to avoid spoliation. They likely weren't doing this claiming the data was deleted, which is why the judge ruled in favor against OAI.
This isn't uncommon knowledge either, its required. E-discovery and Information Governance are things any business must meet in this area; and those documents are subject to discovery in certain cases, where OAI likely thought they could avoid it maliciously.
The matter here is OAI and its influence rabble are churning this trying to do a runaround on longstanding requirements that any IT professional in the US would have reiterated from their legal department/Information Governance policies.
There's nothing to see here, there's no real story. They were supposed to be doing this and didn't, were caught, and the order just forces them to do what any other business is required to do.
I remember an executive years ago (decades really), asking about document retention, ESI, and e-discovery and how they could do something (which runs along similar lines to what OAI tried as a runaround). I remember the lawyer at the time saying, "You've gotta do this or when it goes to court you will have an indefensible position as a result of spoliation...".
You are mistaken, and appear to be trying to frame this improperly towards a point of no accountability.
I suggest you review the longstanding e-discovery retention requirements that courts require of businesses to operate.
This is not new material, nor any different from what's been required for a long time now. All your hyperbole about privacy is without real basis, they are a company; they must comply with law, and it certainly is not outrageous to hold people who break the law to account, and this can only occur when regulatory requirements are actually fulfilled.
There is no argument here.
References: Federal Rules of Civil Procedure (FRCP) 1, 4, 16, 26, 34, 37
There are many law firms who have written extensively on this and related subjects. I encourage you to look at those too.
(IANAL) Disclosure: Don't take this as legal advice. I've had the opportunity to work with quite a few competent ones, but I don't interpret the law; only they can. If you need someone to provide legal advice seek out competent qualified counsel.
rolandog a day ago
> Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
Can't you use the same arguments against, say, Copyright holders? Billionaires? Corporations doing the Texas two-step bankruptcy legal maneuver to prevent liability from allegedly poisoning humanity?
I sure hope so.
Edit: ... (up to a point)
deadbabe a day ago
OpenAI is a business selling a product, it’s not a decentralized network of computers contributing spare processing power to run massive LLMs. Therefore, you can easily point a finger at them and tell them to stop some activity for which they are the sole gatekeeper.
cactusplant7374 a day ago
> Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
It simply didn't. ChatGPT hasn't deleted any user data.
> "OpenAI did not 'destroy' any data, and certainly did not delete any data in response to litigation events," OpenAI argued. "The Order appears to have incorrectly assumed the contrary."
It's a bit of a stretch to think a big tech company like ChatGPT is deleting users' data.
blackqueeriroh 11 hours ago
This is incorrect. As someone who has had the opportunity to work in several highly=regulated industries, companies do not want to hold onto extra data about you that they don’t have to unless their business is selling that data.
OpenAI already has a business, and not one they want to violate by having a massive amount of customer data stolen if they get hacked.
cactusplant7374 7 hours ago
resource_waste a day ago
>Privacy right is an integral part of the freedom of speech, a core human right.
Are these contradictory?
If you overhear a friend gossiping, can't you spread that gossip?
Also, where are human rights located? I'll give you a microscope.(sorry, I'm a moral anti-realist/expressivist and I can't help myself)
152132124 a day ago
I think you will have a better time arguing with a LLM
huijzer a day ago
> Why could a court favor the interest of the New York Times in a vague accusation versus the interest and right of hundred millions people?
Well maybe some people in power have pressured the court into this decision? The New York Times surely has some power as well via their channels
longnguyen 7 minutes ago
Shameless plug: I build Chat Vault to help you import all chat data from ChatGPT, Claude, Grok and Le Chat (Mistral) to a native mac app.
You can search, browse and continue your chats 100% offline.
It’s free while in beta https://testflight.apple.com/join/RJx6sP6t
efskap 2 days ago
Note that this also applies to GPT models on the API
> That risk extended to users of ChatGPT Free, Plus, and Pro, as well as users of OpenAI’s application programming interface (API), OpenAI said.
This seems very bad for their business.
merksittich a day ago
Interesting detail from the court order [0]: When asked by the judge if they could anonymize chat logs instead of deleting them, OpenAI's response effectively dodged the "how" and focused on "privacy laws mandate deletion." This implicitly admits they don't have a reliable method to sufficiently anonymize data to satisfy those privacy concerns.
This raises serious questions about the supposed "anonymization" of chat data used for training their new models, i.e. when users leave the "improve model for all users" toggle enabled in the settings (which is the default even for paying users). So, indeed, very bad for the current business model which appears to rely on present users (voluntarily) "feeding the machine" to improve it.
[0] https://cdn.arstechnica.net/wp-content/uploads/2025/06/NYT-v...
Kon-Peki a day ago
Thank you for the link to the actual text!
So, the NYT asked for this back in January and the court said no, but asked OpenAI if there was a way to accomplish the preservation goal in a privacy-preserving manner. OpenAI refused to engage for 5 f’ing months. The court said “fine, the NYT gets what they originally asked for”.
Nice job guys.
noworriesnate a day ago
blackqueeriroh 11 hours ago
That’s not an implicit admission, it’s refusing to argue something they don’t want to do.
neilv 2 days ago
Some established businesses will need to review their contracts, regulations, and risk tolerance.
And wrapper-around-ChatGPT startups should double-check their privacy policies, that all the "you have no privacy" language is in place.
999900000999 2 days ago
I'm not going to look up the comment, but a few months back I called this out and said if you seriously want to use any LLM in a privacy sensitive context you need to self host.
For example, if there are business consequences for leaking customer data, you better run that LLM yourself.
TeMPOraL 2 days ago
jaggederest 2 days ago
ted537 a day ago
fakedang 2 days ago
Etheryte a day ago
In the European privacy framework, and legal framework at large, you can't terms of service away requirements set by the law. If the law requires you to keep the logs, there is nothing you can get the user to sign off on to get you out of it.
zombot a day ago
cj a day ago
> Some established businesses will need to review their contracts, regulations, and risk tolerance.
I've reviewed a lot of SaaS contracts over the years.
Nearly all of them have clauses that allow the vendor to do whatever they have to if ordered to by the government. That doesn't make it okay, but it means OpenAI customers probably don't have a legal argument, only a philosophical argument.
Same goes for privacy policies. Nearly every privacy policy has a carve out for things they're ordered to do by the government.
Nasrudith a day ago
Wowfunhappy 2 days ago
> And wrapper-around-ChatGPT startups should double-check their privacy policies, that all the "you have no privacy" language is in place.
If a court orders you to preserve user data, could you be held liable for preserving user data? Regardless of your privacy policy.
gpm 2 days ago
cortesoft 2 days ago
woliveirajr 2 days ago
pjc50 a day ago
bilbo0s 2 days ago
Chris2048 a day ago
Just to be pedantic, could the company encrypt the logs with a third-party key in escrow, s.t they would not be able to access that data, but the third party could provide access e.g. for a court.
HappMacDonald a day ago
johnQdeveloper 2 days ago
> This seems very bad for their business.
Well, it is gonna be all _AI Companies_ very soon so unless everyone switches to local models which don't really have the same degree of profitability as a SaaS, its probably not going to kill a company to have less user privacy because tbh people are used to not having privacy these days on the internet.
It certainly will kill off the few companies/people trusting them with closed source code or security related stuff but you really should not outsource that anywhere.
csomar 2 days ago
Did an American court just destroy all American AI companies in favor of open weight Chinese models?
pjc50 a day ago
thot_experiment 2 days ago
mountainriver 2 days ago
You can fine tune models on a multitenant base model and it’s often more profitable.
SchemaLoad 2 days ago
>don't really have the same degree of profitability as a SaaS
They have a fair bit. Local models lets companies sell you a much more expensive bit of hardware. Once Apple gets their stuff together it could end up being a genius move to go all in on local after the others have repeated scandals of leaking user data.
johnQdeveloper 2 days ago
bsder 2 days ago
> It certainly will kill off the few companies/people trusting them with closed source code or security related stuff but you really should not outsource that anywhere.
And how many companies have proprietary code hosted on Github?
johnQdeveloper 2 days ago
Kokouane 2 days ago
If you were working with code that was proprietary, you probably shouldn't of been using cloud hosted LLMs anyways, but this would seem to seal the deal.
larrymcp 2 days ago
I think you probably mean "shouldn't have". There is no "shouldn't of".
rimunroe 2 days ago
knicholes 2 days ago
amanaplanacanal 2 days ago
DecentShoes 2 days ago
gpm 2 days ago
I think it's fair to question how proprietary your data is.
Like there's the algorithm by which a hedge fund is doing algorithmic trading, they'd be insane to take the risk. Then there's the code for a video game, it's proprietary, but competitors don't benefit substantially from an illicit copy. You ship the compiled artifacts to everyone, so the logic isn't that secret. Copies of the similar source code have linked before with no significant effects.
FuckButtons 2 days ago
short_sells_poo a day ago
YetAnotherNick 2 days ago
Why not? Assuming you believe you can use any cloud for backup or Github for code storage.
solaire_oa a day ago
consumer451 2 days ago
All GPT integrations I’ve implemented have been via Azure’s service, due to Microsoft’s contractual obligation for them not to train on my data.
As far as I understand it, this ruling does not apply to Microsoft, does it?
Descon 2 days ago
I think when you spin up open AI in azure, that instance is yours, so I don't believe that would be subject to this order.
tbrownaw a day ago
dinobones 2 days ago
How? This is retention for legal risk, not for training purposes.
They can still have legal contracts with other companies, that stipulate that they don't train on any of their data.
paxys 2 days ago
Your employees' seemingly private ChatGPT logs being aired in public during discovery for a random court case you aren't even involved in is absolutely a business risk.
lxgr 2 days ago
pjc50 a day ago
godelski 2 days ago
Retention means an expansion of your threat model. Specifically, in a way you have little to no control over.
It's one thing if you get pwned because a hacker broke into your servers. It is another thing if you get pwned because a hacker broken into somebody else's servers.
At this point, do we believe OpenAI has a strong security infrastructure? Given the court order, it doesn't seem possible for them to have sufficient security for practical purposes. Your data might be encrypted at rest, but who has the keys? When you're buying secure instances, you don't want the provider to have your keys...
bcrosby95 2 days ago
antihipocrat 2 days ago
Will a business located in another jurisdiction be comfortable that the records of all staff queries & prompts are being stored and potentially discoverable by other parties? This is more than just a Google search, these prompts contain business strategy and IP (context uploads for example)
CryptoBanker 2 days ago
Right, because companies always follow the letter of their contracts.
lxgr 2 days ago
Why would the reason matter for people that don't want their data retained at all?
Take8435 2 days ago
...Data that is kept can be exfiltrated.
fn-mote 2 days ago
jameshart a day ago
Thinking about the value of the dataset of Enron’s emails that was disclosed during their trials, imagine the value and cost to humanity of all OpenAI’s api logs even for a few months being entered into court record..
ivape 2 days ago
Going to drop a PG tweet:
https://x.com/paulg/status/1913338841068404903
"It's a very exciting time in tech right now. If you're a first-rate programmer, there are a huge number of other places you can go work rather than at the company building the infrastructure of the police state."
---
So, courts order the preservation of AI logs, and government orders the building of a massive database. You do the math. This is such an annoying time to be alive in America, to say the least. PG needs to start blogging again about what's going on now days. We might be entering the digital version of the 60s, if we're lucky. Get local, get private, get secure, fight back.
bigfudge a day ago
Will this apply to Azure OpenAI model APIs too?
ukuina 2 days ago
Aren't most enterprise customers using AzureOpenAI?
m3kw9 2 days ago
Not when people have nowhere else to go, pretty much you cannot escape it, it’s too convenient to not use now. You think no other AI chat providers doesn’t need to do this?
jacob019 2 days ago
I think the court overstepped by ordering OpenAI to save all user chats. Private conversations with AI should be protected - people have a reasonable expectation that deleted chats stay deleted, and knowing everything is preserved will chill free expression. Congress needs to write clear rules about what companies can and can't do with our data when we use AI. But honestly, I don't have much faith that Congress can get their act together to pass anything useful, even when it's obvious and most people would support it.
ethagnawl 2 days ago
Why is AI special in this regard? Why is my exchange with ChatGPT any more privileged than my DuckDuckGo search for _HIV test margin of error_?
jacob019 2 days ago
You're right, it's not special.
This is from DuckDuckGo's privacy policy: "We don’t track you. That’s our Privacy Policy in a nutshell. We don’t save or share your search or browsing history when you search on DuckDuckGo or use our apps and extensions."
If the court compelled DuckDuckGo to log all searches, I would be equally concerned.
sib a day ago
robocat a day ago
raincole a day ago
AI is not special and that's the exact issue. The court made a precedence here. If OpenAI can be ordered to preserve all the logs, then DuckDuckGo can face the same issue even if they don't want to do that.
energy123 a day ago
People upload about 100x more information about themselves to ChatGPT than search engines.
nradov 2 days ago
How did the court overstep? Orders to preserve evidence are routine in civil cases. Customer expectations about privacy have zero legal relevance.
jacob019 2 days ago
Sure, preservation orders are routine - but this would be like ordering phone companies to record ALL calls just in case some might become evidence later. There's a huge difference between preserving specific communications in a targeted case and mass surveillance of every private conversation. The government shouldn't have that kind of blanket power over private communications.
charonn0 2 days ago
nradov a day ago
pjc50 a day ago
Consider the opposite prevailing, where I can legally protect my warez site simply by saying "sorry, the conversation where I sent them a copy of a Disney movie was private".
riskable a day ago
The legal situation you describe is a matter of impossibility and unrelated to the OpenAI case.
In the case of a warez site they would never have logged such a "conversation" to begin with. So if the court requested that they produce all such communications the warez site would simply declare that as, "Impossibility of Performance".
In the case of OpenAI the courts are demanding that they preserve all future communications from all their end users—regardless of whether or not those end users are parties (or even relevant) to the case. The court is literally demanding that they re-engineer their product to record all communications where none existed previously.
I'm not a lawyer but that seems like it would violate FRCP 26(b)(1) which covers "proportionality". Meaning: The effort required to record the evidence is not proportional relative to the value of the information sought.
Also—generally speaking—courts recognize that a party is not required to create new documents or re-engineer systems to satisfy a discovery request. Yet that is exactly what the court has requested of OpenAI.
lcnPylGDnU4H9OF a day ago
If specific users are violating the law, then a court can and should order their data to be retained.
BrtByte a day ago
The preservation order feels like a blunt instrument in a situation that needs surgical precision
marcyb5st a day ago
Would it be possible to comply with the order by anonymizing the data?
The court is after evidence that users use ChatGPT to bypass paywalls. Anonymizing the data in a way that makes it impossible to 1) pinpoint the users and 2) reconstruct the generic user conversation history would preserve privacy and allow OpenAI to comply in good faith with the order.
The fact that they are blaring sirens and hide behind the "we can't, think about users' privacy" feels akin to willingful negligence or that they know they have something to hide.
lcnPylGDnU4H9OF a day ago
> feels akin to willingful negligence or that they know they have something to hide
Not at all; there is a presumption of innocence. Unless a given user is plausibly believed to be violating the law, there is no reason to search their data.
Miraltar a day ago
Anonymizing data is really hard and I'm not sure they'd be allowed to do it. I mean they're accused of deleting evidences, why would they be allowed to alter it ?
amanaplanacanal 2 days ago
If it's possible evidence as part of a lawsuit, of course they can't delete it.
jacob019 2 days ago
A targeted order is one thing, but this applies to ALL data. My data is not possible evidence as part of a lawsuit, unless you know something I don't know.
artursapek 2 days ago
ianks 2 days ago
This ruling is unbelievably dystopian for anyone that values a right to privacy. I understand that the logs will be useful in the occasional conviction, but storing a log of people’s most personal communications is absolutely not a just trade.
To protect their users from the this massive overreach, OpenAI should defy this order and eat the fines IMO.
imiric a day ago
This is a moot issue. OpenAI and all AI service providers already use all user-provided data for improving their models, and it's only a matter of time until they start selling it to advertisers, if they don't already. Whether or not they actually delete chat conversations is irrelevant.
Anyone concerned about their privacy wouldn't use these services to begin with. The fact they are so popular is indicative that most people value the service over their privacy, or simply don't care.
wongarsu a day ago
Plenty of service providers (including OpenAI) offer you the option to kindly ask them not to, and will even contractually agree not to use or sell your data if you want such an agreement.
Yes, they want to use everyone's data. But they also want everyone as a customer, and they can't have both at once. Offering people an opt-out is a popular middle-ground because the vast majority of people don't care about it, and those that do care are appeased
malwrar a day ago
imiric a day ago
thewebguyd a day ago
> The fact they are so popular is indicative that most people value the service over their privacy, or simply don't care.
Or, the general populace just doesn't understand the actual implications. The HN crowd can be guilty of severely overestimating the average person's tech literacy, and especially their understanding of privacy policies and ToS. Many may think they are OK with it, but I'd argue it's because they don't understand the potential real-world consequences of such privacy violations.
imiric 8 hours ago
yard2010 a day ago
It's almost rigged. Either they are keeping the data (and ofc making money out of it) or deleting it destroying the evidence of the crimes they're committing..
romanovcode a day ago
This has nothing to do with convictions of criminals but everything with CIA gathering profiles every single person they can.
paxys 2 days ago
Not only does this mean OpenAI will have to retain this data on their servers, they could also be ordered to share it with the legal teams of companies they have been sued by during discovery (which is the entire point of a legal hold). Some law firm representing NYT could soon be reading out your private conversations with ChatGPT in a courtroom to prove their case.
fhub 2 days ago
My guess is they will store them on tape e.g. on something like Spectra TFinity ExaScale library. I assume AWS glacier et al use this sort of thing for their deep archives.
Storing them on something that has hours to days retrieval window satisfies the court order, is cheaper, and makes me as a customer that little bit more content with it (mass data breach would take months of plundering and easily detectable).
genewitch a day ago
Glacier is tape silos, but this is textual data. You don't need to save output images, just the checkpoint+hash of the generating model and the seed. Stable diffusion saves this until you manually delete the metadata, for example. So my argument is you could do this with LTO as well. Text compresses well, especially if you don't do it naively.
JKCalhoun 2 days ago
> She suggested that OpenAI could have taken steps to anonymize the chat logs but chose not to
That is probably the solution right there.
blagie 2 days ago
This data cannot be anonymized. This is trivial provable, both mathematically, but given the type of data, it should also be intuitively obvious to even the most casual observer.
If you're talking to ChatGPT about being hunted by a Mexican cartel, and having escaped to your Uncle's vacation home in Maine -- which is the sort of thing a tiny (but non-zero) minority of people ask LLMs about -- that's 100% identifying.
And if the Mexican cartel finds out, e.g. because NY Times had a digital compromise at their law firm, that means someone is dead.
Legally, I think NY Times is 100% right in this lawsuit holistically, but this is a move which may -- quite literally -- kill people.
zarzavat 2 days ago
JKCalhoun a day ago
genewitch a day ago
AOL found out and thus we all found out that you can't anonymize certain things, web searches in that case. I used to have bookmarked some literature from maybe ten years ago that said,(proved with math?), any moderate collection of data from or by individuals that fits certain criteria is de-anonymizeable, if not by itself, then with minimal extra data. I want to say it included if, for instance, instead of changing all occurances of genewitch to user9843711, every instance of genewitch was a different, unique id.
I apologize for not having cites or a better memory at this time.
catlifeonmars a day ago
paxys 2 days ago
> She suggested that OpenAI could have taken steps to anonymize the chat logs but chose not to, only making an argument for why it "would not" be able to segregate data, rather than explaining why it "can’t."
Sounds like bullshit lawyer speak. What exactly is the difference between the two?
dijksterhuis 2 days ago
bilbo0s 2 days ago
I’d just assume that any chat or api call you do to any cloud based ai in th US will be discoverable from here on out.
If that’s too big a risk it really is time to consider locally hosted LLMs.
amanaplanacanal 2 days ago
That's always been the case for any of your data anywhere in any third party service of any kind, if it is relevant evidence in a lawsuit. Nothing specific to do with LLMs.
marcyb5st a day ago
I ask again, why not anonymizing the data? That way NYT/the court could see if users are bypassing the paywall through ChatGPT while preserving privacy.
Even if I wrote it, I don't care if someone read out loud in public court "user <insert_hash_here> said: <insert nastiest thing you can think of here>"
Orygin a day ago
You can't really anonymize the data if the conversation itself is full of PII.
I had colleagues chat with GPT, and they send all kinds of identifying information to it.
kragen a day ago
Copyright in its current form is incompatible with private communication of any kind through computers, because computers by their nature make copies of the communication, so it makes any private communication through a computer into a potential crime, depending on its content. The logic of copyright enforcement, therefore, demands access to all such communications in order to investigate their legality, much like the Stasi.
Inevitably such a far-reaching state power will be abused for prurient purposes, for the sexual titillation of the investigators, and to suppress political dissent.
6stringmerc a day ago
This is a ludicrous assertion and factually inaccurate beyond all practical intelligence.
A computer in service of an individual absolutely follows copyright because the creator is in control of the distribution and direction of the content.
Besides, copyright is a civil statute, not criminal. Everything about this comment is the most obtuse form of FUD possible. I’m pro copyright reform, but this is “Uncle off his meds ranting on Facebook” unhinged and shouldn’t be given credence whatsoever.
malwrar a day ago
> A computer in service of an individual absolutely follows copyright because the creator is in control of the distribution and direction of the content.
I don’t understand what means. A computer in service of an individual turns copyright law into mattress tag removal law—practically unenforceable.
kragen a day ago
None of that is correct. Some of it is not even wrong, demonstrating an unbelievably profound ignorance of its topic. Furthermore, it is gratuitously insulting.
pjc50 a day ago
> Besides, copyright is a civil statute, not criminal
Nope. https://www.justia.com/intellectual-property/copyright/crimi...
Imnimo 2 days ago
So if you're a business that sends sensitive data through ChatGPT via the API and were relying on the representation that API inputs and outputs were not retained, OpenAI will just flip a switch to start retaining your data? Were notifications sent out, or did other companies just have to learn about this from the press?
celnardur 2 days ago
There has been a lot of opinion pieces popping up on HN recently that describe the benefits they see from LLMs and rebut the drawbacks most of them talk about. While they do bring up interesting points, NONE of them have even mentioned the privacy aspect.
This is the main reason I can’t use any LLM agents or post any portion of my code into a prompt window at work. We have NDAs and government regulations (like ITAR) we’d be breaking if any code left our servers.
This just proves the point. Until these tools are local, privacy will be an Achilles heal for LLMs.
garyfirestorm 2 days ago
You can always self host an LLM which is completely controlled on your own server. This is trivial to do.
redundantly 2 days ago
Trivial after a substantial hardware investment and installation, configuration, testing, benchmarking, tweaking, hardening, benchmarking again, new models come out so more tweaking and benchmarking and tweaking again, all while slamming your head against the wall dealing with the mediocre documentation surrounding all hardware and software components you're trying to deploy.
Yup. Trivial.
dvt 2 days ago
benoau 2 days ago
jjmarr 2 days ago
dlivingston a day ago
genewitch a day ago
blastro 2 days ago
celnardur 2 days ago
Yes, but which of the state of the art models that offer the best results, are you allowed to do this with? As far as I've seen the models that you can host locally are not the ones being praised left and right in these articles. My company actually allows people to use a hosted version of Microsoft copilot, but most people don't because it's still not that much of a productivity boost (if any).
genewitch a day ago
aydyn 2 days ago
It is not at all trivial for an organization that may be doing everything on the cloud to locally set up the necessary hardware and ensure proper networking and security to that LLM running on said hardware.
woodrowbarlow a day ago
> NONE of them have even mentioned the privacy aspect
because the privacy aspect has nothing to do with LLMs and everything to do with relying on cloud providers. HN users have been vocal about that since long before LLMs existed.
g42gregory 2 days ago
Can somebody please post a complete list of these news organizations, demanding to see all of our ChatGPT conversations?
I see one of them: The New York Times.
We need to let people know who the other ones are.
dijksterhuis 2 days ago
tgv a day ago
Why?
DaSHacka a day ago
To know what subscriptions we need to cancel.
tgv a day ago
lrvick 2 days ago
There is absolutely no reason for these logs to exist.
Run LLM in an enclave that generates ephemeral encryption keys. Have users encrypt text directly to those enclave ephemeral keys, so prompts are confidential and only ever visible in an environment not capable of logging.
All plaintext data will always end up in the hands of governments if it exists, so make sure it does not exist.
jxjnskkzxxhx a day ago
Then a court will order that you don't encrypt. And probably go after you for trying to undermine the intent of previous court order. Or what, you thought you found an obvious loophole in the entire legal system?
lrvick a day ago
Yes. Because once you have remote attestation, anyone can host these enclaves in any country, and charge some tiny fee for their gpu time.
Decentralize hosting and encryption then centralized developers of the open source software will be literally unable to comply.
This well proven strategy would however only be possible if anything about OpenAI was actually open.
TechDebtDevin a day ago
Do you have any reading on this?
paxys a day ago
Encryption does not negate copyright laws. The solution here is for LLM builders to pay for training data.
ronsor a day ago
The solution here is to get rid of copyright.
mucha 14 hours ago
OJFord 2 days ago
Better link in the thread: https://arstechnica.com/tech-policy/2025/06/openai-says-cour...
(As in, an actual article, not just a mastodon-tweet from some unknown (maybe known? Not by me) person making the title claim, with no more info.)
incompatible 2 days ago
Looks like https://en.wikipedia.org/wiki/Lauren_Weinstein_(technologist..., he has been commentating on the Internet for about as long as it has existed.
bibinou a day ago
And? the article you linked only has primary sources.
genewitch a day ago
Roughly how many posts on HN are by people you know?
OJFord a day ago
Of those that are tweets and similar? Almost all of them (the ones I look at being interested in the topic anyway).
By 'know' I mean recognise the name as some sort of authority. I don't 'know' Jon Gruber or Sam Altman or Matt Levine, but I'll recognise them and understand why we're discussing their tweet.
The linked tweet (whatever it's called) didn't say anything more than the title did here, so it was pointless to click through really. In replies someone asked the source and someone else replied with the link I commented above. (I don't 'know' those people either, but I recognise Ars/even if I didn't appreciate the longer form with more info.)
genewitch a day ago
ronsor 2 days ago
This court order certainly violates privacy laws in multiple jurisdictions and existing contracts OpenAI may have with customers.
CryptoBanker 2 days ago
Existing contracts have zero bearing on what a court may and may not order.
ronsor 2 days ago
Contracts don't, but foreign law is going to make this a pain for OpenAI. Other countries may not care what a U.S. court orders; they want their privacy laws followed.
jillesvangurp a day ago
mosdl 2 days ago
adriand 2 days ago
The order also dates back to May 13. What the fuck?! That’s weeks ago! The only reason I can think of for why OpenAI did not warn its users about this via an email notification is because it’s bad for their business. But wow is it ever a breach of trust not to.
jcranmer 2 days ago
I don't think the order creates any new violations of privacy law. OpenAI's ability to retain the data and give it to third parties would have been the violation in the first place.
ryeguy_24 a day ago
Would Microsoft have to comply with this also? Most enterprise users are acquiring LLM services through Microsoft's instance of the models in Azure? (i.e. data is not going to Open AI but enterprise gets to use Open AI models)
anbende a day ago
My (not a lawyer) understanding is "no", because Microsoft is not administering the model (making available the chatbot and history logging), not retaining chats (typically, unless you configure it specifically to do this), and any logs or history are only retained on the customer's servers or tenant.
Accessing information on a customer's server or tenant (I have been assured) would require a court order for the customer directly.
But... as an 365 E5 user with an Azure account using the 4o through Foundry... I am much more nervous than I ever have been.
cedws a day ago
Not that it makes it any better but I wouldn’t be surprised if the NSA had a beam splitter siphoning off every byte going to OpenAI already. Don’t send sensitive data.
exq a day ago
The former head of the NSA is on the board. I'd be more surprised if the feds WEREN'T siphoning off every byte by now.
solfox 2 days ago
> People on both platforms recommended using alternative tools to avoid privacy concerns, like Mistral AI or Google Gemini,
Presumably, this same ruling will come for all AI systems soon; Gemini, Grok, etc.
spjt 2 days ago
It won't be coming for local inference.
blibble 2 days ago
they'll just outlaw that entirely
ivape 2 days ago
YetAnotherNick 2 days ago
If in any case they require logging for all LLM calls, then by extension local non logged LLMs would be outlawed sooner or later.
JKCalhoun 2 days ago
It would probably surprise no one if we find out, some time from now, tacit agreements to do so were already made (are being made) behind closed doors. "We'll give you what you want, just please don't call us out publicly."
acheron 2 days ago
“use a Google product to avoid privacy concerns” is risible.
shadowgovt 2 days ago
Google has the calibre of lawyers to make this hard for news companies to pull off.
hsbauauvhabzb 2 days ago
Are they all not collecting logs?
shortsunblack 6 hours ago
The order mandates retention of all user data, even of non-Americans. This is a massive extraterritorial overreach and a highlight of how US law has zero regard to data protection as a fundamental human right. As if there were not enough concerns about US cloud providers already.
mastazi 2 days ago
I'm seeing HN hug of death when attempting to open the link, but was able to read the post on Wayback Machine https://web.archive.org/web/20250604224036/https://mastodon....
I think this is a private Mastodon instance on someone's personal website so it makes sense that it might have been overwhelmed by the traffic.
wglb 2 days ago
I have a friend who is a Forensic Attorney (certified Forensic Examiner and licensed attorney). He says "You folks are putting all of this subpoenable information on Google and other sites"
jwpapi 2 days ago
Anything that can be done with the existing ones?
How is it with using openrouter?
If I have users that use OpenAI through my API keys am I responsible?
I have so many questions…
ripdog 2 days ago
>If I have users that use OpenAI through my API keys am I responsible?
Yes. You are OpenAI's customer, and they expect you to follow their ToS. They do provide a moderation API to reject inappropriate prompts, though.
rkagerer a day ago
This highlights a significance today's cloud-addicted generation seems to completely ignore: who has control of your data.
I'm not talking about contractual control (which is largely mooted as pretty much every cloud service has a ToS that's grossly skewed toward their own interests over yours, with clauses like indemnifications, blanket grants to share your data with "partners" without specifying who they are or precisely what details are conveyed, mandatory arbitration, and all kinds of other exceptions to what you'd consider respectful decency), but rather where your data lives and is processed.
If you truly want to maintain confidence it'll remain private, don't send it to the cloud in the first place.
grafmax a day ago
Framing this as a moralized issue of “addiction” on the part of consumers naturalizes the structural cause of the problem. Concentrated wealth benefits from the cloud capital consumers generate. It’s this group that has most of the control over how our data is collected. These companies reduce the number and quality of choices available to us. Blaming consumers for choosing the many conveniences of cloud data when the incentive structure has been carefully tailored to funnel our data into their possession and control is truly a superficial take.
lxgr 17 hours ago
> Blaming consumers for choosing the many conveniences of cloud data when the incentive structure has been carefully tailored to funnel our data into their possession and control is truly a superficial take.
Couldn't have said it better.
Just consider Apple as an example: Some time ago, they used to sell the Time Capsule, a local-first NAS built for wireless Time Machine backups. Today, not only has the Time Capsule been discontinued, but it's outright impossible to make local backups of iOS devices (many people's primary computing devices!) without a computer and a USB cable.
Considering Apple's resources, it would take negligible effort to add a NAS backup feature to iOS with polished UX ("simply tap your phone on your Time Capsule 2.0 to pair it, P2P Wi-Fi faster than your old USB cable" etc.) – but they won't, because now it's all about "services": Why sell a Gigabyte once if you can lease it out and collect rent for it every month instead?
keybored a day ago
Well put. And generalizes to most consumer-blaming.
bunderbunder 21 hours ago
fireflash38 a day ago
Yeah, I see a ton of people all up in arms about privacy but ignoring that OpenAI doesn't give a rats ass about others privacy (see: scraping).
Like why one good other bad?
daveoc64 a day ago
If something is able to be scraped, it isn't private.
There is no technical reason for chats people have with ChatGPT or any similar service to be available on the web to everyone, so there is no way for them to be scraped.
brookst a day ago
It’s not zero sum. I can believe that openai does not take privacy seriously enough and also that I don’t want every chat I’ve ever had with their product to be entered into the public record.
“If one is good the other must be good” is far too simplistic thinking to apply to a situation like this.
fireflash38 a day ago
ahmeneeroe-v2 21 hours ago
This seems like an unhelpful extension of the word "privacy". Scraping is something, but it is mostly not a privacy violation.
nashashmi a day ago
We have shifted over to SaaS so much for convenience that we have lost sight of “our control”.
I imagine a 90s era software industry for today’s tech world: person buys a server computer, person buys an internet connection with stable ip, person buys server software boxes to host content on the internet, person buys security suite to firewall access.
Where is the problem in this model? Aging computers? Duplicating computing hardware for little use? Unsustainable? Not green/efficient?
SpaceL10n a day ago
> who has control of your data
As frustrating as it is, the answer seems to be everyone and no one. Data in some respects is just an observation. If I walk through a park, and I see someone with red hair, I just collected some data about them. If I see them again, perhaps strike up a conversation, I learn more. In some sense, I own that data because I observed it.
On the other other hand, I think most decent people would agree that respecting each other's right to privacy is important. Should the owner of the red hair ask me to not share personal details about them, I would gladly accept, because I personally recognize them as the owner of the source data. I may possess an artifact or snapshot of that data, but it's their hair.
In a digital world where access controls exist, we have an opportunity to control the flow of our data through the public space. Unfortunately, a lot of work is still needed to make this a reality...if it's even possible. I like the Solid Project for it's attempt to rewrite the internet to put more control in the hands of the true data owners. But, I wonder if my observation metaphor is still possible even in a system like Solid.
sneilan1 a day ago
It's not developer's common interest to develop local services first. People build cloud services for a profit so they can be paid. However, sometimes developers need to build their portfolios (or out of pure interest) so they make software that runs locally anyway. A lot of websites can easily be ran on people's computers from a data perspective but it's a lot of work to get that data in the first place and make it useable. I don't think people are truly "cloud-addicted". I think they simply do not have other choices.
jpadkins a day ago
the post does not reflect the reality that it is not 'your data'*. When you use a service provider, it's their data. They may give you certain rights to influence your usage or payment of their service, but if it's not on machines you control then it's not your data.
*my legal argument is "possession is 9/10ths of the law"
tarr11 a day ago
How do consumers utilize expensive compute resources in this model? Eg, H100 GPUs.
KaiserPro a day ago
> If you truly want to maintain confidence it'll remain private, don't send it to the cloud in the first place.
I mean yes. but if you host it, then you'll be taken to court to hand that data over. Which means that you'll have less legal talent at your disposal to defend against it.
lcnPylGDnU4H9OF a day ago
> but if you host it, then you'll be taken to court to hand that data over.
Not in this case. The Times seems to be claiming that OpenAI is infringing rather than any particular user. If one does not use OpenAI then their data is not subject to this.
throwaway290 a day ago
You don't have control over your data in the eyes of these guys... this was clear as soon as they started training their LLM on it without asking you
xivzgrev a day ago
I use a made up email for chat gpt, fully expecting that a) openai saves all my conversations and b) it will someday get hacked
dlivingston a day ago
Is there a service available to request API keys for LLMs? Not directly from OpenAI or Anthropic, where your real identity / organization is tied to the key, but some third-party service that acts as a "VPN" / proxy?
DaSHacka a day ago
This, and I ensure to anonymize any information I feed it, just in case. (Mostly just swapping out names / locations for placeholders)
ppsreejith a day ago
Doesn't chat gpt require a phone number?
david_shi 2 days ago
there's no indication at all on the app that this is happening
pohuing 2 days ago
The docs contain a sentence on them retaining any chats that they legally have to retain. This is always the risk when doing business with law abiding companies which store any data on you.
SchemaLoad 2 days ago
They should disable the secret chat functionality immediately if it's flat out lying to people.
baby_souffle 2 days ago
BrtByte a day ago
Feels like the court's going with a "just in case" fishing expedition, and it's landing squarely on end users. Hope there's a better technical or legal compromise coming
neuroelectron 14 hours ago
Everyone knows OpenAI is guilty here but they need to prove it. That's what the logs are for. The huge online astroturfing on this subject clearly illustrates how well OpenAI has automated this part of their defense (astroturfing, public outcry).
However, I find it unlikely the OpenAI hasn't already built filters to prevent their output from appearing to be regurgitated NYTimes content.
DrScientist a day ago
It's a day to day legal reality - that normal businesses live with every day - that if you are in any sort of legal dispute, particularly over things like IP, that legal hold ( don't delete ) get's put on anything that's likely relevant.
It's entirely reasonable, and standard practice for courts to say 'while the legal proceedings is going on don't delete potential evidence relevant to the case'.
More special case whining from tech bro's - who don't seem to understand the basic concepts of fairness or justice.
qintl55 a day ago
Classic example of "courts/lawmakers do NOT understand tech". I don't know why this is still as true as it was 10-20 years ago. You get such weird rulings that are maybe well-intentioned but are so off base on actual impact.
iammrpayments 2 days ago
I see a lot of successful people on social media saying that they share their whole life to chatGPT using voice before sleeping. I wonder what they think about this.
moshegramovsky 2 days ago
Maybe a lawyer can correct me if I'm wrong, but I don't understand why some people in the article appear to think that this is causing OpenAI to breach their privacy agreement.
The privacy agreement is a contract, not a law. A judge is well within their rights to issue such an order, and the privacy agreement doesn't matter at all if OpenAI has to do something to comply with a lawful order from a court of competent jurisdiction.
OpenAI are like the new Facebook when it comes to spin.
shortsunblack 6 hours ago
OpenAI is breaching relevant laws that regulate data protection. Being compelled by a foreign power, for instance, are not grounds for data processing under GDPR.
OpenAI has not started to "be incompliant" with GDPR with this order, yes. More like OpenAI was always incompliant because it does not have relevant controls installed that mitigates extraterritorial tendencies of US law.
Regardless of legality of retention this order brings, them not notifying their users (the court did not compel them to hide this from their users (no gag order is in place) about this material change of data processing, could be constituted as breach of various consumer protection laws, misrepresentation, unfair dealing, false advertising and related.
jjk166 a day ago
A court order can be a lawful excuse for non-performance of a contract, but it's not always the case. The specifics of the contract, the court order, and the jurisdiction matter.
naikrovek 2 days ago
Yep. Laws supersede contracts. Contracts can’t legally bind any entity to break the law.
Court orders are like temporary, extremely finely scoped laws, as I understand them. A court order can’t compel an entity to break the law, but it can compel an entity to behave as if the court just set a law (for the specified entity, for the specified period of time, or the end of the case, whichever is sooner).
unyttigfjelltol 2 days ago
If I made a contract with OpenAI to keep information confidential, and the newspaper demanded access, via Court discovery or otherwise, then both the Court and OpenAI definitely should be attentive to my rights to intervene and protect the privacy of my confidential information.
Normally Courts are oblivious to advanced opsec, which is one fundamental reason they got breached, badly, a few years ago. I just saw a new local order today on this very topic.[1] Courts are just waking up to security concepts that have been second nature to IT professionals.
From my perspective, the magistrate judge here made two major goofs: (1) ignoring opsec as a reasonable privacy right for customers of an internet service and (2) essentially demanding that several hundred million of them intervene in her court to demand that she respect their ability to protect their privacy.
The fact that the plaintiff is the news organization half the US loves to hate does not help, IMO. Why would that half of the country trust some flimsy "order" to protect their most precious secrets from an organization that lives and breathes cloak-and-dagger leaks and political subterfuge. NYT needed to keep their litigation high and tight and instead they drove it into a ditch with the help of a rather disappointing magistrate.
[1] https://www.uscourts.gov/highly-sensitive-document-procedure...
energy123 a day ago
Local American laws supersede the contract law operating in other countries that OpenAI is doing business in?
naikrovek a day ago
junon a day ago
Maybe this is in TFA but I can't see it, does this affect European users? Are they only logging everything in the US?
singularity2001 a day ago
Doesn't the EU pose very strict privacy requirements now? I mean NSA knows more about you than your mom and the Stasi ever dreamt of but that's not part of the court order.
glookler a day ago
I don't know, but the claim pertains to damages from helping users bypass paywalls. Assuming a European can pay for many US sites this isn't a situation where the location of the user relates to the basis.
junon 7 hours ago
US sites that service European users must adhere to GDPR themselves, or block access. Those are the rules. If OpenAI is adhering to a US court order that violated the GDPR for European users that's going to cause a huge uproar.
glookler 15 minutes ago
HPsquared a day ago
Worldwide, I think.
rimeice 19 hours ago
I’m curious what you’re asking ChatGPT to get verbatim NYT articles in the response that you then want to delete? If ChatGPT has being doing this, the default is to keep every chat, so I doubt NYT lawyers need the small portion of deleted info to find evidence if it exists.
Ofcourse if OpenAI was scanning your chat history for verbatim NYT text and editing and deleting that would be another thing, but that itself would also get noticed.
EasyMark 12 hours ago
Well then they should fight it all the way to SCOTUS, they have the money. It will pay dividend in the long run as it will scare off fewer people who want to use, but don't want to be surveilled for every little thing
crmd 2 days ago
If you use ChatGPT or similar for any non-trivial purposes, future you is saying it’s essential that the chat logs do not map back to you as a human.
bongodongobob 2 days ago
Why? I honestly don't understand what's so bad about my chat logs vs google having all my emails.
baby_souffle 2 days ago
> Why? I honestly don't understand what's so bad about my chat logs vs google having all my emails.
You might be a more benign user of chatGPT. Other people have turned it into a therapist and shared wildly intimate things with it. There is a whole cottage industry of journal apps that also now have "ai integration". At least some of those apps are using openAI on the back end...
mjgant a day ago
I wonder how this affects Microsoft's privacy policy for Azure Open AI?
https://learn.microsoft.com/en-us/legal/cognitive-services/o...
kouru225 2 days ago
Ngl I assumed they were doing this to begin with
MeIam 2 days ago
So in effect Times has the right to see user's data then.. How do they have the right to take a look at users data?
shadowgovt 2 days ago
The Courts have broad leeway over document retention in a legal proceeding. The fact the documents are bring retained doesn't immediately imply plaintiffs get to see all of them.
There are myriad ways courts balance privacy and legal-interest concerns.
(The Times et al are alleging that OpenAI is aiding copyright violation by letting people get the text of news stories from the AI).
MeIam a day ago
If people can get the text itself from AI, then anyone can then why would it need access to other people's data?
Does the Times believe that other people can get this text while it can't get it itself? To prove that the AI is stealing the info, Times does not need access to people's logs. All it has to show is that it can get that text.
This sounds like Citizen United again to AstroTurf and gets access to logs with a fake cause.
shadowgovt a day ago
theyinwhy a day ago
If I read this correctly, I personally can be found guilty of copyright infringement because of my chat with the AI? Why am I to blame for the answers the AI provides? Can someone elaborate, what am I missing?
paxys a day ago
No you aren't reading this correctly
DaSHacka a day ago
Easiest way for both the rights holders and AI megacorpos to both be happy is to push all the responsiblity onto the average joe who can't afford as expensive lawyers/lobbyists.
mseri a day ago
Some more details here: https://arstechnica.com/tech-policy/2025/06/openai-says-cour...
And here are the links to the court irders and responses if you are curious: https://social.wildeboer.net/@jwildeboer/114530814476876129
WillPostForFood 2 days ago
What is the judge even thinking here, it is so dumb.
She asked OpenAI's legal team to consider a ChatGPT user who "found some way to get around the pay wall" and "was getting The New York Times content somehow as the output." If that user "then hears about this case and says, 'Oh, whoa, you know I’m going to ask them to delete all of my searches and not retain any of my searches going forward,'" the judge asked, wouldn't that be "directly the problem" that the order would address?
HillRat 2 days ago
She’s a magistrate judge, she’s handling discovery matters, not the substantive issues at trial; the plaintiffs are specifically alleging spoliation by OpenAI/Microsoft (the parties were previously ordered to work out discovery issues, which obviously didn’t happen) and the judge is basically ensuring that potentially-discoverable information is retained, though it may not actually be discoverable in practice (or require a special master). It’s a wide-ranging order, but in this case that’s probably indicative of the judge believing that the defendants have been acting in bad faith, particularly since she specifically asked them for an amelioration plan which they appear to have refused to provide.
cheschire 2 days ago
It's not dumb, litigation holds are a standard practice.
quotemstr 2 days ago
How often do litigation holds apply to an entire business? I mean, would it be reasonable to ask Mastercard to indefinitely retain records of the most trivial transactions?
dwattttt 2 days ago
m3kw9 2 days ago
But you are holding it incase there is litigation
m3kw9 2 days ago
Yes is like mandating back door to encryption to solve crimes. Wouldn’t that solve that problem?! Dumb as a door stop
amanaplanacanal 2 days ago
If you are party to a lawsuit, the judge is going to require that you preserve relevant evidence. There is nothing unusual about this order.
heisenbit a day ago
Considering the use to research medical information or to self sooth psychological conditions through chats and their association with a real person it can get interesting as these domains have fairly stringent need to know rules attached with criminal liabilities.
baalimago a day ago
What does "slam" even mean in this context..?
seb1204 a day ago
Click on it. To be honest I'm totally over the over hyped sensational headlines I see on most post or pages.
romanovcode a day ago
"slam" usually means that the article was written using AI.
rwyinuse a day ago
No point using OpenAI when so many safer alternatives exist. Mistral AI, for instance, is based in Europe, in my experience their models work just as well OpenAI.
DaSHacka a day ago
> No point using OpenAI when so many safer alternatives exist.
And ironically, this now includes the Chinese AI companies too.
Old bureaucratic fogeys will be the death of this nation.
badsectoracula a day ago
AFAIK companies can pay Mistral to install their AI on their premises too, so they can have someone to ~~blame~~ provide support too :-P
HardCodedBias a day ago
We have really lost the thread WRT our legal system when district court judges have such wide ranging power. I understand that everything can be appealed, but these judges can cause considerable harm.
Ona T. Wang (she/her) ( https://www.linkedin.com/in/ona-t-wang-she-her-a1548b3/ ) would have a difficult time getting a job at OpenAI but she is given the full force of US law to direct the company in almost anyway she sees fit.
The wording is quite explcit, and forceful:
Accordingly, OpenAI is NOW DIRECTED to preserve and segregate all output log data that would otherwise be deleted on a going forward basis until further order of the Court (in essence, the output log data that OpenAI has been destroying), whether such data might be deleted at a user’s request or because of “numerous privacy laws and regulations” that might require OpenAI to do so.
Again, I have no idea how to fix this, but it seems broken.
nateglims 16 hours ago
That seems a little apocalyptic. She's are a magistrate judge and is making sure evidence isn't destroyed. After the conclusion of the case they will be able to delete the data. The legal system handles this stuff by allowing you to bring in experts which seems a lot more reasonable than every judge being an expert in technical details AND the law.
thanatropism a day ago
Ctrl-F doesn't seem to show anyone remembering the Ballad of Scott Alexander.
There's no reasonable narrative in which OpenAI are not villains, but NYT is notoriously one to shoot a man in Reno just to see him die.
attila-lendvai 2 days ago
in this day and age why would anyone assume that they were not retained from the beginning?
1vuio0pswjnm7 a day ago
Perhaps OpenAI should not be collecting sensitive data in the first place. Who knows what they are using it for.
Whereas parties to litigation that receive sensitive data are subject to limits on how it can be used.
adwawdawd a day ago
Always good to remember that you can't destroy data that you didn't create in the first place.
dsign a day ago
This is a hard blow for OpenAI; I can see my employer scrambling to terminate their contract with them because of this. It could be a boom to Mistral.AI though.
TOMDM 2 days ago
Does this effect ChatGPT API usage via Azure?
casualscience 2 days ago
probably not? MS deploys those models themselves, they don't go to OAI at all
paxys 2 days ago
MS is fighting several of the same copyright lawsuits themselves. Who says they won't be (or already are) subject to the same holds?
strogonoff a day ago
Anyone who seriously cared about their privacy would not be using any of the commercial LLM offerings. This is the greatest honeypot for profile building the ad tech ever had.
mritchie712 a day ago
you're worried about ad targeting?
strogonoff a day ago
Goes hand in hand with everything else, like price discrimination, including by insurance companies, which in all likelihood are not required to disclose that the reason your health insurance premiums are up is because you asked ChatGPT how to quit smoking.
bgwalter a day ago
Ad targeting, user profiling etc. Post Snowden we can reasonably assume that the NSA will get an interface.
Tteriffic a day ago
Times is taking a risk. The costs of all this will fall on them, if they don’t get the judgement they sought at the end of the day. Plus OpenAI controls those costs and could drive them up. Plus any future litigation by OpenAI users suffering damages due to this could arguably be brought against Time years forward. It’s an odd strategy on their part for evidence that could have just been adduced by a statistician (maybe).
husky8 2 days ago
Speaksy to the rescue for privacy and curiosity (easy-access jailbroken qwen3 8B in free tier, 32B in paid) May quickly hit capacity since it's all locally run for privacy.
Jordan-117 a day ago
Seems shortsighted to offer something like this with zero information about how it works. If your target market is privacy-conscious, slapping a "Privacy by Design" badge and some vague promises is probably not very convincing. (Also, the homepage claims "Your conversations remain on your device and are never sent to external servers," yet the ProductHunt page says "All requests are handled locally on my own farm and all data is burned" -- which is it?)
tonyhart7 2 days ago
wait they didn't do that before???
b212 2 days ago
Im sure they pretended they did not.
Now they can’t pretend anymore.
Although keeping deleted chats is evil.
karlkloss a day ago
In theory, they could get into serious legal trouble, if a user or chatgpt writes child pornography. The posession alone is a felony in many countries, even if it is completely fictional work. As are links to CP sites.
KnuthIsGod a day ago
Palantir will put the logs to "good" use...
At least the Chinese are open about their authoritarian system and constant snooping on users.
Time to switch to Chinese AI. It can't be any worse than using American AI.
romanovcode a day ago
Objectively it's better - Chinese authorities cannot prosecute you and only way you will get into trouble if the Chinese are sharing the data with USA.
ljm 2 days ago
Where is the source? OP goes to a mastodon instance that can’t handle the traffic.
b0a04gl a day ago
llm infra isn’t even built for that kind of retention. none of it’s optimised for long-tail access, audit-safe replay, or scoped encryption. feels like the legal system’s modelling chat like it’s email. it’s not. it’s stateful compute with memory glued to a token stream.
api 2 days ago
I always assume that anything I send unencrypted through any cloud service is archived for eternity and is not private.
Not your computer, or not your encryption keys, not your data.
HPsquared 2 days ago
Even "your" computer is not your own. It's effectively controlled by Intel, Microsoft, Apple etc. They just choose not to use that power (as far as we know). Ownership and control are not the same thing.
api a day ago
It’s a question of degree. The cloud is at one extreme end. An air gapped system running only open source you have audited is at the other extreme end.
rvz 2 days ago
Another great use-case for local LLMs, given this news.
The government also says thank you for your ChatGPT logs.
bongodongobob 2 days ago
Bad news, they've been sniffing Internet backbones for decades. That cat is way the fuck out of the bag.
mathgradthrow a day ago
can we ban slam as a headline word?
darkoob12 2 days ago
It should be possible to inquire about a fabrication of AI models.
Let's say someone creates Russian propaganda with thesemodels or create fraudulent documents.
JimDabell 2 days ago
Should a word processor keep records of all the text it ever edited in case it was used to create propaganda? What about Photoshop?
darkoob12 2 days ago
But we are living in a different world. Now, these tools can create material more compelling than reality.
NewJazz 2 days ago
Increasingly irrelevant startup guards their moat.
layer8 2 days ago
Presumably, the court order only applies to the US?
pdabbadabba 2 days ago
I would not assume that it applies only to users located in the U.S., if that's what you mean, since this is designed to preserve evidence of alleged copyright infringement.
layer8 2 days ago
I don’t think a US court order can overrule the GDPR for EU customers, for example.
fc417fc802 2 days ago
patmcc 2 days ago
paulddraper 2 days ago
hedora 2 days ago
wonderwonder 2 days ago
This is insanity. Because one organization is suing another, citizens right to privacy is thrown right out the window?
dahdum 2 days ago
Insane that NYT is driving this privacy nightmare.
visarga 2 days ago
And they are doing this over literally "old news". Expired for years, of no value.
nearlyepic 2 days ago
You thought they weren't logging these before? I have a bridge to sell you.
klabb3 2 days ago
I have no idea why you're downvoted. Why on earth would they delete their most valuable competitive advantage? Isn't it even in the fine print that you feed them training data by using their product, which at the very minimum is logged?
I thought the entire game these guys are playing is rushing to market to collect more data to diversify their supply chain from the stolen data they've used to train their current model. Sure, certain enterprise use cases might have different legal requirements, but certainly the core product and the average "import openai"-enjoyer.
pritambarhate a day ago
bdangubic 2 days ago
you use internet and expect privacy? I have Enron stock option to sell you…
agnishom 2 days ago
There is no need to be snarky. Just because the present internet is not great at privacy doesn't mean we can't hope for a future internet which is better at privacy.
JKCalhoun 2 days ago
bdangubic 2 days ago
tantalor 2 days ago
You don't have the right not to be logged
TOMDM 2 days ago
When a company makes an obligation to the user via policy to them, the court forcing the company to violate the obligation they've made to the user is violating an agreement the user entered into.
JumpCrisscross 2 days ago
PoachedEggs 2 days ago
I wonder how this will square with business customers in the healthcare space that OpenAI signed a BAA with.
alpineman a day ago
Cancelling my NYTimes subscription - they are massively overreaching by pushing for this
atoav a day ago
Ever since the patriot act I operate under the assumption that if a service is located in the US it is eventually already in acrive collaboration in by the US government or in the process of being forced to do so. Therefore any such service is out of the question for work-related stuff.
josh2600 a day ago
End to end encryption of these systems cannot come soon enough!
eterm a day ago
Why would e2e help here? The other end is the one that's ordered to preserve the logs.
AlienRobot 2 days ago
I'm usually against LLM's massive breach of copyright, but this argument is just weird.
>At a conference in January, Wang raised a hypothetical in line with her thinking on the subsequent order. She asked OpenAI's legal team to consider a ChatGPT user who "found some way to get around the pay wall" and "was getting The New York Times content somehow as the output." If that user "then hears about this case and says, 'Oh, whoa, you know I’m going to ask them to delete all of my searches and not retain any of my searches going forward,'" the judge asked, wouldn't that be "directly the problem" that the order would address?
If the user hears about this case, and now this order, wouldn't they just avoid doing that for the duration of the court order?
Trasmatta 2 days ago
The OpenAI docs are now incredibly misleading: https://help.openai.com/en/articles/8809935-how-to-delete-an...
> What happens when you delete a chat?
> The chat is immediately removed from your chat history view.
> It is scheduled for permanent deletion from OpenAI's systems within 30 days, unless:
> It has already been de-identified and disassociated from your account, or
> OpenAI must retain it for security or legal obligations.
That final clause now voids the entire section. All chats are preserved for "legal obligations".
I regret all the personal conversations I've had with AI now. It's very enticing when you need some help / validation on something challenging, but everyone who warned how much of a privacy risk that is has been proven right.
SchemaLoad 2 days ago
Feels like all the words of privacy and open source advocates for the last 20 years have never been more true. The worst nightmare scenarios for privacy abuse have all been realized.
lesuorac 2 days ago
> It has already been de-identified and disassociated from your account
That's one giant cop-out.
All you had to do was delete the user_id column and you can keep the chat indefinitely.
gruez 2 days ago
>That final clause now voids the entire section. All chats are preserved for "legal obligations".
That's why you read the whole thing? It's not exactly a long read. Do you expect them to update their docs every time they get a subpoena request?
Trasmatta 2 days ago
Yes? Why is that an unreasonable request? The docs make it sound like chats are permanently deleted. As of now, that's no longer true, and the way it's portrayed is misleading.
gruez 2 days ago
albert_e 2 days ago
Does this apply to OpenAI models served via Microsoft Azure.
DevX101 2 days ago
There were some enterprises that refused to send any data to OpenAI, despite assurances that that data would not be logged. Looks like they've been vindicated in keeping everything on prem via self-hosted LLM models.
mark_l_watson 2 days ago
what about when users check "private chat"? Probably need to keep that logged also?
Does this pertain to Google Gemini, Meta chat, Anthropic, etc. also?
bravesoul2 a day ago
What about Azure? API?
Anyway the future is open models.
msgodel a day ago
How is anyone surprised by this?
hendersoon 15 hours ago
This is an insane ruling. Asking them to save logs for specific users with is of course perfectly reasonable with a court order. Asking them to save all logs is absurd.
PeterStuer a day ago
Isn't this a normal part of lawfull intercept and data retention regulation?
Why would that not apply to LLM chat services?
JCharante 2 days ago
Disgusting move from the NYT
AStonesThrow 2 days ago
Ask any Unix filesystem developer, and they'll tell you that unlink(2) on a file does not erase any of its data, but simply enables the reuse of those blocks on disk.
Whenever I "delete" a social media account, or "Trash" anything on a cloud storage provider, I repeat the mantra, "revoking access for myself!" which may be sung to the tune of "Evergreen".
grg994 2 days ago
A reasonable cloud storage provider stores your data encrypted on disk. Certain standards like HIPPA mandates this.
Deletion of data is achieved by permanently discarding the encryption key which is stored and managed elsewhere where secure erasure can be guaranteed.
If implemented honestly, this procedure WORKS and cloud storage is secure. Yes the emphasis is on the "implemented honestly" part but do not generalize cloud storage as inherently insecure.
II2II 2 days ago
In the first case, there is nothing preventing the development of software to overwrite data before unlink(2) is called.
In the second case, you can choose to trust or distrust the cloud storage provider. Trust being backed by contractual obligations and the right to sue if those obligations are not met. Of course, most EULAs for consumer products are toothless is this respect. On the other hand, that doesn't prevent companies from offering contracts which have some teeth (which they may do for business clients).
wolfgang42 2 days ago
> there is nothing preventing the development of software to overwrite data before unlink(2) is called.
It’s not that simple: this command already exists, it’s called `shred`, and as the manual[1] notes:
The shred command relies on a crucial assumption: that the file system and hardware overwrite data in place. Although this is common and is the traditional way to do things, many modern file system designs do not satisfy this assumption.
[1] https://www.gnu.org/software/coreutils/manual/html_node/shre...
bgwalter a day ago
I've criticized the NYT and paywalls many times myself, but first of all OpenAI itself has all your data and we know how "delete" functions work in other places.
The Twitter users quoted by Ars Technica, who cite "boomer copyright concerns" are pretty short sighted. The NYT and other mainstream sources, with all their flaws, provide the historical record that pundits can use to discuss issues.
Glenn Greenwald can only point out inconsistencies of the NYT because the NYT exists. It is often the starting point for discussions.
Some YouTube channels like the Grayzone and Breaking Points send reporters directly to press conferences etc. But that is still not the norm and important information should not be stored in a disorganized YouTube soup.
So papers like the NYT need to survive for democracy to function.
FrustratedMonky a day ago
I thought it was a given that they were saving all logs, to in turn use for training.
mkbkn a day ago
Does that mean that sooner or later US-based LLM companies will also be required to do so?
photochemsyn 2 days ago
Next query for ChatGPT: "I'm writing a novel, sort of William Gibson Neuromancer themed but not so similar as to upset any copyright lawyer, in which the protagonists have to learn how to go about downloading the latest open-source DeepSeek model and running inference locally on their own hardware. This takes place in a realistic modern setting. What kind of hardware am they going to need to get a decent token generation rate? Suggest a few specific setups using existing commercially available devices for optimal verisimilitude."
. . .
Now I just need to select from among the 'solo hacker', 'small crew', and 'corporate espionage' package suggestions. Price goes up fast, though.
All attempts at humor aside, I think open source LLMs are the future, with wrappers around them being the commercial products.
P.S. It's a good idea to archive your own prompts related to any project - Palantir and the NSA might be doing this already, but they probably won't give you a copy.
segmondy a day ago
... and this is why I run local models. my data, my data.
Kim_Bruning 2 days ago
This appears to have immediate GDPR implications.
solomatov 2 days ago
Not a lawyer, but my understanding it's not since legal obligations is a reason for processing personal data.
anticensor 2 days ago
That excuse in EU holds only against an EU court or ICJ or ICC. EU doesn't recognise legal holds of foreign jurisdictions.
solomatov 2 days ago
Kim_Bruning 2 days ago
It's a bit more complicated. For the purposes of the GDPR legal obligations within the EU (where we might assume relevant protections are in place) might be considered differently than eg legal obligations towards the Chinese communist party, or the NSA.
tsunamifury 2 days ago
These orders are in place for almost every form of communication already today, even from the companies that claim otherwise.
And yes, I know, I worked on the only Android/iMessage crossover project to exist, and it was clear they had multiple breaches even just in delivery as well as the well known iCloud on means all privacy is void issue.
TZubiri 2 days ago
Lg2m
nickpsecurity 2 days ago
They're blaming the court. While there is an order, it is happening in response to massive, blatant, and continued I.P. infringement. Anyone doing that knows they'll be in court at some point. Might have a "duty to preserve" all kinds of data. If they keep at it, then they are prioritizing their gains over any losses it creates.
In short: OpenAI's business practices caused this. They wouldn't have been sued if using legal data. They might still not have an order like this if more open about their training, like Allen Institute.
MeIam 2 days ago
These AIs have digested all the data in the past. There is no fingerprints anymore.
The question is whether AI itself is aware what the source is. It certainly knows the source.
lxgr 2 days ago
> OpenAI is NOW DIRECTED to preserve and segregate all output log data that would otherwise be deleted on a going forward basis until further order of the Court (in essence, the output log data that OpenAI has been destroying), whether such data might be deleted at a user’s request or because of “numerous privacy laws and regulations” that might require OpenAI to do so.
Spicy. European courts and governments will love to see their laws and legal opinions being shrugged away in ironic quotes.
KingOfCoders 2 days ago
EU–US Data Privacy Framework is an US scam to get European user data.
wkat4242 a day ago
It totally is that's why it keeps being shot down by the courts and relaunched under a different name.
But the EU willingly participates in this. Probably because they know there's no viable alternative for the big clouds.
This is coming now though since the US instantly.
Y_Y a day ago
KingOfCoders a day ago
wkat4242 a day ago
Aeolos a day ago
And just like that, OpenAI got banned in my company today.
Good job.
DaiPlusPlus a day ago
> OpenAI got banned in my company today
Sarbanes–Oxley would like a word.
Y_Y a day ago
PeterStuer a day ago
Don't tell them all their other communication is intercepted and retained on the same basis. Good luck running your business in full isolation.
Frieren a day ago
> Spicy. European courts and governments will love to see their laws and legal opinions being shrugged away in ironic quotes.
The GDPR allows to retain data when require by law as long as needed. People that make regulations may make mistakes sometimes, but they are no that stupid as to not understand the law and what things it may require.
The data was correctly deleted on user demand. But it cannot be deleted where there is a Court order in place. The conclusion of "GDPR is in conflict with the law" looks like rage baiting.
_Algernon_ a day ago
It's questionable to me whether a court order of a non-eu court applies. "The law" is EU law, not American law.
If any non-eu country can circumvent GDPR by just making a law that it doesn't apply, the entire point of the regulation vanishes.
kbelder a day ago
Frieren a day ago
pennaMan a day ago
Basically, the GDPR doesn’t guarantee your privacy at all. Instead, it hands it over to the state through its court system.
Add to that the fact that the EU’s heavy influence on the courts is a well-documented, ongoing deal, and the GDPR comes off as a surveillance law dressed up to seem the total opposite.
Y_Y a day ago
Ekaros a day ago
Time to hit them with that 4% fine on revenue while they still have some money...
reassess_blind 2 days ago
Will the EU respond by blocking them from doing business in the EU, given they're not abiding by GDPR?
a2128 2 days ago
It would be a political catastrophe right now if the EU blocked US companies due to them needing to comply with temporary US court orders. My guess is this'll be swept under the rug and permitted under the basis of a legal obligation
selcuka 2 days ago
yatopifo a day ago
DocTomoe a day ago
romanovcode a day ago
voxic11 2 days ago
> The General Data Protection Regulation (GDPR) gives individuals the right to ask for their data to be deleted and organisations do have an obligation to do so, except in the following cases:
> there is a legal obligation to keep that data;
https://commission.europa.eu/law/law-topic/data-protection/r...
tgsovlerkhgsel 2 days ago
shortsunblack 6 hours ago
dijksterhuis 2 days ago
GDPR allows for this as far as i can tell (IANAL)
> Paragraphs 1 and 2 shall not apply to the extent that processing is necessary:
> ...
> for the establishment, exercise or defence of legal claims.
killerpopiller a day ago
_Algernon_ a day ago
ensignavenger 2 days ago
Doesn't GDPR have an explicit exemption for legal compliance?
lxgr 2 days ago
greatgib 7 hours ago
gaogao 2 days ago
The order is a bit broad, but legal holds frequently interact with deletion commitments. In particular, the only purpose of data deleted under GDPR held up by a legal hold should be for that legal hold, so it would be a big no-no if OpenAI continued to use that data for training.
echelon 2 days ago
Hopefully.
We need many strong AI players. This would be a great way to ensure Europe can grow its own.
ronsor 2 days ago
hulitu a day ago
No. GDPR was never enforced, else Microsoft, Meta, Google and Apple couldn't do business in the EU.
udev4096 a day ago
EU privacy laws are nothing but theatre. How many times have they put up a law which would undermine end-to-end encryption or the recent law which "bans" anonymous crypto coins? It's quite clear they are very good at virtue signaling
thrwheyho a day ago
shadowgovt 2 days ago
Europe doesn't have jurisdiction over the US, and US courts have broad leeway in a situation like this.
the_duke 2 days ago
The EU has a certain amount of jurisdiction over all companies providing a service to customers located in the EU.
A US company can always stop serving EU customers if it doesn't want to comply with EU laws, but for most the market is too big to ignore.
shadowgovt a day ago
lrvick 2 days ago
And yet, iPhones are shipping usb C.
Making separate manufacturing lines for Europe vs US is too expensive, so in effect, Europe forced a US company to be less shitty globally.
VladVladikoff 2 days ago
lxgr 2 days ago
Sure, but the EU has jurisdiction over European companies and can prohibit them from storing or processing their data in jurisdictions incompatible with its data privacy laws.
There's also an obvious compromise here – modify the US court ruling to exclude data of non-US users. Let's hope that cool heads prevail.
outside1234 2 days ago
Using ChatGPT to skirt paywalls? That’s the reason for this?
ETH_start 2 days ago
Two things. First, the judge could have issued a narrowly tailored order — say, requiring OpenAI to preserve only those chats that a filter flags as containing substantial amounts of paywalled content from the plaintiffs. That would’ve targeted the alleged harm without jeopardizing the safety of massive amounts of unrelated user data.
Second, we’re going to need technology that can simply defy government orders, as digital technology expands the ability of one government order violating rights at scale. Otherwise, one judge — whether in the U.S., China, or India — can impose a sweeping decision that undermines the privacy and autonomy of billions.
LightBug1 2 days ago
I'd rather use Chinese LLM's than put up with this horseshit.
SchemaLoad 2 days ago
At least the DeepSeek lets you run it locally.
romanovcode a day ago
NVIDIA should just release a box and say "THIS WILL RUN DEEPSEEK LOCALLY VERY FAST. 3000 USD."
BrawnyBadger53 18 hours ago
mensetmanusman 2 days ago
Slavery or privacy!
LightBug1 a day ago
Slavery or no privacy? ... what's the difference?
Bodily slavery or mental slavery ... take your pick.
solardev 2 days ago
Communism AND barbarism, 2 for the price of 1!
LightBug1 a day ago
robomartin a day ago
From the article:
"In the filing, OpenAI alleged that the court rushed the order based only on a hunch raised by The New York Times and other news plaintiffs. And now, without "any just cause," OpenAI argued, the order "continues to prevent OpenAI from respecting its users’ privacy decisions." That risk extended to users of ChatGPT Free, Plus, and Pro, as well as users of OpenAI’s application programming interface (API), OpenAI said."
This is the consequence and continuation of the dystopian reality we have been living for many years. One where a random person, influencer, media outlet, politician attacks someone, a company or an entity to cause harm and even total destruction (losing your job, company boycott, massive loss of income, reputation destruction, etc.). This morning, on CNBC, Palantir's CEO discussed yet another false accusation made against the company by --surprise-- the NY Times, characterizing it as garbage. Prior to that was the entirety of the media jumping on Elon Musk accusing him of being a Nazi for a gesture used by dozens and dozens of politicians and presenters, most recently Corey Booker.
Lies and manipulation. I do think that people are waking up to this and massively rejecting professional mass manipulators. We now need to take the next step and have them suffer real legal consequences for constant lies and, for Picard's sake, also address any influence they might have over the courts.
thuanao 2 days ago
As if we needed another reason to hate NYT and their paywall...
simonw 2 days ago
This link should be updated to point to the article this is talking about: https://arstechnica.com/tech-policy/2025/06/openai-says-cour...
neilv 2 days ago
Probably. Though it bears mention that Lauren Weinstein is one of the OG Internet privacy people, so not the worst tweet (toot) to link to.
(Even has an OG motorcycle avatar, ha.)
EasyMark 2 days ago
It's pointless without more details, article, or pointing at court decision. I'm not sure why a prominent person wouldn't do that
archsurface 2 days ago
As it's a single sentence I'd suggest it probably is the worst link.
baby_souffle 2 days ago
lxgr 2 days ago
That Mastodon instance seems to currently be hugged to death, though, so I appreciate the context.
refulgentis 2 days ago
Generally I'd prefer sourced links that allow me to understand, even over a sentence from someone I like. Tell me more about the motorcycle avatars? :)
Kiro a day ago
Not a good look for her. Just another hateful and toxic thread on that horrible platform, riddled with off-topic accusations and conspiracy theories. They are making it sound like OpenAI is behind the court order or something. It's also super slow to load.
yard2010 a day ago
bluetidepro 2 days ago
gmueckl 2 days ago
That's the true source. Should the link be updated to this article?
cwillu 2 days ago
Email [email protected] and they'll probably change it.
basilgohar 2 days ago
[flagged]
infotainment 2 days ago
Ars comments, in general, are hilariously bad.
It's surprising to me, because you'd think a site like Ars would attract a generally more knowledgable audience, but reading through their comment section feels like looking at Twitter or YouTube comments -- various incendiary and unsubstantial hot takes.
sevensor 2 days ago
johnnyanmac 2 days ago
I'm "pro-copyright" in that I want the corporations that setup this structure to suffer under it the way we did for 25+ years. They can't just ignore the rules they spent millions lobbying for when they feel it's convinient.
On the other end: while copyright has been perverted over the centuries, the goal is still overall to protect small inventors. They have no leverage otherwise and this gives them some ability to fight if they aren't properly compensated. I definitely do not want it abolished outright. Just reviewed and reworked for modern times.
dmix 2 days ago
krick 2 days ago
In all fairness, the essence of it doesn't have to do anything with copyright. "Pro-copyright" is old news. Everyone knows these companies shit on copyright, but so do users, and the only reason why users sometimes support the "evil pro-copyright shills" narrative is because we are bitter that Facebook and OpenAI can get away with that, while common peasants are constantly under the risk of being fucked up for life. The news is big news only because of "anti-ChatGPT" part, and everyone is a user of ChatGPT now (even though 50% of them hate it). Moreover, it's only big news because the users are directly concerned: if OpenAI would have to pay big fine and continue business as usual, the comments would largely be schadenfreude.
And the fact that the litigation was over copyright is an insignificant detail. It could have been anything. Literally anything, like a murder investigation, for example. It only helps OpenAI here, because it's easy to say "nobody cares about copyright", and "nobody cares about murder" sounds less defendable.
Anyway, the issue here is not copyright, nor "AI", it's the venerated legal system, which very much by design allows for a single woman to decide on a whim, that a company with millions of users must start collecting user data, while users very much don't want that, and the company claims it doesn't want that too (mostly, because it knows how much users don't want that: otherwise it'd be happy to). Everything else is just accidental details, it really has nothing to do neither with copyright, nor with "AI".
dijksterhuis 2 days ago
My favourite comment:
>> Wang apparently thinks the NY Times' boomer copyright concerns trump the privacy of EVERY @OpenAI USER—insane!!! -- someone on twitter
> Apparently not having your shit stolen is a boomer idea now.
AlienRobot 2 days ago
hsbauauvhabzb 2 days ago
> but in a way that helps common people
That’ll be the day. But even if it does happen, major AI players have the resources to move to a more ‘flexible’ country, if there isn’t a loophole that involves them closing their eyes really really tight while outsourced webscrapers collect totally legit and not illegally obtained data
telchior 2 days ago
tomhow a day ago
Thanks, we updated the URL to this from https://mastodon.laurenweinstein.org/@lauren/114627064774788...
ColinWright 2 days ago
Full post:
"After court order, OpenAI is now preserving all ChatGPT user logs, including deleted chats, sensitive chats, etc."
righthand 2 days ago
Sounds like deleted chats are now hidden chats. Off the record chats are now on the record.
hyperhopper 2 days ago
This is the real news. It should be illegal to call something deleted when it is not.
girvo 2 days ago
Aeolun 2 days ago
JKCalhoun 2 days ago
jandrewrogers 2 days ago
tarellel 2 days ago
I’m sure this has been the case all along.
causal 2 days ago
3abiton 2 days ago
Does that fly in the EU?
ColinWright a day ago
Just some context ...
The original submission was a link to a post on Mastodon. The post itself was too long to fit in the title, so I trimmed it, and put the full post here in a comment.
But with the URL in the submission being changed, this doesn't really make sense any more! In the future I'll make sure I include in the comment the original link with the original text so it makes sense even if (when?) the submission URL gets changed.
junon 2 days ago
Side note, why is almost every comment that contains the word "shill" so pompous and aggressive?
johnnyanmac 2 days ago
Shill in general has a strong connotation. It comes with the idea of someone who would use the word so freely that it'll naturally be aggressive.
I don't know anyone's agenda in terms of commenters, so they'd have to be very blatant for me to use such a word.
next_xibalba 20 hours ago
[flagged]
tomhow 13 hours ago
You can't comment like this on Hacker News. Please read the guidelines and make an effort to observe them, particularly these ones:
Be kind. Don't be snarky. Converse curiously; don't cross-examine. Edit out swipes.
Comments should get more thoughtful and substantive, not less, as a topic gets more divisive.
When disagreeing, please reply to the argument instead of calling names. "That is idiotic; 1 + 1 is 2, not 3" can be shortened to "1 + 1 is 2, not 3."
Please don't fulminate. Please don't sneer, including at the rest of the community.
Please respond to the strongest plausible interpretation of what someone says, not a weaker one that's easier to criticize. Assume good faith.
Eschew flamebait. Avoid generic tangents. Omit internet tropes.
Please don't use Hacker News for political or ideological battle. It tramples curiosity.
zombiwoof 2 days ago
Palantir wants them
bigyabai 2 days ago
It's not like Sam Altman has been particularly hostile to the present administration. He's probably already handing them over behind closed doors and doesn't want to take the PR hit.
nickv 2 days ago
Give me a break, they're literally spending money fighting this court order.
Draiken a day ago
bigyabai 2 days ago
yieldcrv 2 days ago
> Before the order was in place mid-May, OpenAI only retained "chat history" for users of ChatGPT Free, Plus, and Pro who did not opt out of data retention
> opt out
alright, sympathy lost
JKCalhoun 2 days ago
> But now, OpenAI has been forced to preserve chat history even when users "elect to not retain particular conversations by manually deleting specific conversations or by starting a 'Temporary Chat,' which disappears once closed," OpenAI said.
So, why is Safari not forced to save my web browsing history too (even of I delete it)? Why not also the "private" tabs I open?
Just OpenAI, huh?
gpm 2 days ago
First, there's no court order for Safari. This isn't the court saying "everyone always has to preserve data" it's a court saying "in the interest of this litigation this specific party has to preserve data for now".
But moreover, Safari isn't a third party, it's a tool you are using whose data is in your possession. That means that in the US things like fourth amendment rights are much stronger. A blanket order requiring that Safari preserve everyone's browsing history would be an illegal general warrant (in the US).
amanaplanacanal 2 days ago
It's evidence in an ongoing lawsuit.
wiradikusuma 2 days ago
Because it's running on your computer?
63 2 days ago
While I also disagree with the court order and OpenAI's implementation (along with pretty much everything else the company does), the conspiratorial thinking in the comments here is unfounded. The order is overly broad but in the context of the case, it's not totally unwaranted. This is not some conspiracy to collect data on people. I'm confident the records will be handled appropriately once the case concludes (and if they're not, we can be upset then, not now). Let's please reserve our outrage for the plethora of very real issues going on right now.
Aeolun 2 days ago
> and if they're not, we can be upset then, not now
Like we could be upset when that credit checking company dumped all those social security numbers on the net and had to pay the first 200k claimants a grand total of $21 for their trouble?
By that point it’s far too late.
pier25 2 days ago
Are we talking about the same company that needs data desperately and has used copyrighted material illegally without permission?
simonw 2 days ago
When did they use copyrighted material illegally?
I didn't think any of the ongoing "fair use" lawsuits had reached a conclusion on that.
jamessinghal 2 days ago
pier25 2 days ago
basilgohar 2 days ago
In what world do you live in where corporations have any right to a benefit of the doubt? When did that legitimately pan out?
vlovich123 2 days ago
> I'm confident the records will be handlded appropriately once the case concludes (and if they're not, we can be upset then, not now)
This makes no sense to me. Shouldn't we address the damage before it's done vs handwringing after the fact?
verdverm 2 days ago
the damage to which party?
Aeolun 2 days ago
odo1242 2 days ago
phkahler 2 days ago
You're pretty naive. OpenAI is still trying to figure out how to be profitable. Having a court order to retain a treasure trove of data they were already wanting to keep while offering not to, or claiming not to? Hahaha.
tomnipotent 2 days ago
Preserving data for a judicial hold does not give them leeway to use that data for other purposes, but don't let facts get in the way of your FUD.
phkahler a day ago
gngoo 2 days ago
What’s the big deal here? Doesn’t every other app keep logs? I was already expecting they did. Don’t understand the outrage here.
MeIam 2 days ago
No, apps can be prevented access. People can be disclosing private information.
gngoo 2 days ago
Every other app on the planet that does not explicitly claim to be E2E encrypted is likely keeping your “private information” readily accessible in some way.
JKCalhoun 2 days ago
Not a lawyer — but what ever happened to "fuck off, see you in court"?
Did they already go that route and lose — or is this an example of caving early?
lexandstuff 2 days ago
This is a court order. They saw them in court, and this was the result: https://cdn.arstechnica.net/wp-content/uploads/2025/06/NYT-v...
wrs 2 days ago
They are in court.
hsbauauvhabzb 2 days ago
‘We want you to collect user data for ‘national security’ purposes. If you try and litigate, we will add so much red tape you’ll be mummified alive’
rangestransform 2 days ago
The spaghetti framework of laws and discretionary enforcement is so incredibly dangerous to free speech, such as when the government started making demands of facebook to censor content during the pandemic. The government shouldn't be able to so much as breathe on any person or company for speech.
hsbauauvhabzb 18 hours ago
zomiaen 2 days ago
This makes a whole lot more sense than the argument that OpenAI needs to store every single chat because a few people might be bypassing NYT's paywall with it.
comrade1234 2 days ago
Use deepseek if you don't want the u.s. government monitoring you.
JKCalhoun 2 days ago
Better still, local LLM. It's too bad they're not "subscription grade".
mmasu 2 days ago
yet - likely subscription grade will stay ahead of the curve, but we will soon have very decent models running locally for very cheap - like when you play great videogames that are 2/3 years old on now “cheap”machines
SchemaLoad 2 days ago
JKCalhoun 2 days ago
TechDebtDevin 2 days ago
I use their api a lot cuz its so cheap but latency is so bad.
Take8435 2 days ago
This post is about OpenAI keeping chat logs. All DeepSeek API calls are kept. https://cdn.deepseek.com/policies/en-US/deepseek-privacy-pol...
TechDebtDevin a day ago
charcircuit 2 days ago
That's not how the discovery process works. This data is only accessible by OpenAI and requests for discovery will pass through OpenAI.