Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".
-
@simonzerafa @raymaccarthy this is just a dumb attack on Cory as a person which I will not accept. You can talk about what he or I wrote (both things can be criticized) but have some respect
@tante @simonzerafa
A brilliant person isn't right about everything.
It's only a criticism of one view/idea. -
Checking for punctuation errors is does not discourage critical thinking. It's weird to laud "critical thinking" and also make this claim.
@pluralistic @simonzerafa on this one for example I fully agree with Cory. This is not him having a genAI system write or anything like that.
-
Good critique. And I say this, as well as like your arguments, while also believing there are certain domains where LLMs do have reasonable utility value - nowhere near the value required to make OpenAI, Antrophic or the rest profitable, but some value nonetheless.
@tante@osma@mas.to @tante@tldr.nettime.org It has debatable utility in some uses, but nowhere near enough to make the industry worth keeping around given the ethical concerns. The utility is effectively immaterial compared to the self-parody levels of evil on display from OpenAI and its ilk.
-
@raymaccarthy @simonzerafa @tante
Again, what does checking the punctuation on a single essay per day have to do with "play[ing] Russian Roulette every day for hours?"
@pluralistic I'd be disappointed if I didn't see myself in the pattern of engaging with people on a post like this who are worlds away from having a fair discussion...
They literally can't see the reality of AI beyond their arguments, they've decided it's inherently evil and wrong and locked in their viewpoint.
So their "russian roulette every day for hours" is because, despite you saying what you use it for, they can't comprehend how it can be used outside of the worst possible use cases.
Same reason they're accusing you of being a libertarian, but that's already the purity culture you were originally calling out.
-
@osma@mas.to @tante@tldr.nettime.org It has debatable utility in some uses, but nowhere near enough to make the industry worth keeping around given the ethical concerns. The utility is effectively immaterial compared to the self-parody levels of evil on display from OpenAI and its ilk.
Whatever I just wrote, thanks. Don't see why we should debate it.
@flesh -
> I am not clear on how this connects to discussing origins of technologies
Because the arguments against running an LLM on your own computer boil down to, "The LLM was made by bad people, or in bad ways."
This is a purity culture standard, a "fruit of the poisoned tree" argument, and while it is often dressed up in objectivity ("I don't use the fruit of the poisoned tree"), it is just special pleading ("the fruits of the poisoned tree that I use don't count, because __").
-
> I am not clear on how this connects to discussing origins of technologies
Because the arguments against running an LLM on your own computer boil down to, "The LLM was made by bad people, or in bad ways."
This is a purity culture standard, a "fruit of the poisoned tree" argument, and while it is often dressed up in objectivity ("I don't use the fruit of the poisoned tree"), it is just special pleading ("the fruits of the poisoned tree that I use don't count, because __").
> Almost everyone using LLMs will use the online kind, so objections to LLMs are (reasonably IMHO) based on that scenario.
Except that in this specific instance, you are weighing on an article that claims that it is wrong to run a local LLM for the purposes of checking for punctuation errors.
-
@pluralistic @simonzerafa on this one for example I fully agree with Cory. This is not him having a genAI system write or anything like that.
@tante @pluralistic @simonzerafa I agree in principle with Cory, but I really wish that he had clarified that:
1. Ollama is not an LLM, it's a server for various models, of varying degrees of openness.
2. Open weights is not open source, the model is still a black box. We should support projects like OLMO, which are completely open, down to the training data set and checkpoints.
3. It's quite difficult to "seize that technology" without using Someone Else's Computer to do so (a.k.a clown/cloud) -
There is no virtue in being constrained or regulated per se.
Regulation isn't a good unto itself.
Regulation that is itself good - drawn up for a good purpose, designed to be administrable, and then competently administered - is good.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante It seems to me Doctorow is obviously correct about this. But I don't think it matters too much if you don't agree... the trajectory of LLMs is going to be whatever it is going to be.
If you don't like it and have buddies that don't like it either, that's not a bad thing especially if you are undergoing real negative effects from it.
It's just if you stray from reality (whatever that will be) too far for too long, you will end up with a big shock when forced to rejoin it.
-
@skyfaller I think you should be able to answer these questions yourself, but clearly are struggling...
On your mink fur argument: the one ethical way to wear something like that is to only purchase used and old. The harm is done regardless of whether you purchase, you don't increase demand because your refusal to purchase new or recent means there's no profit in it. (This argument is also flawed because it's assuming local LLMs are made for profit when no profit is made on them)
And on your Luddite argument: When someone is using a machine to further oppress workers, the issue is not the machine but the person using it. You attack the machine to deprive them of it. But when an individual is using a completely separate instance of the machine, contributing nothing to those who are using the machine to abuse people... attacking them is simply attacking the worker.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante Didn't we already have spelling and grammar tools before everyone decided LLMs needed to be pushed at every problem?
-
@pluralistic Thank you for the fact check. I was paraphrasing that text from the popular Nib comic: https://thenib.com/im-a-luddite/
If this contains factual inaccuracies I will need to do more research and perhaps stop sharing that comic.
-
@pluralistic Thank you for the fact check. I was paraphrasing that text from the popular Nib comic: https://thenib.com/im-a-luddite/
If this contains factual inaccuracies I will need to do more research and perhaps stop sharing that comic.
@skyfaller @FediThing @tante I strongly recommend Brian Merchant's "Blood in the Machine" as the best modern history of the Luddites.
-
@Colman @FediThing @tante That's interesting. I've never wondered that about you.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante cory is, at his heart, a conservative/liberal USian, putting him far to the right of mainstream European thought and politics.
He constantly refuses to apply his beliefs to underlying structures, arguing that AI or enshittification are aberrations in capitalism, refusing to acknowledge and blocking anyone who argues that it's just capitalism acting as intended.
It doesn't surprise me at all that he's acting hypocritically here.
-
@pluralistic I don't think mink fur or LLMs are comparable to criticizing the origins of the internet or transistors. It's the process that produced mink fur and LLMs that is destructive, not merely that it's made by bad people.
For example, LLM crawlers regularly take down independent websites like Codeberg, DDoSing, threatening the small web. You may say "but my LLM is frozen in time, it's not part of that scraping now", but it would not remain useful without updates.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante Frankly, I don't think there are any ethical concerns with how he's using it.
The reason AI is a violation when it trains on openly available data and then outputs similar stuff is that it's creating derivative works. Something that reads everything produced by man and then uses that information to score similar output does NOT. It's completely fair use and it's a GOOD application of AI.
IFF all the evil crap that the people who made it are up to wasn't a concern there'd be none.
-
I am astonished that I have to explain this,
but very simply in words even a small child could understand:
using these products *creates further demand*
- surely you know this?
Well, either you know this and are being facetious, or you are a lot stupider than I ever thought possible for someone with your privilege and resources.
I am absolutely floored at this reveal, just wow, "where's Cory and what have you done with him?"

Massive loss of respect!
@kel it sounds like your respect is rooted only in someone agreeing with you. If you respected them you'd maybe take a minute to listen to their arguments and ask yourself more about why they might disagree with you.
Namely the fact that you don't understand how "using these products creates further demand" doesn't relate to their arguments at all?
-
@pluralistic I don't think mink fur or LLMs are comparable to criticizing the origins of the internet or transistors. It's the process that produced mink fur and LLMs that is destructive, not merely that it's made by bad people.
For example, LLM crawlers regularly take down independent websites like Codeberg, DDoSing, threatening the small web. You may say "but my LLM is frozen in time, it's not part of that scraping now", but it would not remain useful without updates.
No. Literally the same LLM that currently finds punctuation errors will continue to do so. I'm not inventing novel forms of punctuation error that I need an updated LLM to discover.