The LLM discourse on the Fediverse has really irked me the last few days.
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse
ALL OF THIS! -
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse If you couldn't be bothered to write it, why should I be bothered to read it?
-
> can be useful to digest and explore information at great speed
Nope. Still wrong. This is in fact something they are extremely and *dangerously* bad at.
@lproven @xs4me2 @reading_recluse
For generating content of any kind, I think there's a reckoning to come. Especially in the 'agentic' space.
But for Information Retrieval, LLMs are great, tbh... I'd argue that also includes those far out stories about prompts leading to new scientific theories, or mathematical proofs.
The tool is a big part of that, but it's the user ('operator'?) that writes the prompts, guides the outcomes, and validates them.
That's a worthy advance.
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse They’re also built on the exploitation of the Global South: https://stephvee.ca/blog/artificial%20intelligence/generative-ai-is-built-on-the-exploitation-of-the-global-south/
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse @Furthering Well said! With you 100%.
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
-
@lproven @xs4me2 @reading_recluse
For generating content of any kind, I think there's a reckoning to come. Especially in the 'agentic' space.
But for Information Retrieval, LLMs are great, tbh... I'd argue that also includes those far out stories about prompts leading to new scientific theories, or mathematical proofs.
The tool is a big part of that, but it's the user ('operator'?) that writes the prompts, guides the outcomes, and validates them.
That's a worthy advance.
The problem is that LLMs just make things up. There are no new discovers, there is no accurate information retrieval. But people don't notice, because they lack the expertise, they lack the ability to check.
LLMs cannot be trusted with anything. They are a sheer waste of our world's resources.
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse I can relate to your stance. But ultimately decide to take action demanding attribution and compensation for the unpaid labor and externalities that goes into LLMs development. Have you considered engaging from that perspective?
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse Absolutely. LLMs are the biggest, most bloody useless con ever invented by the vacuous arseholes in charge of the tech industry.
The extra annoying thing is that there are other potential approaches to AI out there that are ultimately likely to be more useful, less destructive and work better (e.g. some expert systems, decision support systems, etc.) But so many folks are just playing with probabilistic horseshit generators instead.
-
> can be useful to digest and explore information at great speed
Nope. Still wrong. This is in fact something they are extremely and *dangerously* bad at.
Well as I said it is a tool, a hammer is not right or wrong. It can be used right or wrong.
As a domain expert, I use LLM in my work, but I will always judge and validate if it is right... I have indeed seen colleagues use it out of their zone of work, where I had to tell them yes this is right what LLM said, but not in this context. The real problem is LLM will never tell you context or probability of it telling you something is correct.
-
The problem is that LLMs just make things up. There are no new discovers, there is no accurate information retrieval. But people don't notice, because they lack the expertise, they lack the ability to check.
LLMs cannot be trusted with anything. They are a sheer waste of our world's resources.
@Firlefanz @dynamite_ready @lproven @reading_recluse
Well, yes and no, see my reply below:
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse I do agree, but I'd like to add something. After all, the manipulative scheme on users isn't much different from what has happened in the last twentysomething years. The companies behind it are still the same ones, almost all of them were born less than three decades ago.
LLMs have just refined the decoy, polished the deceptive honey-pot.
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse that's a great point, you're right to point that out, and you've touched on a classic issue between humanity and LLMs
-
@lproven @xs4me2 @reading_recluse
For generating content of any kind, I think there's a reckoning to come. Especially in the 'agentic' space.
But for Information Retrieval, LLMs are great, tbh... I'd argue that also includes those far out stories about prompts leading to new scientific theories, or mathematical proofs.
The tool is a big part of that, but it's the user ('operator'?) that writes the prompts, guides the outcomes, and validates them.
That's a worthy advance.
@dynamite_ready @lproven @reading_recluse
It is the user and their skills indeed. A hammer can be used skillfully or wrong...
-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse My first thought was that the people wittering on about "purity culture" literally can't grasp the concept of collective action. But then it struck me that framing everything as an individual choice is a classic neoliberal tactic to defuse and dismantle opposition when it becomes a threat. So I say: Good work, keep it up!

-
The LLM discourse on the Fediverse has really irked me the last few days.
Refusing to read writing made with the use of LLMs and refusing to give time to writers who use, promote or justify the use of LLMs is not purity culture, it's a boycott. It's a political act of withdrawing my time, resources and support for something that I find deeply morally wrong. It's protest. I have a choice and I refuse.
LLMs are exploitative, destructive, biased, mediocre parroting machines. Using them has a negative impact on the climate, the arts, the quality of the internet, the job market, the economy, the accessibility of electronics, even on skill development, creativity and mental health. LLMs are made and trained on the unpaid labour of millions -if not billions- of people who didn't consent. Their generic output litter the path to finding anything by true human creators.
Wherever I can, for as long as I can, I reject LLMs and anything that is related to them. I'm boycotting.
@reading_recluse A small aid in assisting boycotting "A.I.": https://codeberg.org/just_a_husk/uBlockOrigin-AI-Blocklist
-
@dynamite_ready @lproven @reading_recluse
It is the user and their skills indeed. A hammer can be used skillfully or wrong...
@xs4me2 @dynamite_ready @reading_recluse But it can't be used for brain surgery.
No, this is not a skills issue. It is based on profound misunderstanding. No they are not good search tools. No they are not good for research or learning, because they work only and entirely by *making stuff up* and if you're learning then you're not an expert and you can't tell true from false.
-
> can be useful to digest and explore information at great speed
Nope. Still wrong. This is in fact something they are extremely and *dangerously* bad at.
@lproven@social.vivaldi.net @xs4me2@mastodon.social @reading_recluse@c.im
Hasn't been my experience. What have you tested it with?
Even tiny models in the 4-12B range have been able to handle the things I need (though granted, not as well as the 24-30B range).
My use-case is saving my hands from typing up repetitive patterns, analyzing my journals on several angles (e.g. what's my average mood based on the wording I use in my journals, how does that relate to some medical things like migraines, etc.) and as a parrot that'll repeat my plans/ calendar to me in different words, so I can overcome my own biases easier.
I have found the available models entirely sufficient for these tasks.
Not for coding, though. Even the Qwen3-Coder-Next, which is an 80B behemoth just plain sucks at code.
Now to be clear - I'm not saying they're always accurate when I use LLMs. I'm saying that because I use them with data I type up by hand and am intrinsically familiar with they save me time and mental effort, because spotting problems is easy.
I wouldn't use them for any subject which I'm not already well grounded in, and in that specific way, I agree with you.
But I also wouldn't say they're extremely or dangerously bad at digesting and exploring information, as such. Not moreso than code written by juniors without supervision.
Ultimately it's on the user to ensure the tool's output meets requirements.
Anecdotally, people aren't great at processing large amounts of information either. I work in infosec, and curate a rather complex inventory/risk/audit/reporting toolkit. I pull data from over a dozen critical systems and sub-systems, networks, etc, including vast amounts of hand-written documentation, guides and explanations about how all of this works together.
I'm still the only person capable of actually using the entire toolset in concert - not even going into further development/ integrations. Others rely on Cursor/ Claude Code to use them. And that's fine by me - I'd rather have tools that get used than tools that are entirely dependent on me.
I guess my point is that in this scenario the problem isn't LLMs themselves. The problem is people who don't take time to read and understand the requirements, input and output.
(Of course, this is putting aside the ethical/ political/ economic/ ecological problems, to keep this conversation more focused on the technical merits/demerits.) -
@lproven@social.vivaldi.net @xs4me2@mastodon.social @reading_recluse@c.im
Hasn't been my experience. What have you tested it with?
Even tiny models in the 4-12B range have been able to handle the things I need (though granted, not as well as the 24-30B range).
My use-case is saving my hands from typing up repetitive patterns, analyzing my journals on several angles (e.g. what's my average mood based on the wording I use in my journals, how does that relate to some medical things like migraines, etc.) and as a parrot that'll repeat my plans/ calendar to me in different words, so I can overcome my own biases easier.
I have found the available models entirely sufficient for these tasks.
Not for coding, though. Even the Qwen3-Coder-Next, which is an 80B behemoth just plain sucks at code.
Now to be clear - I'm not saying they're always accurate when I use LLMs. I'm saying that because I use them with data I type up by hand and am intrinsically familiar with they save me time and mental effort, because spotting problems is easy.
I wouldn't use them for any subject which I'm not already well grounded in, and in that specific way, I agree with you.
But I also wouldn't say they're extremely or dangerously bad at digesting and exploring information, as such. Not moreso than code written by juniors without supervision.
Ultimately it's on the user to ensure the tool's output meets requirements.
Anecdotally, people aren't great at processing large amounts of information either. I work in infosec, and curate a rather complex inventory/risk/audit/reporting toolkit. I pull data from over a dozen critical systems and sub-systems, networks, etc, including vast amounts of hand-written documentation, guides and explanations about how all of this works together.
I'm still the only person capable of actually using the entire toolset in concert - not even going into further development/ integrations. Others rely on Cursor/ Claude Code to use them. And that's fine by me - I'd rather have tools that get used than tools that are entirely dependent on me.
I guess my point is that in this scenario the problem isn't LLMs themselves. The problem is people who don't take time to read and understand the requirements, input and output.
(Of course, this is putting aside the ethical/ political/ economic/ ecological problems, to keep this conversation more focused on the technical merits/demerits.)@phil @xs4me2 @reading_recluse My current favourite paper on this:
https://ea.rna.nl/2024/05/27/when-chatgpt-summarises-it-actually-does-nothing-of-the-kind/
-
@xs4me2 @dynamite_ready @reading_recluse But it can't be used for brain surgery.
No, this is not a skills issue. It is based on profound misunderstanding. No they are not good search tools. No they are not good for research or learning, because they work only and entirely by *making stuff up* and if you're learning then you're not an expert and you can't tell true from false.
@lproven @dynamite_ready @reading_recluse
In my opinion, you are incorrect here, and a user is always responsible for digesting the assumed truth as they observe it. Especially on tools. There is no substitute for critical thinking. And there never will be.
Truth and social surrounds are infinitesimally more complex than analyzing a game of chess.
