Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. 👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

Planlagt Fastgjort LĂĄst Flyttet Ikke-kategoriseret
llmopensource
310 Indlæg 57 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne trĂĄd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • ossguy@fedi.copyleft.orgO ossguy@fedi.copyleft.org

    @davidgerard @wwahammy @silverwizard @firefly_lightning @cwebber Yes, which is why it's important to allow people to identify when they have used LLM/AI assistants to help. New contributors will see this is the norm, and then it will be easier to help them, because we'll know a bit about where any potential knowledge gaps might be coming from.

    If we "ban" LLM/AI-assisted contributions, people will use them anyway but hide their use, which is a trickier problem to solve.

    gulfie@mastodonapp.ukG This user is from outside of this forum
    gulfie@mastodonapp.ukG This user is from outside of this forum
    gulfie@mastodonapp.uk
    wrote sidst redigeret af
    #257

    @ossguy @davidgerard @wwahammy @silverwizard @firefly_lightning @cwebber big nope.

    1 Reply Last reply
    0
    • jedbrown@hachyderm.ioJ jedbrown@hachyderm.io

      @cwebber
      I agree about the hazard. LLM outputs should be considered derivative of all their inputs unless established otherwise. LLMs manipulate expression, not ideas, and the propensity of verbatim reproduction (up to and including entire books) is evidence of that process. Note that the purpose of the "substantial similarity" test is as circumstantial evidence of process.

      I think the counterpoints are "mutually-assured destruction" and/or "yolo denial-of-service attack on copyright will win because power likes it". "AI" companies are still delaying cases from 2022 (like Doe v GitHub) because they want a jury who believes it is inevitable. Plaintiffs seek to win their cases, not to establish broad precedent. OpenAI has already lost (in German court) on copyright infringement of their outputs, arguing unsuccessfully that the infringement is the sole responsibility of their customers for prompting. The political reality of public sentiment is changing and collapse of the financial bubble will greatly alter the power held by "AI" companies.

      Meanwhile, I think the words of the DCO ought to mean something, even for those who are certain they are a smol bean.

      https://hachyderm.io/@jedbrown/114931171543347621

      @bkuhn @ossguy @richardfontana

      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
      bkuhn@fedi.copyleft.org
      wrote sidst redigeret af
      #258

      @jedbrown

      A case from 2022 still not a trial in 2026 doesn't indicate unreasonable or manipulative delay by Defendants. Such cases really do take that long.

      Also, Doe vs. Microsoft's Github is a terribly constructed case and actually pushes us toward compulsory licensing of #FOSS works for #LLM-backed gen-#AI training— since the Plaintiff's lawyers in that case are clearly chasing their own avarice, not software freedom.

      Background:
      https://sfconservancy.org/news/2022/nov/04/class-action-lawsuit-filing-copilot/

      @cwebber @ossguy @richardfontana

      jedbrown@hachyderm.ioJ 1 Reply Last reply
      0
      • cwebber@social.coopC cwebber@social.coop

        @bkuhn @ossguy @richardfontana So let me summarize:

        - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
        - The idea of a copyleft-only LLM is a joke and we should not rely on it
        - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

        That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

        But we can't ignore this right now.

        We're putting FOSS codebases at risk.

        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
        bkuhn@fedi.copyleft.org
        wrote sidst redigeret af
        #259

        @cwebber

        Re: “polluting”, my reply is: https://fedi.copyleft.org/@bkuhn/116426437134023846 (elsewhere in thread).

        Re: “copyleft-only #LLM”: I didn't propose that. I proposed copylefting the human-modified output of LLMs.

        Re: “two scenarios”: IMO you propose a false dichotomy.

        I hope you come to one of #SFC's public sessions on this, as I'd be glad to talk more about it, & this discussion doesn't lend itself to online debate because it's so complex.

        cc: @ossguy @richardfontana
        @jedbrown

        #AI #OpenSource #FOSS

        cwebber@social.coopC 1 Reply Last reply
        0
        • mu@mastodon.nzM mu@mastodon.nz

          @bkuhn @firefly_lightning @josh @kees @ossguy ok Neville Chamberlain

          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
          bkuhn@fedi.copyleft.org
          wrote sidst redigeret af
          #260

          @mu

          A WWII reference is never helpful in a discussion unless the topic *is actually* WWII.

          I'd be glad to have a serious discussion with you, but if you follow Godwin's law again, I probably will block you.

          I know emotions are frayed and the FOSS community is frightened and worried, so I forgive you. But there is no reason to claim the situation with LLM-backed AI is tantamount to Hitler's violent invasion of Europe.

          Cc: @firefly_lightning @josh @kees @ossguy @cwebber

          mu@mastodon.nzM 1 Reply Last reply
          0
          • evan@cosocial.caE evan@cosocial.ca

            I might ask ChatGPT to give it a try, and give it some extra incentive to dig deeper because if it digs up some dirt on Claude it'd be good for business.

            @bkuhn @richardfontana @cwebber @ossguy

            bkuhn@fedi.copyleft.orgB This user is from outside of this forum
            bkuhn@fedi.copyleft.orgB This user is from outside of this forum
            bkuhn@fedi.copyleft.org
            wrote sidst redigeret af
            #261

            @evan 🤣 … but I know you're only half joking.

            Frankly part of the problem here is that people are either taking this situation *too* seriously or not serious enough. I'm guessing you're right in the happy medium, but your comment made me think of that point.

            Cc: @richardfontana @cwebber @ossguy

            1 Reply Last reply
            0
            • evan@cosocial.caE evan@cosocial.ca

              I think you could make the case that Claude is not an uninterested party in this discussion, since Blanchard used Claude to generate the code, so maybe it's lying to cover up its tracks.

              @bkuhn @richardfontana @cwebber @ossguy

              bkuhn@fedi.copyleft.orgB This user is from outside of this forum
              bkuhn@fedi.copyleft.orgB This user is from outside of this forum
              bkuhn@fedi.copyleft.org
              wrote sidst redigeret af
              #262

              @evan
              I have a speculative suspicion that the “leak” of Claude's front-end code was a false flag operation *hoping* someone would so-called “clean-room-with-Claude” their own UI.
              I have this theory b/c the UI code is not what Claude needs to IPO (it's all the server side stuff that matters), and it behooves them & their investors if they themselves take a “fair's fair” position on the leak of their own code.
              I'm meanwhile working on the chardet situation.

              Cc: @richardfontana @cwebber @ossguy

              1 Reply Last reply
              0
              • evan@cosocial.caE evan@cosocial.ca

                I gave it a try. It's quite wordy! Claude thought that a lot of Pilgrim's work would be filtered since it was a direct port from the Mozilla C++ codebase. I pushed back that they shared the same license, and it loosened up that constraint.

                https://claude.ai/share/e4aae73c-14d1-462e-9773-4381adde54f7

                Warning: if you read this document, it will get AI in you, and it will make you AI and you will become an AI-booster like me and Sam Altman. It will also burn down the rainforest.

                @bkuhn @richardfontana @cwebber @ossguy

                bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                bkuhn@fedi.copyleft.org
                wrote sidst redigeret af
                #263

                @evan

                I don't mind that you tried (and I even clicked on the link so I guess I burnt down a rainforest?), but this reads like LLM-backed gen-AI slop to me. Full of truthiness but seems to lack depth of understanding of the AFC test.

                I hope you can make it to one of SFC's chats on this topic.

                Cc: @richardfontana @cwebber @ossguy

                1 Reply Last reply
                0
                • mu@mastodon.nzM mu@mastodon.nz

                  @bkuhn @wwahammy @silverwizard @cwebber "Nothing ever got done politically that was good when both sides become more entrenched, refuse to even concede the other side has some valid points, & each say the other is the Enemy. … "

                  Now that is a really strange thing to hear from someone who is representing a FOSS community, because that's basically what FOSS *is*

                  bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                  bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                  bkuhn@fedi.copyleft.org
                  wrote sidst redigeret af
                  #264

                  @mu

                  I saw this comment after I saw you elsewhere in the thread comparing the LLM-backed genAI situation to WWII, so I am have a lot of trouble taking this seriously.

                  Plus your comment is snarky, sarcastic, mean, and slightly ad hominem. There is no reason for all that in civil debate.

                  Cc: @wwahammy @silverwizard @cwebber @richardfontana

                  mu@mastodon.nzM 1 Reply Last reply
                  0
                  • richardjacton@fosstodon.orgR richardjacton@fosstodon.org

                    @cwebber @bkuhn @ossguy @richardfontana Under this view it doesn't matter how the training data was licensed as it's a fair use defense. The outputs being uncopyrightable / effectively public domain allows people to claim they wrote it when it's convenient and they want to be able to copyright it as it's hard to prove if it was AI generated or human authored. And simultaneously to claim that it was the output of and LLM when they want to strip inconvenient licensing terms.

                    bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                    bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                    bkuhn@fedi.copyleft.org
                    wrote sidst redigeret af
                    #265

                    @RichardJActon
                    The copyleft-ish hack I propose is *we* (FOSS community) assume that any output of an LLM-backed genAI system *is* copylefted (since we are pretty sure all such systems — at least those designed for software development assist — have been trained on copylefted codebases).
                    Then, we copyleft any work that comes out of the system.
                    The only threat is proprietary software in the training set, & the industry can't abide enforcing *that*!
                    @cwebber @ossguy @richardfontana
                    @evan
                    @kees

                    richardfontana@mastodon.socialR 1 Reply Last reply
                    0
                    • cwebber@social.coopC cwebber@social.coop

                      @bkuhn @evan @richardfontana @ossguy Probably a ton of people here think I am anti-AI-output, and that I would be upset to find out that the chardet rewrite were legal.

                      Actually, I'm not! I'd be fine with the ability to copyright launder software to some degree, as long as we could do the same for proprietary software (including in binary form).

                      I'm concerned about whether or not we have an *equitable* situation, though. And I'm *more concerned* that we need to advise people, who are incorporating code *today*.

                      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                      bkuhn@fedi.copyleft.org
                      wrote sidst redigeret af
                      #266

                      @cwebber

                      We already know the situation isn't equitable & probably won't become such in our lifetimes. Microsoft already all-but-admitted they will never train Copilot on their code. No proprietary software company is going to offer training data back to other vendors.

                      The goal here obviously was to LLM-wash away copyleft. *That* we must resist, and use their own tools against them: which is the very spirit that made copyleft in the first place!

                      Cc: @evan @richardfontana @ossguy @kees
                      @karen

                      1 Reply Last reply
                      0
                      • evan@cosocial.caE evan@cosocial.ca

                        I consider myself an expert on this process since I learned about it 45 minutes ago, but it seems like AFC follows the hierarchical layers of modern programming-in-the-large -- statements, functions, modules, packages, program. That is the stuff that LLMs handle pretty well.

                        @richardfontana @cwebber @bkuhn @ossguy

                        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                        bkuhn@fedi.copyleft.org
                        wrote sidst redigeret af
                        #267

                        @evan wrote:

                        > “I consider myself an expert on this process since I learned about it 45 minutes ago ”

                        This is the second time you've made me 🤣 in this thread. Thanks for being comic relief (and I know that's not *all* you're doing, but that part is particularly helpful). Thank you!

                        Cc:
                        @richardfontana @cwebber @ossguy
                        @karen

                        evan@cosocial.caE 1 Reply Last reply
                        0
                        • sfoskett@techfieldday.netS sfoskett@techfieldday.net

                          @richardfontana @evan @cwebber @bkuhn @ossguy I feel like it’s 3 questions for the court:
                          1 Can a non-human actor produce a copyrightable work? Likely no.
                          2 Is the human prompt and review enough to apply copyright to LLM content? Maybe?
                          3 Does this have implications for open source? I guess not.

                          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                          bkuhn@fedi.copyleft.org
                          wrote sidst redigeret af
                          #268

                          @sfoskett

                          *Thaler is limited to DC Circuit & very narrow. It's a registration question, & even *its* dicta hints there is no way we can know the answer on (1).

                          I think (2) is a strong argument.

                          As for (3), there is huge value to be extracted by applying copyleft-ish principles (and copyleft licenses themselves) to LLM-backed genAI output.

                          In worse case: a big complex mix of public domain + copylefted-human-authored stuff can't easily be separated.

                          @richardfontana @evan @cwebber @ossguy

                          sfoskett@techfieldday.netS 1 Reply Last reply
                          0
                          • wwahammy@social.treehouse.systemsW wwahammy@social.treehouse.systems

                            @ossguy @cwebber @LordCaramac @bkuhn @richardfontana proprietary software companies extensively use GitHub and yet SFC's position is "don't use GitHub".

                            There are so many things we do in free software and in the interactions with SFC and FSF that would be simpler if we used proprietary software. How many janky experiences have people been asking to tolerate to participate? Why shouldn't we use proprietary software there?

                            bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                            bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                            bkuhn@fedi.copyleft.org
                            wrote sidst redigeret af
                            #269

                            @wwahammy

                            Indeed, SFC's position is #GiveUpGithub, but N.B. the https://giveupgithub.com/ site itself admits most people will uses it & suggests a “using Github under protest” README.md.

                            I use proprietary software every day. I've been convinced for ≥ 10yrs: one can't succeed in an industrialized nation at *anything* w/out sometimes doing so.

                            The difficulty is figuring out when to compromise. I remain open-minded.
                            Few of us will be FOSS monks.

                            @ossguy @cwebber @LordCaramac @richardfontana

                            1 Reply Last reply
                            0
                            • richardfontana@mastodon.socialR richardfontana@mastodon.social

                              @evan oh I mean of course you could use LLMs to help with the analysis @cwebber @bkuhn @ossguy

                              bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                              bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                              bkuhn@fedi.copyleft.org
                              wrote sidst redigeret af
                              #270

                              @richardfontana wrote:
                              > “oh I mean of course you could use LLMs to help with the analysis ”

                              I'm catching up backwards on this thread, but do you see now the monster you created by telling @evan that?

                              🤣

                              cc: @cwebber @ossguy @karen

                              evan@cosocial.caE 1 Reply Last reply
                              0
                              • evan@cosocial.caE evan@cosocial.ca

                                @richardfontana @cwebber @bkuhn @ossguy Yeah, I thought my job couldn't be automated, either, and yet here we are.

                                bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                bkuhn@fedi.copyleft.org
                                wrote sidst redigeret af
                                #271

                                LLM-backed genAI never makes as good jokes as you do, @evan

                                But are you finally coming clean with us here today that, in fact, #EvanPoll's are all created by a genAI system?

                                Cc: @richardfontana @cwebber @ossguy @karen

                                evan@cosocial.caE 1 Reply Last reply
                                0
                                • sfoskett@techfieldday.netS sfoskett@techfieldday.net

                                  @evan @cwebber @bkuhn @ossguy @richardfontana Another major concern is that works generated by AI are not copyrightable per the US Supreme Court. So code generated by an LLM can not be licensed at all, open or closed. https://www.reuters.com/legal/government/us-supreme-court-declines-hear-dispute-over-copyrights-ai-generated-material-2026-03-02/

                                  bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                  bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                  bkuhn@fedi.copyleft.org
                                  wrote sidst redigeret af
                                  #272

                                  @sfoskett
                                  I responded in detail in another post to your conclusions later, but the assumption is wrong too. It's just pure FUD to say: “works generated by AI are not copyrightable per the US Supreme Court”.
                                  https://sfconservancy.org/blog/2026/mar/04/scotus-deny-cert-dc-circuit-thaler-appeal-llm-ai/
                                  TL;DR: *DC Circuit* held that a specific copyright registration *for a digital painting* that lists a computer program as the sole author is not eligible *at this time* for copyright *registration*. SCOTUS decided to not hear the case.

                                  @evan @cwebber @richardfontana

                                  1 Reply Last reply
                                  0
                                  • richardfontana@mastodon.socialR richardfontana@mastodon.social

                                    @cwebber I think adequate compliance might be possible with good enough detection/matching tools but I don't necessarily expect such tools to be developed (let alone available to foss projects) (my assumption is that the few such tools in use today are pretty bad) @bkuhn @ossguy

                                    bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                    bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                    bkuhn@fedi.copyleft.org
                                    wrote sidst redigeret af
                                    #273

                                    @richardfontana

                                    I'm with @cwebber, there is no way to automate compliance. But, again, we should use that to our advantage in a copyleft-ish way.

                                    Cc: @ossguy

                                    1 Reply Last reply
                                    0
                                    • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                                      @evan wrote:

                                      > “I consider myself an expert on this process since I learned about it 45 minutes ago ”

                                      This is the second time you've made me 🤣 in this thread. Thanks for being comic relief (and I know that's not *all* you're doing, but that part is particularly helpful). Thank you!

                                      Cc:
                                      @richardfontana @cwebber @ossguy
                                      @karen

                                      evan@cosocial.caE This user is from outside of this forum
                                      evan@cosocial.caE This user is from outside of this forum
                                      evan@cosocial.ca
                                      wrote sidst redigeret af
                                      #274

                                      @bkuhn @richardfontana @cwebber @ossguy @karen thanks! I hope I wasn't too flip.

                                      1 Reply Last reply
                                      0
                                      • cwebber@social.coopC cwebber@social.coop

                                        @bkuhn @ossguy @richardfontana So let me summarize:

                                        - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
                                        - The idea of a copyleft-only LLM is a joke and we should not rely on it
                                        - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

                                        That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

                                        But we can't ignore this right now.

                                        We're putting FOSS codebases at risk.

                                        larsmb@mastodon.onlineL This user is from outside of this forum
                                        larsmb@mastodon.onlineL This user is from outside of this forum
                                        larsmb@mastodon.online
                                        wrote sidst redigeret af
                                        #275

                                        @cwebber @bkuhn @ossguy @richardfontana FWIW, I'd be delighted to read this as a blog post.

                                        I'm still baffled that chardet just sidestepped this via 0BSD, sort of.

                                        A thought that recently struck me that, if code is essentially impossible to license now, will we see a resurgence in other forms of IP, like ... software patents?

                                        Those *would* be defensible post-laundering ...

                                        1 Reply Last reply
                                        0
                                        • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                                          LLM-backed genAI never makes as good jokes as you do, @evan

                                          But are you finally coming clean with us here today that, in fact, #EvanPoll's are all created by a genAI system?

                                          Cc: @richardfontana @cwebber @ossguy @karen

                                          evan@cosocial.caE This user is from outside of this forum
                                          evan@cosocial.caE This user is from outside of this forum
                                          evan@cosocial.ca
                                          wrote sidst redigeret af
                                          #276

                                          @bkuhn @richardfontana @cwebber @ossguy @karen sadly no!

                                          I really don't like having anyone, including AI systems, write for me under my own name. Not least because I don't like the style and tone of ChatGPT and friends. They just write very blandly.

                                          1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper