Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. 👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

Planlagt Fastgjort LĂĄst Flyttet Ikke-kategoriseret
llmopensource
310 Indlæg 57 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne trĂĄd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • zacchiro@mastodon.xyzZ zacchiro@mastodon.xyz

    @cwebber @bkuhn @ossguy @richardfontana

    My current answer to your "is it safe" question is to answer a slightly different question. Namely: "is it any less safe than accepting code from a random employee that claims to be submitting under a inbound=outbound regime, whereas in fact they cannot?". The latter we have been doing for decades, with limited damages to the commons.

    (I *also* think the legal odds are more in our favor with AI-assisted contributions than in the previous case.)

    ? Offline
    ? Offline
    Gæst
    wrote sidst redigeret af
    #210

    @zacchiro @cwebber @bkuhn @ossguy @richardfontana I would say it's dramatically less safe. First, there's very little incentive to go after some OSS project over an unauthorized inbound=outbound contribution. Second, if someone did, the damage would likely be a small part of a single project. Third, only a small number of parties (the employer, or maybe some other single party whose code was copied) have the ability to sue.

    With LLMs, it's different. When the authors sued Anthropic, they all sued. Is a shell script that Claude generated a derivative work of, say, the romantasy novel A Court of Thorns and Roses (to pick a random thing included in Anthropic's training set)? Well, it's hard to show that it's not, in the sense that that novel is one of the zillion things that went into generating the weights that generated the shell script.

    Now it happens that the authors sued Anthropic (and settled). But I don't know if their settlement covers users of Claude (and even if it did, there are two other big models). And that's only the book authors -- there's still all of the code authors in the world.

    So yes, I think the risk is high. I mean, in some sense -- in another sense, it seems unlikely that Congress would say, "sorry, LLMs as code generators are toast because of some century-old laws". At most, they would set up a statutory licensing scheme for LLM providers which covers LLM outputs. Of course, Europe might go a different way, but I think they would probably do the same. Under this hypothetical scheme, if your code were used to train Claude, you would get a buck or two in the mail every year. Authors got I think $3k per book as a one-time payment, but that was a funny case because of how Anthropic got access to the books.

    Still, there's a risk that Congress wouldn't act (due to standard US government dysfunction).

    It seems like most people are willing to take this risk, which I think says something interesting about most people's moral intuitions.

    bkuhn@fedi.copyleft.orgB 1 Reply Last reply
    0
    • evan@cosocial.caE evan@cosocial.ca

      @cwebber

      This is probably a healthy concern.

      I think there might be some good ways to hedge one's bets, though.

      Use LLMs for rubber ducking, code scanning and review, rather than code generation.

      Keep LLM code contributions minimal and unremarkable, too.

      Don't make them load-bearing. If the code is central to the program, it's too unique.

      @richardfontana @bkuhn @ossguy

      fay@lingo.lolF This user is from outside of this forum
      fay@lingo.lolF This user is from outside of this forum
      fay@lingo.lol
      wrote sidst redigeret af
      #211

      @evan
      @cwebber @richardfontana @bkuhn @ossguy or just... not at all

      1 Reply Last reply
      0
      • evan@cosocial.caE evan@cosocial.ca

        @cwebber

        This is probably a healthy concern.

        I think there might be some good ways to hedge one's bets, though.

        Use LLMs for rubber ducking, code scanning and review, rather than code generation.

        Keep LLM code contributions minimal and unremarkable, too.

        Don't make them load-bearing. If the code is central to the program, it's too unique.

        @richardfontana @bkuhn @ossguy

        evan@cosocial.caE This user is from outside of this forum
        evan@cosocial.caE This user is from outside of this forum
        evan@cosocial.ca
        wrote sidst redigeret af
        #212

        I think the worst case scenario is that the inserted code matches exactly one snippet in the training data.

        So you could try to go for zero matches, by using such idiosyncratic and unrecommended coding conventions that nobody else has code like yours.

        Or you could try to go for lots of matches, by using bog standard coding conventions and software patterns.

        @cwebber @richardfontana @bkuhn @ossguy

        evan@cosocial.caE 1 Reply Last reply
        0
        • cwebber@social.coopC cwebber@social.coop

          @evan @richardfontana @bkuhn @ossguy Yeah! I actually already said elsewhere in the thread I don't think we need to worry about using these tools for such scenarios from a *licensing* perspective, only when the genAI is explicitly checked into the codebase

          evan@cosocial.caE This user is from outside of this forum
          evan@cosocial.caE This user is from outside of this forum
          evan@cosocial.ca
          wrote sidst redigeret af
          #213

          @cwebber the weights themselves?

          @richardfontana @bkuhn @ossguy

          cwebber@social.coopC 1 Reply Last reply
          0
          • evan@cosocial.caE evan@cosocial.ca

            @cwebber the weights themselves?

            @richardfontana @bkuhn @ossguy

            cwebber@social.coopC This user is from outside of this forum
            cwebber@social.coopC This user is from outside of this forum
            cwebber@social.coop
            wrote sidst redigeret af
            #214

            @evan @richardfontana @bkuhn @ossguy Sorry, I missed a word when I edited the sentence, I meant "genAI output"

            evan@cosocial.caE 1 Reply Last reply
            0
            • evan@cosocial.caE evan@cosocial.ca

              I think the worst case scenario is that the inserted code matches exactly one snippet in the training data.

              So you could try to go for zero matches, by using such idiosyncratic and unrecommended coding conventions that nobody else has code like yours.

              Or you could try to go for lots of matches, by using bog standard coding conventions and software patterns.

              @cwebber @richardfontana @bkuhn @ossguy

              evan@cosocial.caE This user is from outside of this forum
              evan@cosocial.caE This user is from outside of this forum
              evan@cosocial.ca
              wrote sidst redigeret af
              #215

              But maybe that's wrong; I don't know. Maybe if I wrote a Person.setName() method that was in the training set, and the LLM generated an identical Person.setName() code snippet for someone else, I could claim that the code is a copyright violation, even if there were thousands of other identical and independent Person.setName() methods in the training set.

              @cwebber @richardfontana @bkuhn @ossguy

              promovicz@chaos.socialP 1 Reply Last reply
              0
              • cwebber@social.coopC cwebber@social.coop

                @evan @richardfontana @bkuhn @ossguy Sorry, I missed a word when I edited the sentence, I meant "genAI output"

                evan@cosocial.caE This user is from outside of this forum
                evan@cosocial.caE This user is from outside of this forum
                evan@cosocial.ca
                wrote sidst redigeret af
                #216

                @cwebber it's sometimes a distinction that people blur!

                @richardfontana @bkuhn @ossguy

                1 Reply Last reply
                0
                • evan@cosocial.caE evan@cosocial.ca

                  @cwebber

                  Are you concerned that the LLMs generate nontrivial verbatim excerpts of copyrighted works?

                  Or that there is a hidden "intellectual property" in the deep patterns that they use?

                  Say, when an LLM was trained on a file I made with an interesting loop structure, and it emits code with a similar loop structure, even if the variable names, problem domain, details, or programming language differ.

                  What if a court says I can demand royalties for my "IP"?

                  @bkuhn @ossguy @richardfontana

                  sfoskett@techfieldday.netS This user is from outside of this forum
                  sfoskett@techfieldday.netS This user is from outside of this forum
                  sfoskett@techfieldday.net
                  wrote sidst redigeret af
                  #217

                  @evan @cwebber @bkuhn @ossguy @richardfontana Another major concern is that works generated by AI are not copyrightable per the US Supreme Court. So code generated by an LLM can not be licensed at all, open or closed. https://www.reuters.com/legal/government/us-supreme-court-declines-hear-dispute-over-copyrights-ai-generated-material-2026-03-02/

                  cwebber@social.coopC evan@cosocial.caE richardfontana@mastodon.socialR bkuhn@fedi.copyleft.orgB 4 Replies Last reply
                  0
                  • evan@cosocial.caE evan@cosocial.ca

                    But maybe that's wrong; I don't know. Maybe if I wrote a Person.setName() method that was in the training set, and the LLM generated an identical Person.setName() code snippet for someone else, I could claim that the code is a copyright violation, even if there were thousands of other identical and independent Person.setName() methods in the training set.

                    @cwebber @richardfontana @bkuhn @ossguy

                    promovicz@chaos.socialP This user is from outside of this forum
                    promovicz@chaos.socialP This user is from outside of this forum
                    promovicz@chaos.social
                    wrote sidst redigeret af
                    #218

                    @evan That’s not enough code for copyright enforcement. People have been finding identical code in the output - you just need something “rare”. It’s similar for subjects with little text in the corpus - I’ve been seeing listings that *can only have one source* (retro datasheets by AMD, in my case).

                    1 Reply Last reply
                    0
                    • cwebber@social.coopC cwebber@social.coop

                      @richardfontana @bkuhn @ossguy That's a problem so hard it throws the "NP complete" debate out the window in favor of something brand new. Given that these codebases have no trouble "translating" from one language's source code into another, how on *earth* could you possibly hope to build a compliance tool around that?

                      Laughable, to anyone who tries.

                      evan@cosocial.caE This user is from outside of this forum
                      evan@cosocial.caE This user is from outside of this forum
                      evan@cosocial.ca
                      wrote sidst redigeret af
                      #219

                      @cwebber

                      This is a really interesting question! TIL about CA vs. Altai and the abstraction-filtration-comparison test.

                      https://w.wiki/LbiW

                      I'm not sure how automatable it is. Interesting to try though!

                      @richardfontana @bkuhn @ossguy

                      richardfontana@mastodon.socialR 1 Reply Last reply
                      0
                      • sfoskett@techfieldday.netS sfoskett@techfieldday.net

                        @evan @cwebber @bkuhn @ossguy @richardfontana Another major concern is that works generated by AI are not copyrightable per the US Supreme Court. So code generated by an LLM can not be licensed at all, open or closed. https://www.reuters.com/legal/government/us-supreme-court-declines-hear-dispute-over-copyrights-ai-generated-material-2026-03-02/

                        cwebber@social.coopC This user is from outside of this forum
                        cwebber@social.coopC This user is from outside of this forum
                        cwebber@social.coop
                        wrote sidst redigeret af
                        #220

                        @sfoskett @evan @bkuhn @ossguy @richardfontana That outcome I am not worried about; code that's not copyrightable is considered in the public domain within the US, which means there aren't any real risks to incorporating into FOSS projects. But the Supreme Court punted on it, they didn't rule that way.

                        1 Reply Last reply
                        0
                        • sfoskett@techfieldday.netS sfoskett@techfieldday.net

                          @evan @cwebber @bkuhn @ossguy @richardfontana Another major concern is that works generated by AI are not copyrightable per the US Supreme Court. So code generated by an LLM can not be licensed at all, open or closed. https://www.reuters.com/legal/government/us-supreme-court-declines-hear-dispute-over-copyrights-ai-generated-material-2026-03-02/

                          evan@cosocial.caE This user is from outside of this forum
                          evan@cosocial.caE This user is from outside of this forum
                          evan@cosocial.ca
                          wrote sidst redigeret af
                          #221

                          @sfoskett you can incorporate public domain code into a licensed work.

                          @cwebber @bkuhn @ossguy @richardfontana

                          sfoskett@techfieldday.netS 1 Reply Last reply
                          0
                          • evan@cosocial.caE evan@cosocial.ca

                            @cwebber

                            This is a really interesting question! TIL about CA vs. Altai and the abstraction-filtration-comparison test.

                            https://w.wiki/LbiW

                            I'm not sure how automatable it is. Interesting to try though!

                            @richardfontana @bkuhn @ossguy

                            richardfontana@mastodon.socialR This user is from outside of this forum
                            richardfontana@mastodon.socialR This user is from outside of this forum
                            richardfontana@mastodon.social
                            wrote sidst redigeret af
                            #222

                            @evan I feel pretty confident in saying the abstraction-filtration-comparison test cannot possibly be automated @cwebber @bkuhn @ossguy

                            evan@cosocial.caE 1 Reply Last reply
                            0
                            • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                              👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993. I was on USENET extensively then; I confirm the disruption was indeed similar. I urge you to read his essay, think about it, & join Denver, me, & others at the following datetimes…
                              $ date -d '2026-04-21 15:00 UTC'
                              $ date -d '2026-04-28 23:00 UTC'
                              …in https://bbb-new.sfconservancy.org/rooms/welcome-llm-gen-ai-users-to-foss/join
                              #AI #LLM #OpenSource

                              redchrision@mastodon.socialR This user is from outside of this forum
                              redchrision@mastodon.socialR This user is from outside of this forum
                              redchrision@mastodon.social
                              wrote sidst redigeret af
                              #223

                              Sorry for interfering in the discussion out of the blue, but the topic is really interesting. I really hope that the conclusion of this will not be engineers saying they are not lawyers, and lawyers saying that it's for the courts to decide.

                              redchrision@mastodon.socialR 1 Reply Last reply
                              0
                              • redchrision@mastodon.socialR redchrision@mastodon.social

                                Sorry for interfering in the discussion out of the blue, but the topic is really interesting. I really hope that the conclusion of this will not be engineers saying they are not lawyers, and lawyers saying that it's for the courts to decide.

                                redchrision@mastodon.socialR This user is from outside of this forum
                                redchrision@mastodon.socialR This user is from outside of this forum
                                redchrision@mastodon.social
                                wrote sidst redigeret af
                                #224

                                To chip in, a situation where AI-generated code is completely unacceptable is hard, if not impossible, to implement. This also puts a lot of pressure on reviewers, who have the difficult job of determining whether a piece of code is AI-generated. Sometimes it's easy; sometimes it's impossible under the conditions they operate in. If the code is good, it should be accepted.

                                redchrision@mastodon.socialR 1 Reply Last reply
                                0
                                • redchrision@mastodon.socialR redchrision@mastodon.social

                                  To chip in, a situation where AI-generated code is completely unacceptable is hard, if not impossible, to implement. This also puts a lot of pressure on reviewers, who have the difficult job of determining whether a piece of code is AI-generated. Sometimes it's easy; sometimes it's impossible under the conditions they operate in. If the code is good, it should be accepted.

                                  redchrision@mastodon.socialR This user is from outside of this forum
                                  redchrision@mastodon.socialR This user is from outside of this forum
                                  redchrision@mastodon.social
                                  wrote sidst redigeret af
                                  #225

                                  The practical issue is the fact that reviewers risk facing large amounts of PRs done completely with LLMs, and even by LLMs, under the name of a human who uses them. That generates enormous risk for code quality. Probably a practical way of handling this is to accept PRs only from community members who are validated with a status of “valid contributor”, making it more official, basically.

                                  redchrision@mastodon.socialR 1 Reply Last reply
                                  0
                                  • redchrision@mastodon.socialR redchrision@mastodon.social

                                    The practical issue is the fact that reviewers risk facing large amounts of PRs done completely with LLMs, and even by LLMs, under the name of a human who uses them. That generates enormous risk for code quality. Probably a practical way of handling this is to accept PRs only from community members who are validated with a status of “valid contributor”, making it more official, basically.

                                    redchrision@mastodon.socialR This user is from outside of this forum
                                    redchrision@mastodon.socialR This user is from outside of this forum
                                    redchrision@mastodon.social
                                    wrote sidst redigeret af
                                    #226

                                    This would mean putting responsibility on those who are contributing: if they make PRs, those should be carefully analyzed by them beforehand, even if they use AI, since it's hard to control that.

                                    redchrision@mastodon.socialR 1 Reply Last reply
                                    0
                                    • redchrision@mastodon.socialR redchrision@mastodon.social

                                      This would mean putting responsibility on those who are contributing: if they make PRs, those should be carefully analyzed by them beforehand, even if they use AI, since it's hard to control that.

                                      redchrision@mastodon.socialR This user is from outside of this forum
                                      redchrision@mastodon.socialR This user is from outside of this forum
                                      redchrision@mastodon.social
                                      wrote sidst redigeret af
                                      #227

                                      This will avoid a situation where a reviewer ends up reviewing what an LLM has written and then becomes, in some sense, an author without their consent of whatever the LLM outputs next, through comments on the PR and suggestions for improvement that will turn into future prompts.

                                      redchrision@mastodon.socialR 1 Reply Last reply
                                      0
                                      • redchrision@mastodon.socialR redchrision@mastodon.social

                                        This will avoid a situation where a reviewer ends up reviewing what an LLM has written and then becomes, in some sense, an author without their consent of whatever the LLM outputs next, through comments on the PR and suggestions for improvement that will turn into future prompts.

                                        redchrision@mastodon.socialR This user is from outside of this forum
                                        redchrision@mastodon.socialR This user is from outside of this forum
                                        redchrision@mastodon.social
                                        wrote sidst redigeret af
                                        #228

                                        The sanction for not respecting that should be related to reputation within that community and decided locally: whether they further allow that person to contribute, completely ban contributions, close PRs from the beginning, etc.

                                        redchrision@mastodon.socialR 1 Reply Last reply
                                        0
                                        • redchrision@mastodon.socialR redchrision@mastodon.social

                                          The sanction for not respecting that should be related to reputation within that community and decided locally: whether they further allow that person to contribute, completely ban contributions, close PRs from the beginning, etc.

                                          redchrision@mastodon.socialR This user is from outside of this forum
                                          redchrision@mastodon.socialR This user is from outside of this forum
                                          redchrision@mastodon.social
                                          wrote sidst redigeret af
                                          #229

                                          On the topic of proprietary code generated by LLMs and then accepted in OSS, the responsibility should be on the LLM company; the code should naturally inherit the OS license it is associated with. On the topic of LLM companies using OSS code inappropriately, the responsibility should again be on the LLM company. In both situations, courts will probably have opinions in the future, and LLM companies might consider adapting their use policies further.

                                          redchrision@mastodon.socialR 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper