Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. 👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

Planlagt Fastgjort LĂĄst Flyttet Ikke-kategoriseret
llmopensource
310 Indlæg 57 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne trĂĄd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • evan@cosocial.caE evan@cosocial.ca

    The comparison seems tough, but I'd put an LLM to the task. "How similar are the database abstraction layers in activitypub-bot and Fedify?" Again, I'd probably want some human review, but for that code stuff LLMs are pretty good.

    @richardfontana @cwebber @bkuhn @ossguy

    evan@cosocial.caE This user is from outside of this forum
    evan@cosocial.caE This user is from outside of this forum
    evan@cosocial.ca
    wrote sidst redigeret af
    #241

    I consider myself an expert on this process since I learned about it 45 minutes ago, but it seems like AFC follows the hierarchical layers of modern programming-in-the-large -- statements, functions, modules, packages, program. That is the stuff that LLMs handle pretty well.

    @richardfontana @cwebber @bkuhn @ossguy

    bkuhn@fedi.copyleft.orgB 1 Reply Last reply
    0
    • evan@cosocial.caE evan@cosocial.ca

      For filtration, it seems like merger or scènes à faire would also be kind of automatable, maybe with human oversight. Is there a way to make a mailing daemon without a logging module? Maybe, but it's so common that everyone does it that way. Could you have a Person class without a getter and setter for the name? Probably not?

      @richardfontana @cwebber @bkuhn @ossguy

      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
      bkuhn@fedi.copyleft.org
      wrote sidst redigeret af
      #242

      @evan

      I actually think that these copyright concepts aren't particularly automatable.
      Even if we try, it's pure arms race.

      And the merger doctrine isn't the big problem here, it is the more complex analysis where merger doctrine clearly doesn't apply that needs analysis and I suspect the analysis is difficult to (even partially) automate.

      But I'm looking into it.

      Cf: chardet situation https://github.com/chardet/chardet/issues/355#issuecomment-4145369025

      @richardfontana @cwebber @ossguy

      cwebber@social.coopC evan@cosocial.caE 2 Replies Last reply
      0
      • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

        @novalis
        I agree with your supporting arguments but not the conclusion.

        It goes back to the mutually assured destruction idea: no one in the for profit proprietary software industry is going to bring a lawsuit because they are so invested in LLM-backed AI succeeding.

        That's where our commons differs widely from other creative works of expression.

        I am worried about compulsory licensing for *training* —could be a disaster — but unrelated to output.

        @zacchiro @cwebber @ossguy @richardfontana

        ? Offline
        ? Offline
        Gæst
        wrote sidst redigeret af
        #243

        @bkuhn @zacchiro @cwebber @ossguy @richardfontana I don't even know if I agree with my supporting arguments. But I don't even think that it has to be someone in the proprietary world that brings a lawsuit -- it could be anyone whose code or text is trained on.

        1 Reply Last reply
        0
        • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

          @evan

          I actually think that these copyright concepts aren't particularly automatable.
          Even if we try, it's pure arms race.

          And the merger doctrine isn't the big problem here, it is the more complex analysis where merger doctrine clearly doesn't apply that needs analysis and I suspect the analysis is difficult to (even partially) automate.

          But I'm looking into it.

          Cf: chardet situation https://github.com/chardet/chardet/issues/355#issuecomment-4145369025

          @richardfontana @cwebber @ossguy

          cwebber@social.coopC This user is from outside of this forum
          cwebber@social.coopC This user is from outside of this forum
          cwebber@social.coop
          wrote sidst redigeret af
          #244

          @bkuhn @evan @richardfontana @ossguy One thing I worry about is that the chardet rewrite might not generalize. The chardet maintainer used *more* care in the rewrite than most projects which have followed suit for laundering would. https://dan-blanchard.github.io/blog/chardet-rewrite-controversy/

          Even then, it raises questions, because even the maintainer admits, chardet was part of the training set.

          It's very similar to how a friend recently sent me, "Claude managed to reverse engineer Bubble Bobble without using any reverse engineering tools, just inspecting the binary!" https://kotrotsos.medium.com/we-pointed-an-ai-at-raw-binary-files-from-1986-662ba30120f3

          Which like, Claude is enough of a black box already but Bubble Bobble is also one of the most studied ROMs in history, so that's hard to evaluate whether it's true. You'd have to choose a less studied ROM as a test case, not Bubble Bobble, which the internet has discussed to death.

          cwebber@social.coopC 1 Reply Last reply
          0
          • cwebber@social.coopC cwebber@social.coop

            @bkuhn @evan @richardfontana @ossguy One thing I worry about is that the chardet rewrite might not generalize. The chardet maintainer used *more* care in the rewrite than most projects which have followed suit for laundering would. https://dan-blanchard.github.io/blog/chardet-rewrite-controversy/

            Even then, it raises questions, because even the maintainer admits, chardet was part of the training set.

            It's very similar to how a friend recently sent me, "Claude managed to reverse engineer Bubble Bobble without using any reverse engineering tools, just inspecting the binary!" https://kotrotsos.medium.com/we-pointed-an-ai-at-raw-binary-files-from-1986-662ba30120f3

            Which like, Claude is enough of a black box already but Bubble Bobble is also one of the most studied ROMs in history, so that's hard to evaluate whether it's true. You'd have to choose a less studied ROM as a test case, not Bubble Bobble, which the internet has discussed to death.

            cwebber@social.coopC This user is from outside of this forum
            cwebber@social.coopC This user is from outside of this forum
            cwebber@social.coop
            wrote sidst redigeret af
            #245

            @bkuhn @evan @richardfontana @ossguy Probably a ton of people here think I am anti-AI-output, and that I would be upset to find out that the chardet rewrite were legal.

            Actually, I'm not! I'd be fine with the ability to copyright launder software to some degree, as long as we could do the same for proprietary software (including in binary form).

            I'm concerned about whether or not we have an *equitable* situation, though. And I'm *more concerned* that we need to advise people, who are incorporating code *today*.

            bkuhn@fedi.copyleft.orgB 1 Reply Last reply
            0
            • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

              @evan

              I actually think that these copyright concepts aren't particularly automatable.
              Even if we try, it's pure arms race.

              And the merger doctrine isn't the big problem here, it is the more complex analysis where merger doctrine clearly doesn't apply that needs analysis and I suspect the analysis is difficult to (even partially) automate.

              But I'm looking into it.

              Cf: chardet situation https://github.com/chardet/chardet/issues/355#issuecomment-4145369025

              @richardfontana @cwebber @ossguy

              evan@cosocial.caE This user is from outside of this forum
              evan@cosocial.caE This user is from outside of this forum
              evan@cosocial.ca
              wrote sidst redigeret af
              #246

              @bkuhn I just did an abstraction and filtration pass on a medium-sized application framework (~30K LOC), and as an expert on the code I think it did a good job:

              https://claude.ai/share/071ccb69-5d22-4673-905a-362d9663e7d0

              It missed a few things (e.g. relay specs). Then again, I have no idea how this kind of review is supposed to work. I didn't go down to the function or statement level -- that'd probably be much noisier.

              Maybe chardet 2 and 7 would be a better test of the technique?

              @richardfontana @cwebber @ossguy

              evan@cosocial.caE 2 Replies Last reply
              0
              • cwebber@social.coopC cwebber@social.coop

                @bkuhn @ossguy @richardfontana I say "good outcome", and I'm not saying it's an outcome I want, because "what I want" is pretty complicated here. I'm saying, it's the only one where there is the possibility of legal output from these tools that can safely be incorporated into FOSS projects *at all* that is *equitable* for both FOSS and proprietary situations.

                And yup, unfortunately, that would mean copyright-laundering of FOSS codebases through LLMs would be possible to strip copyleft.

                It would also mean the same for proprietary codebases.

                Frankly I think it would kind of rule if we stabbed copyright in the gut that badly, but there's so much vested interest from various copyright holding corporations, I don't think we're likely to get that. Do you?

                richardjacton@fosstodon.orgR This user is from outside of this forum
                richardjacton@fosstodon.orgR This user is from outside of this forum
                richardjacton@fosstodon.org
                wrote sidst redigeret af
                #247

                @cwebber @bkuhn @ossguy @richardfontana I'd don't see a great way out of the copyright stripping conclusions for them without changes to the law. As I understand their defense for training on copyrighted materials - it's predicated on the models being a "transformative" and not competing directly with the original works in the market. The models themselves don't compete with the training material only their outputs do - and the LLM companies want any liability for that to be on users not them.

                richardjacton@fosstodon.orgR 1 Reply Last reply
                0
                • richardjacton@fosstodon.orgR richardjacton@fosstodon.org

                  @cwebber @bkuhn @ossguy @richardfontana I'd don't see a great way out of the copyright stripping conclusions for them without changes to the law. As I understand their defense for training on copyrighted materials - it's predicated on the models being a "transformative" and not competing directly with the original works in the market. The models themselves don't compete with the training material only their outputs do - and the LLM companies want any liability for that to be on users not them.

                  richardjacton@fosstodon.orgR This user is from outside of this forum
                  richardjacton@fosstodon.orgR This user is from outside of this forum
                  richardjacton@fosstodon.org
                  wrote sidst redigeret af
                  #248

                  @cwebber @bkuhn @ossguy @richardfontana Under this view it doesn't matter how the training data was licensed as it's a fair use defense. The outputs being uncopyrightable / effectively public domain allows people to claim they wrote it when it's convenient and they want to be able to copyright it as it's hard to prove if it was AI generated or human authored. And simultaneously to claim that it was the output of and LLM when they want to strip inconvenient licensing terms.

                  bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                  0
                  • evan@cosocial.caE evan@cosocial.ca

                    @bkuhn I just did an abstraction and filtration pass on a medium-sized application framework (~30K LOC), and as an expert on the code I think it did a good job:

                    https://claude.ai/share/071ccb69-5d22-4673-905a-362d9663e7d0

                    It missed a few things (e.g. relay specs). Then again, I have no idea how this kind of review is supposed to work. I didn't go down to the function or statement level -- that'd probably be much noisier.

                    Maybe chardet 2 and 7 would be a better test of the technique?

                    @richardfontana @cwebber @ossguy

                    evan@cosocial.caE This user is from outside of this forum
                    evan@cosocial.caE This user is from outside of this forum
                    evan@cosocial.ca
                    wrote sidst redigeret af
                    #249

                    If I were going to productize this, I'd do AF passes on a huge training dataset like The Stack and generate some kind of fingerprint for each program. (Estimated cost: billions!)

                    https://huggingface.co/datasets/bigcode/the-stack

                    Then, I'd have a tool to let you fingerprint your own code and C it against the big database -- maybe give you a list of high-similarity codebases.

                    And you could re-run the comparison each time you push to Git -- maybe only Cing what changed.

                    @bkuhn @richardfontana @cwebber @ossguy

                    1 Reply Last reply
                    0
                    • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                      (2/5) … In https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ ,
                      Denver's key points are: we *have* to (a) be open to *listening* to people who want to contribute #FOSS with #LLM-backed generative #AI systems, & (b) work collaboratively on a *plan* of how we can solve the current crisis.

                      Nothing ever got done politically that was good when both sides become more entrenched, refuse to even concede the other side has some valid points, & each say the other is the Enemy. …

                      Cc: @wwahammy @silverwizard @cwebber

                      #OpenSource

                      mu@mastodon.nzM This user is from outside of this forum
                      mu@mastodon.nzM This user is from outside of this forum
                      mu@mastodon.nz
                      wrote sidst redigeret af
                      #250

                      @bkuhn @wwahammy @silverwizard @cwebber "Nothing ever got done politically that was good when both sides become more entrenched, refuse to even concede the other side has some valid points, & each say the other is the Enemy. … "

                      Now that is a really strange thing to hear from someone who is representing a FOSS community, because that's basically what FOSS *is*

                      bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                      0
                      • evan@cosocial.caE evan@cosocial.ca

                        @bkuhn I just did an abstraction and filtration pass on a medium-sized application framework (~30K LOC), and as an expert on the code I think it did a good job:

                        https://claude.ai/share/071ccb69-5d22-4673-905a-362d9663e7d0

                        It missed a few things (e.g. relay specs). Then again, I have no idea how this kind of review is supposed to work. I didn't go down to the function or statement level -- that'd probably be much noisier.

                        Maybe chardet 2 and 7 would be a better test of the technique?

                        @richardfontana @cwebber @ossguy

                        evan@cosocial.caE This user is from outside of this forum
                        evan@cosocial.caE This user is from outside of this forum
                        evan@cosocial.ca
                        wrote sidst redigeret af
                        #251

                        I gave it a try. It's quite wordy! Claude thought that a lot of Pilgrim's work would be filtered since it was a direct port from the Mozilla C++ codebase. I pushed back that they shared the same license, and it loosened up that constraint.

                        https://claude.ai/share/e4aae73c-14d1-462e-9773-4381adde54f7

                        Warning: if you read this document, it will get AI in you, and it will make you AI and you will become an AI-booster like me and Sam Altman. It will also burn down the rainforest.

                        @bkuhn @richardfontana @cwebber @ossguy

                        evan@cosocial.caE bkuhn@fedi.copyleft.orgB 2 Replies Last reply
                        0
                        • evan@cosocial.caE evan@cosocial.ca

                          I gave it a try. It's quite wordy! Claude thought that a lot of Pilgrim's work would be filtered since it was a direct port from the Mozilla C++ codebase. I pushed back that they shared the same license, and it loosened up that constraint.

                          https://claude.ai/share/e4aae73c-14d1-462e-9773-4381adde54f7

                          Warning: if you read this document, it will get AI in you, and it will make you AI and you will become an AI-booster like me and Sam Altman. It will also burn down the rainforest.

                          @bkuhn @richardfontana @cwebber @ossguy

                          evan@cosocial.caE This user is from outside of this forum
                          evan@cosocial.caE This user is from outside of this forum
                          evan@cosocial.ca
                          wrote sidst redigeret af
                          #252

                          I think you could make the case that Claude is not an uninterested party in this discussion, since Blanchard used Claude to generate the code, so maybe it's lying to cover up its tracks.

                          @bkuhn @richardfontana @cwebber @ossguy

                          evan@cosocial.caE bkuhn@fedi.copyleft.orgB 2 Replies Last reply
                          0
                          • evan@cosocial.caE evan@cosocial.ca

                            I think you could make the case that Claude is not an uninterested party in this discussion, since Blanchard used Claude to generate the code, so maybe it's lying to cover up its tracks.

                            @bkuhn @richardfontana @cwebber @ossguy

                            evan@cosocial.caE This user is from outside of this forum
                            evan@cosocial.caE This user is from outside of this forum
                            evan@cosocial.ca
                            wrote sidst redigeret af
                            #253

                            I might ask ChatGPT to give it a try, and give it some extra incentive to dig deeper because if it digs up some dirt on Claude it'd be good for business.

                            @bkuhn @richardfontana @cwebber @ossguy

                            bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                            0
                            • wwahammy@social.treehouse.systemsW wwahammy@social.treehouse.systems

                              @bkuhn @kees @glitzersachen @josh @silverwizard @ossguy @xgranade

                              This is not a remotely accurate analogy. The level of rage in this country over AI is uncontrollable and it's accelerating. Two people tried to kill Sam Altman in the last week. An Indiana planning official's house was shot after they approved a new data center.

                              In the political realm, the shift is unimaginably swift. Ex: 6 months ago, no Democrat for WI governor had a policy on data centers because building unions wanted them. Now every one of them is fighting over how strict their ban on data centers is.

                              The best analogy I think of is the opioid crisis. When people were ready to kill the Sacklers and everyone at Purdue Pharma, you can't come in and say anything that people think you are tolerant of the damage. You can't even argue "we can punish these people but we have to protect access to opioids". Everyone KNOWS there are uses but you can't build a policy around that because the public doesn't care. At all.

                              The only time you can have this discussion was years ago or years in the future after the public has taken their pound of flesh. Right now, it's an immensely dangerous idea for SFC.

                              mu@mastodon.nzM This user is from outside of this forum
                              mu@mastodon.nzM This user is from outside of this forum
                              mu@mastodon.nz
                              wrote sidst redigeret af
                              #254

                              @wwahammy @bkuhn @kees @glitzersachen @josh @silverwizard @ossguy @xgranade the people wanting to kill Sam Altman are doing so because they are afraid of the AI Doomer stories, this discussion about including slop in software is very different.

                              1 Reply Last reply
                              0
                              • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                                @firefly_lightning
                                You're not overstepping, and these are very good perspectives. I hope you'll come to the real-time discussion sessions and talk about this.
                                I am concerned that maintainers are already overwhelmed with #AI #slop right now but yelling at the problem has not helped.

                                We're close to an arms race here & I'd rather be the voice of reason to find a compromise that advances FOSS & doesn't complicate maintainer's jobs rather than take a side in the arms race.
                                Cc: @josh @kees @ossguy

                                mu@mastodon.nzM This user is from outside of this forum
                                mu@mastodon.nzM This user is from outside of this forum
                                mu@mastodon.nz
                                wrote sidst redigeret af
                                #255

                                @bkuhn @firefly_lightning @josh @kees @ossguy ok Neville Chamberlain

                                bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                                0
                                • cwebber@social.coopC cwebber@social.coop

                                  @bkuhn @ossguy @richardfontana So let me summarize:

                                  - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
                                  - The idea of a copyleft-only LLM is a joke and we should not rely on it
                                  - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

                                  That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

                                  But we can't ignore this right now.

                                  We're putting FOSS codebases at risk.

                                  jedbrown@hachyderm.ioJ This user is from outside of this forum
                                  jedbrown@hachyderm.ioJ This user is from outside of this forum
                                  jedbrown@hachyderm.io
                                  wrote sidst redigeret af
                                  #256

                                  @cwebber
                                  I agree about the hazard. LLM outputs should be considered derivative of all their inputs unless established otherwise. LLMs manipulate expression, not ideas, and the propensity of verbatim reproduction (up to and including entire books) is evidence of that process. Note that the purpose of the "substantial similarity" test is as circumstantial evidence of process.

                                  I think the counterpoints are "mutually-assured destruction" and/or "yolo denial-of-service attack on copyright will win because power likes it". "AI" companies are still delaying cases from 2022 (like Doe v GitHub) because they want a jury who believes it is inevitable. Plaintiffs seek to win their cases, not to establish broad precedent. OpenAI has already lost (in German court) on copyright infringement of their outputs, arguing unsuccessfully that the infringement is the sole responsibility of their customers for prompting. The political reality of public sentiment is changing and collapse of the financial bubble will greatly alter the power held by "AI" companies.

                                  Meanwhile, I think the words of the DCO ought to mean something, even for those who are certain they are a smol bean.

                                  https://hachyderm.io/@jedbrown/114931171543347621

                                  @bkuhn @ossguy @richardfontana

                                  bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                                  0
                                  • ossguy@fedi.copyleft.orgO ossguy@fedi.copyleft.org

                                    @davidgerard @wwahammy @silverwizard @firefly_lightning @cwebber Yes, which is why it's important to allow people to identify when they have used LLM/AI assistants to help. New contributors will see this is the norm, and then it will be easier to help them, because we'll know a bit about where any potential knowledge gaps might be coming from.

                                    If we "ban" LLM/AI-assisted contributions, people will use them anyway but hide their use, which is a trickier problem to solve.

                                    gulfie@mastodonapp.ukG This user is from outside of this forum
                                    gulfie@mastodonapp.ukG This user is from outside of this forum
                                    gulfie@mastodonapp.uk
                                    wrote sidst redigeret af
                                    #257

                                    @ossguy @davidgerard @wwahammy @silverwizard @firefly_lightning @cwebber big nope.

                                    1 Reply Last reply
                                    0
                                    • jedbrown@hachyderm.ioJ jedbrown@hachyderm.io

                                      @cwebber
                                      I agree about the hazard. LLM outputs should be considered derivative of all their inputs unless established otherwise. LLMs manipulate expression, not ideas, and the propensity of verbatim reproduction (up to and including entire books) is evidence of that process. Note that the purpose of the "substantial similarity" test is as circumstantial evidence of process.

                                      I think the counterpoints are "mutually-assured destruction" and/or "yolo denial-of-service attack on copyright will win because power likes it". "AI" companies are still delaying cases from 2022 (like Doe v GitHub) because they want a jury who believes it is inevitable. Plaintiffs seek to win their cases, not to establish broad precedent. OpenAI has already lost (in German court) on copyright infringement of their outputs, arguing unsuccessfully that the infringement is the sole responsibility of their customers for prompting. The political reality of public sentiment is changing and collapse of the financial bubble will greatly alter the power held by "AI" companies.

                                      Meanwhile, I think the words of the DCO ought to mean something, even for those who are certain they are a smol bean.

                                      https://hachyderm.io/@jedbrown/114931171543347621

                                      @bkuhn @ossguy @richardfontana

                                      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                      bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                      bkuhn@fedi.copyleft.org
                                      wrote sidst redigeret af
                                      #258

                                      @jedbrown

                                      A case from 2022 still not a trial in 2026 doesn't indicate unreasonable or manipulative delay by Defendants. Such cases really do take that long.

                                      Also, Doe vs. Microsoft's Github is a terribly constructed case and actually pushes us toward compulsory licensing of #FOSS works for #LLM-backed gen-#AI training— since the Plaintiff's lawyers in that case are clearly chasing their own avarice, not software freedom.

                                      Background:
                                      https://sfconservancy.org/news/2022/nov/04/class-action-lawsuit-filing-copilot/

                                      @cwebber @ossguy @richardfontana

                                      jedbrown@hachyderm.ioJ 1 Reply Last reply
                                      0
                                      • cwebber@social.coopC cwebber@social.coop

                                        @bkuhn @ossguy @richardfontana So let me summarize:

                                        - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
                                        - The idea of a copyleft-only LLM is a joke and we should not rely on it
                                        - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

                                        That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

                                        But we can't ignore this right now.

                                        We're putting FOSS codebases at risk.

                                        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                        bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                        bkuhn@fedi.copyleft.org
                                        wrote sidst redigeret af
                                        #259

                                        @cwebber

                                        Re: “polluting”, my reply is: https://fedi.copyleft.org/@bkuhn/116426437134023846 (elsewhere in thread).

                                        Re: “copyleft-only #LLM”: I didn't propose that. I proposed copylefting the human-modified output of LLMs.

                                        Re: “two scenarios”: IMO you propose a false dichotomy.

                                        I hope you come to one of #SFC's public sessions on this, as I'd be glad to talk more about it, & this discussion doesn't lend itself to online debate because it's so complex.

                                        cc: @ossguy @richardfontana
                                        @jedbrown

                                        #AI #OpenSource #FOSS

                                        cwebber@social.coopC 1 Reply Last reply
                                        0
                                        • mu@mastodon.nzM mu@mastodon.nz

                                          @bkuhn @firefly_lightning @josh @kees @ossguy ok Neville Chamberlain

                                          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                          bkuhn@fedi.copyleft.orgB This user is from outside of this forum
                                          bkuhn@fedi.copyleft.org
                                          wrote sidst redigeret af
                                          #260

                                          @mu

                                          A WWII reference is never helpful in a discussion unless the topic *is actually* WWII.

                                          I'd be glad to have a serious discussion with you, but if you follow Godwin's law again, I probably will block you.

                                          I know emotions are frayed and the FOSS community is frightened and worried, so I forgive you. But there is no reason to claim the situation with LLM-backed AI is tantamount to Hitler's violent invasion of Europe.

                                          Cc: @firefly_lightning @josh @kees @ossguy @cwebber

                                          mu@mastodon.nzM 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper