Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. 👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

👀 … https://sfconservancy.org/blog/2026/apr/15/eternal-november-generative-ai-llm/ …my colleague Denver Gingerich writes: newcomers' extensive reliance on LLM-backed generative AI is comparable to the Eternal September onslaught to USENET in 1993.

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
llmopensource
310 Indlæg 57 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • miss_rodent@girlcock.clubM miss_rodent@girlcock.club

    @bkuhn parts being as free as the source code, and including their outputs should be treated as a binary blob - since there is no way to investigate the process behind how it ended up there or was created. It can't even be meaningfully reverse-engineered in most cases.

    I'd also add to any project refusal of any generated content that carries the climate, freedom, rights, labour rights, etc. concerns as well - but at a minimum, the outputs of an LLM can not be considered free. (5/5)

    miss_rodent@girlcock.clubM This user is from outside of this forum
    miss_rodent@girlcock.clubM This user is from outside of this forum
    miss_rodent@girlcock.club
    wrote sidst redigeret af
    #144

    @bkuhn The *people* coming in, if they mean well and will contribute in ethical ways, are fine, and worth welcoming in.
    But the LLMs themselves - the use of the tool - as it exists undermines the ethical core of the free software movement, and carries too many other ethical problems to be acceptable.
    (6/5)

    miss_rodent@girlcock.clubM 1 Reply Last reply
    0
    • miss_rodent@girlcock.clubM miss_rodent@girlcock.club

      @bkuhn The *people* coming in, if they mean well and will contribute in ethical ways, are fine, and worth welcoming in.
      But the LLMs themselves - the use of the tool - as it exists undermines the ethical core of the free software movement, and carries too many other ethical problems to be acceptable.
      (6/5)

      miss_rodent@girlcock.clubM This user is from outside of this forum
      miss_rodent@girlcock.clubM This user is from outside of this forum
      miss_rodent@girlcock.club
      wrote sidst redigeret af
      #145

      @bkuhn This leaves room for an ethical, actually free, version of the tech, should it appear at some point, which is a compromise, my instinct is that there can be no ethical version, but, I could be wrong.
      As-is though, the LLMs & Genrative systems in use are a black box - even 'open source' ones, if they do not also provide full access to the training data & methodology, including them in free software is no better than proprietary code. The definitions & licenses need to reflect this. (7/5)

      1 Reply Last reply
      0
      • kees@hachyderm.ioK kees@hachyderm.io

        @glitzersachen @josh @silverwizard @ossguy @bkuhn @karen @wwahammy @xgranade

        Right, yeah, this is why I've cautioned people about *how* they use LLMs. You've distilled it more clearly and lines up with my own intuition that reminds me about how human memory systems work: retrieval is effectively erasure, so "remembering" requires retrieval and storage. Research into treating PTSD (IIRC?) and such found that blocking storage (with drugs or EM) and then triggering recall would wipe memories. You're describing a potentially purely experiential way to do this, which is terrifying.

        I feel like using an LLM can lead to a Dunning-Kruger like effect, in that you think you know what it did, but you don't. And this belief satisfies the need/instinct to learn/know what happened without having actually done so. (Reminds me of making a TODO list and now the Dopamine hit from that kills the need to actually *do* the list.)

        kees@hachyderm.ioK This user is from outside of this forum
        kees@hachyderm.ioK This user is from outside of this forum
        kees@hachyderm.io
        wrote sidst redigeret af
        #146

        @glitzersachen @josh @silverwizard @ossguy @bkuhn @karen @wwahammy @xgranade

        I lump my experiences of software engineering use of LLMs into 3 modes:

        1) "work together", I am watching everything it is doing, reviewing every step, and contributing to the result in tandem. This doesn't feel to me like anything is being eroded on my end. But I'm also a deep sceptic of its output.

        2) "do the thing I know how to do for me", this is super dangerous, as I think I'm solving problems I am familiar with, but I didn't follow the results closely and I'm left with deep erosion of my comprehension of both problem and solution.

        3) "vibe coding", I have no idea what it is doing with a thing I don't know about and I know I have no idea what it is doing. This doesn't seem to erode anything. It does create a new problem for me, though, if the LLM can't solve some problem because also neither can I.

        I've felt #2 a few times, and I had the alarm bells in place to shift myself back to #1, which required doing full review and looking back through the reasoning and checking the work. The risk of being drawn into #2 is high given the sychophancy of the models, but I think my suspicion of it has helped avoid this a bit. 😅 (And perhaps I am more deluded than I think.)

        #3 I have done for educational/amusement purposes, but it's an uncommon mode for me because what's the point of creating a thing I don't understand and can't fix?

        ("I can quit any time!")

        1 Reply Last reply
        0
        • kees@hachyderm.ioK kees@hachyderm.io

          @glitzersachen @josh @silverwizard @ossguy @bkuhn @karen @wwahammy @xgranade

          Right, yeah, this is why I've cautioned people about *how* they use LLMs. You've distilled it more clearly and lines up with my own intuition that reminds me about how human memory systems work: retrieval is effectively erasure, so "remembering" requires retrieval and storage. Research into treating PTSD (IIRC?) and such found that blocking storage (with drugs or EM) and then triggering recall would wipe memories. You're describing a potentially purely experiential way to do this, which is terrifying.

          I feel like using an LLM can lead to a Dunning-Kruger like effect, in that you think you know what it did, but you don't. And this belief satisfies the need/instinct to learn/know what happened without having actually done so. (Reminds me of making a TODO list and now the Dopamine hit from that kills the need to actually *do* the list.)

          G This user is from outside of this forum
          G This user is from outside of this forum
          glitzersachen@hachyderm.io
          wrote sidst redigeret af
          #147

          @kees @josh @silverwizard @ossguy @bkuhn @karen @wwahammy @xgranade

          > making a TODO list and now the Dopamine hit from that

          Well, frankly, I don't get Dopamine hits from TODO lists. I rather get depression ...

          1 Reply Last reply
          0
          • josh@social.joshtriplett.orgJ josh@social.joshtriplett.org
            People shouldn't be *abused*, ever. If people are *shilling* AI and trying to force it on others, they might deserve some amount of shame and disapprobation. But nobody deserves abuse.
            G This user is from outside of this forum
            G This user is from outside of this forum
            glitzersachen@hachyderm.io
            wrote sidst redigeret af
            #148

            @josh @silverwizard @ossguy @bkuhn @karen @wwahammy

            Agreed. I only want to make clear that it should not be clear flying for AI shills. A certain amount of actually painful headwind should meet them.

            On the other side I consider what some middle management does to their employees with AI abuse (not sure about the apllicability of the English word here. I am not a native speaker).

            And in line with Popper's paradox of tolerance I wonder whether abuser _in a position of power_ should really be countered with slight slap on the hand and mild mannered words.

            After all it was this kind of behavior in such a power constellation that triggered revolutions historically. And I have a hard time rooting for the oppressors.

            1 Reply Last reply
            0
            • ossguy@fedi.copyleft.orgO ossguy@fedi.copyleft.org

              @davidgerard @wwahammy @silverwizard @firefly_lightning @cwebber Yes, which is why it's important to allow people to identify when they have used LLM/AI assistants to help. New contributors will see this is the norm, and then it will be easier to help them, because we'll know a bit about where any potential knowledge gaps might be coming from.

              If we "ban" LLM/AI-assisted contributions, people will use them anyway but hide their use, which is a trickier problem to solve.

              silverwizard@convenient.emailS This user is from outside of this forum
              silverwizard@convenient.emailS This user is from outside of this forum
              silverwizard@convenient.email
              wrote sidst redigeret af
              #149

              @ossguy @firefly_lightning @davidgerard @wwahammy @cwebber so is the theory "we should accept people who have been abused by big tech and think LLM usage is ok, and we must help them understand why it's not", or is "we must accept LLM code commits because there's people who think they're ok and the longterm damage is a risk we'll just need to figure out later"?

              I keep thinking it's the very reasonable first one, which I don't think is up for debate. But you and Bradley keep implying it's the second, which *is* up for debate.

              1 Reply Last reply
              0
              • bkuhn@fedi.copyleft.orgB bkuhn@fedi.copyleft.org

                @cwebber I think maybe you missed https://sfconservancy.org/blog/2026/mar/04/scotus-deny-cert-dc-circuit-thaler-appeal-llm-ai/ where #SFC analyzed that situation?
                Also, follow @ai_cases & see the *firehose* of litigation on this & remember the “Work Based on the Program” issue under GPLv2 has still never been litigated directly but lots of cases about 100% proprietary software have bolstered GPL's strength.

                Big Content has legal battles with Big Tech on 100s of fronts rn. Yes, we're adrift on their sea, but the situation is not as dire as you imagine.

                #AI #LLW

                cwebber@social.coopC This user is from outside of this forum
                cwebber@social.coopC This user is from outside of this forum
                cwebber@social.coop
                wrote sidst redigeret af
                #150

                @bkuhn @ossguy @richardfontana Continuing here, because it's the relevant subthread.

                I am sympathetic to choosing to narrow a topic. However, the post, in implying that we should start accepting partially AIgen contributions, inherently pulls in the topic of whether or not that is legally safe.

                Yes, I have read the previous Conservancy post about the existing cases. This partly contributes to my surprise and confusion about the post.

                Acknowledging that the plan is to have continued conversations and meetings about this, I still feel it is important to lay down my current concerns, even before such a meeting. I am leaving the "quality of contributions" and many other details out of here, and instead focusing on whether of not it is *safe to accept* contributions on copyright grounds at the moment, and what the implications of thinking on that are.

                (cotd)

                cwebber@social.coopC shaadra@mastodon.gamedev.placeS 2 Replies Last reply
                0
                • cwebber@social.coopC cwebber@social.coop

                  @bkuhn @ossguy @richardfontana Continuing here, because it's the relevant subthread.

                  I am sympathetic to choosing to narrow a topic. However, the post, in implying that we should start accepting partially AIgen contributions, inherently pulls in the topic of whether or not that is legally safe.

                  Yes, I have read the previous Conservancy post about the existing cases. This partly contributes to my surprise and confusion about the post.

                  Acknowledging that the plan is to have continued conversations and meetings about this, I still feel it is important to lay down my current concerns, even before such a meeting. I am leaving the "quality of contributions" and many other details out of here, and instead focusing on whether of not it is *safe to accept* contributions on copyright grounds at the moment, and what the implications of thinking on that are.

                  (cotd)

                  cwebber@social.coopC This user is from outside of this forum
                  cwebber@social.coopC This user is from outside of this forum
                  cwebber@social.coop
                  wrote sidst redigeret af
                  #151

                  @bkuhn @ossguy @richardfontana So the question is: is it safe, from a legal perspective, given the current state of uncertainty of copyright of such contributions, to encourage accepting such contributions into repositories?

                  Now clearly, many projects are: the Linux kernel most famously is, and their recent policy document says effectively, "You can contribute AI generated code, but the onus is on you whether or not you legally could have".

                  Which is not very helpful of a handwave, I would say, since few contributors are equipped to assess such a thing. I've left myself and three others addressed in this portion of the thread, and all of us *have* done licensing work, and my suspicion is, *especially* based on what's been written, that none of us could confidently project where things are going to go.

                  cwebber@social.coopC zacchiro@mastodon.xyzZ 2 Replies Last reply
                  0
                  • cwebber@social.coopC cwebber@social.coop

                    @bkuhn @ossguy @richardfontana So the question is: is it safe, from a legal perspective, given the current state of uncertainty of copyright of such contributions, to encourage accepting such contributions into repositories?

                    Now clearly, many projects are: the Linux kernel most famously is, and their recent policy document says effectively, "You can contribute AI generated code, but the onus is on you whether or not you legally could have".

                    Which is not very helpful of a handwave, I would say, since few contributors are equipped to assess such a thing. I've left myself and three others addressed in this portion of the thread, and all of us *have* done licensing work, and my suspicion is, *especially* based on what's been written, that none of us could confidently project where things are going to go.

                    cwebber@social.coopC This user is from outside of this forum
                    cwebber@social.coopC This user is from outside of this forum
                    cwebber@social.coop
                    wrote sidst redigeret af
                    #152

                    @bkuhn @ossguy @richardfontana Part of the problem here is that the AI companies have set the stage themselves. Their presumption is that it's fine to absorb effectively all open and "indie" content, and that this is entirely fair to pull into a model without any legal implications, whereas potentially yes, you may need to "license" something that looks like a Disney character. In the land of code, I also sense that Microsoft is perfectly fine with the idea that you can "copyright launder" a codebase from the GPL to perhaps the public domain, but if someone did that to their own leaked source code, they would be very upset.

                    Meanwhile, a friend of mine who works in films has said that he keeps hearing rumors that OpenAI would like a cut of stuff made with their stuff. We should presume tthat true.

                    Regardless, I'm sure everyone on this thread wants an *equitable* situation for proprietary and FOSS licensing. I'll expand on that more in a moment though.

                    cwebber@social.coopC zleap@techhub.socialZ 2 Replies Last reply
                    0
                    • cwebber@social.coopC cwebber@social.coop

                      @bkuhn @ossguy @richardfontana Part of the problem here is that the AI companies have set the stage themselves. Their presumption is that it's fine to absorb effectively all open and "indie" content, and that this is entirely fair to pull into a model without any legal implications, whereas potentially yes, you may need to "license" something that looks like a Disney character. In the land of code, I also sense that Microsoft is perfectly fine with the idea that you can "copyright launder" a codebase from the GPL to perhaps the public domain, but if someone did that to their own leaked source code, they would be very upset.

                      Meanwhile, a friend of mine who works in films has said that he keeps hearing rumors that OpenAI would like a cut of stuff made with their stuff. We should presume tthat true.

                      Regardless, I'm sure everyone on this thread wants an *equitable* situation for proprietary and FOSS licensing. I'll expand on that more in a moment though.

                      cwebber@social.coopC This user is from outside of this forum
                      cwebber@social.coopC This user is from outside of this forum
                      cwebber@social.coop
                      wrote sidst redigeret af
                      #153

                      However, it's not actually the laundering angle I am concerned with here entirely, it's whether we're turning FOSS codebases into potential legal toxic waste dumps that we will have a hell of a time cleaning up later.

                      The previous Conservancy post, which @bkuhn linked upthread, indicates that Conservancy does indeed consider the matter unsettled.

                      Current LLMs wouldn't "default to copyleft", since they also include all-rights-reserved mixed in there. If the result of output of these systems is a slurry of inputs which carry their licensing somehow, their default licensing output situation is one of a hazard.

                      I note that @bkuhn and @ossguy seem to be hinting at hoping a "copyleft based LLM" with all-copyleft output it a winning scenario. I'm going to state plainly: I believe that's an impossible outcome.

                      @richardfontana

                      cwebber@social.coopC evan@cosocial.caE dalias@hachyderm.ioD 3 Replies Last reply
                      0
                      • cwebber@social.coopC cwebber@social.coop

                        However, it's not actually the laundering angle I am concerned with here entirely, it's whether we're turning FOSS codebases into potential legal toxic waste dumps that we will have a hell of a time cleaning up later.

                        The previous Conservancy post, which @bkuhn linked upthread, indicates that Conservancy does indeed consider the matter unsettled.

                        Current LLMs wouldn't "default to copyleft", since they also include all-rights-reserved mixed in there. If the result of output of these systems is a slurry of inputs which carry their licensing somehow, their default licensing output situation is one of a hazard.

                        I note that @bkuhn and @ossguy seem to be hinting at hoping a "copyleft based LLM" with all-copyleft output it a winning scenario. I'm going to state plainly: I believe that's an impossible outcome.

                        @richardfontana

                        cwebber@social.coopC This user is from outside of this forum
                        cwebber@social.coopC This user is from outside of this forum
                        cwebber@social.coop
                        wrote sidst redigeret af
                        #154

                        @bkuhn @ossguy @richardfontana Rather than focus on the GPL, let's choose a different copyleft license. In fact, let's choose a gradient of licenses.

                        - CC0's public domain declaration w/ minimal fallback license
                        - CC BY
                        - CC BY-SA

                        Imagine for a moment an LLM trained entirely on the above three licenses, and then one that's CC BY and CC0, and then one that's just CC0.

                        Let's look at both extremes and then we'll find out the real dangers come from observing the middle.

                        cwebber@social.coopC 1 Reply Last reply
                        0
                        • cwebber@social.coopC cwebber@social.coop

                          @bkuhn @ossguy @richardfontana Rather than focus on the GPL, let's choose a different copyleft license. In fact, let's choose a gradient of licenses.

                          - CC0's public domain declaration w/ minimal fallback license
                          - CC BY
                          - CC BY-SA

                          Imagine for a moment an LLM trained entirely on the above three licenses, and then one that's CC BY and CC0, and then one that's just CC0.

                          Let's look at both extremes and then we'll find out the real dangers come from observing the middle.

                          cwebber@social.coopC This user is from outside of this forum
                          cwebber@social.coopC This user is from outside of this forum
                          cwebber@social.coop
                          wrote sidst redigeret af
                          #155

                          @bkuhn @ossguy @richardfontana First let's imagine the only-CC0 based LLM.

                          I would fully agree that no matter the law and legal case law passed and established, the CC0 based input LLM is clearly effectively in the public domain, or like CC0 itself, equivalent to it. This one is relatively simple.

                          Let's make things more complicated.

                          cwebber@social.coopC 1 Reply Last reply
                          0
                          • cwebber@social.coopC cwebber@social.coop

                            @bkuhn @ossguy @richardfontana First let's imagine the only-CC0 based LLM.

                            I would fully agree that no matter the law and legal case law passed and established, the CC0 based input LLM is clearly effectively in the public domain, or like CC0 itself, equivalent to it. This one is relatively simple.

                            Let's make things more complicated.

                            cwebber@social.coopC This user is from outside of this forum
                            cwebber@social.coopC This user is from outside of this forum
                            cwebber@social.coop
                            wrote sidst redigeret af
                            #156

                            @bkuhn @ossguy @richardfontana Regarding the one containing CC0, CC BY, and CC BY-SA, the situation is more uncertain and seems highly affected by legal outcomes in upcoming law and cases to be set. There is the possibility that indeed, the LLM is considered a slurry of inputs and this is legally acceptable, and effectively any output which is not verbatim of its inputs in some way is effectively under the public domain.

                            Now, of course, the problem is that we don't have to just worry about the US, we have to worry *internationally*. When considered from this angle, that FOSS is an international endeavour, this hope that things are in the public domain feels a lot dicier.

                            The assumption is that then this effectively leads to the output being under the terms of CC BY-SA. This is fine, great even, right?! Because effectively everything is share-alike (Bradley I don't wanna get into whether BY-SA is copyleft or something weaker). We slap CC BY-SA on the output, it's fine. Right??????

                            cwebber@social.coopC 1 Reply Last reply
                            0
                            • cwebber@social.coopC cwebber@social.coop

                              @bkuhn @ossguy @richardfontana Regarding the one containing CC0, CC BY, and CC BY-SA, the situation is more uncertain and seems highly affected by legal outcomes in upcoming law and cases to be set. There is the possibility that indeed, the LLM is considered a slurry of inputs and this is legally acceptable, and effectively any output which is not verbatim of its inputs in some way is effectively under the public domain.

                              Now, of course, the problem is that we don't have to just worry about the US, we have to worry *internationally*. When considered from this angle, that FOSS is an international endeavour, this hope that things are in the public domain feels a lot dicier.

                              The assumption is that then this effectively leads to the output being under the terms of CC BY-SA. This is fine, great even, right?! Because effectively everything is share-alike (Bradley I don't wanna get into whether BY-SA is copyleft or something weaker). We slap CC BY-SA on the output, it's fine. Right??????

                              cwebber@social.coopC This user is from outside of this forum
                              cwebber@social.coopC This user is from outside of this forum
                              cwebber@social.coop
                              wrote sidst redigeret af
                              #157

                              @bkuhn @ossguy @richardfontana Except, I actually believe this scenario isn't legally viable. And it's easier to understand if we scale back to the middle case.

                              Let's now look at the LLM trained on CC0 and CC BY. Because it's the BY aspect that makes everything complicated.

                              There is *NO WAY* in current LLM technology, nor I believe from studying how neural networks work, any viable computationally performant LLM, that they can track provenance. The BY clause cannot be upheld.

                              This isn't a theoretical concern for me; someone built another vibecoded Scheme-to-WASM-GC compiler that looks an awful lot like Spritely's own Hoot compiler in places. They didn't attribute us. They probably didn't know. But like many FOSS licenses, Apache v2 does require certain levels of attribution to be upheld. Most FOSS projects do.

                              You can't uphold the CC BY requirement, as far as I can tell.

                              cwebber@social.coopC richardfontana@mastodon.socialR 2 Replies Last reply
                              0
                              • cwebber@social.coopC cwebber@social.coop

                                @bkuhn @ossguy @richardfontana Except, I actually believe this scenario isn't legally viable. And it's easier to understand if we scale back to the middle case.

                                Let's now look at the LLM trained on CC0 and CC BY. Because it's the BY aspect that makes everything complicated.

                                There is *NO WAY* in current LLM technology, nor I believe from studying how neural networks work, any viable computationally performant LLM, that they can track provenance. The BY clause cannot be upheld.

                                This isn't a theoretical concern for me; someone built another vibecoded Scheme-to-WASM-GC compiler that looks an awful lot like Spritely's own Hoot compiler in places. They didn't attribute us. They probably didn't know. But like many FOSS licenses, Apache v2 does require certain levels of attribution to be upheld. Most FOSS projects do.

                                You can't uphold the CC BY requirement, as far as I can tell.

                                cwebber@social.coopC This user is from outside of this forum
                                cwebber@social.coopC This user is from outside of this forum
                                cwebber@social.coop
                                wrote sidst redigeret af
                                #158

                                @bkuhn @ossguy @richardfontana Now here is a counter-argument: how do people attribute Wikipedia? They generally just attribute Wikipedia! And people seem to be mostly fine with this.

                                It feels fine, when you were a contributor to the Wikipedia project.

                                It feels a lot less fine when you are a contributor to a specific project, to have everything just sucked up into "the generic LLM". Claude did it! Claude did it all by itself.

                                cwebber@social.coopC johl@mastodon.xyzJ 2 Replies Last reply
                                0
                                • cwebber@social.coopC cwebber@social.coop

                                  @bkuhn @ossguy @richardfontana Now here is a counter-argument: how do people attribute Wikipedia? They generally just attribute Wikipedia! And people seem to be mostly fine with this.

                                  It feels fine, when you were a contributor to the Wikipedia project.

                                  It feels a lot less fine when you are a contributor to a specific project, to have everything just sucked up into "the generic LLM". Claude did it! Claude did it all by itself.

                                  cwebber@social.coopC This user is from outside of this forum
                                  cwebber@social.coopC This user is from outside of this forum
                                  cwebber@social.coop
                                  wrote sidst redigeret af
                                  #159

                                  @bkuhn @ossguy @richardfontana If we are pushing for an *equitable* scenario for copyright output, there is only one "good outcome" in terms of copyright, and that is that everything is effectively in the public domain. The dream of having a "copyleft LLM" doesn't work.

                                  And even if it did, there are several problems:

                                  - Nobody is using that *now*, and contributors are facing contributions *now*, and there is legal uncertainty about accepting those contributions *right now*.
                                  - It is unlikely that the "copyleft LLM" would be very useful. The way people use these tools is conversational in a way that requires them to effectively have to be trained on the entire internet to be functional. Not just copyleft codebases.

                                  The copyleft LLM dream is a joke.

                                  cwebber@social.coopC riley@toot.catR 2 Replies Last reply
                                  0
                                  • cwebber@social.coopC cwebber@social.coop

                                    @bkuhn @ossguy @richardfontana If we are pushing for an *equitable* scenario for copyright output, there is only one "good outcome" in terms of copyright, and that is that everything is effectively in the public domain. The dream of having a "copyleft LLM" doesn't work.

                                    And even if it did, there are several problems:

                                    - Nobody is using that *now*, and contributors are facing contributions *now*, and there is legal uncertainty about accepting those contributions *right now*.
                                    - It is unlikely that the "copyleft LLM" would be very useful. The way people use these tools is conversational in a way that requires them to effectively have to be trained on the entire internet to be functional. Not just copyleft codebases.

                                    The copyleft LLM dream is a joke.

                                    cwebber@social.coopC This user is from outside of this forum
                                    cwebber@social.coopC This user is from outside of this forum
                                    cwebber@social.coop
                                    wrote sidst redigeret af
                                    #160

                                    @bkuhn @ossguy @richardfontana I say "good outcome", and I'm not saying it's an outcome I want, because "what I want" is pretty complicated here. I'm saying, it's the only one where there is the possibility of legal output from these tools that can safely be incorporated into FOSS projects *at all* that is *equitable* for both FOSS and proprietary situations.

                                    And yup, unfortunately, that would mean copyright-laundering of FOSS codebases through LLMs would be possible to strip copyleft.

                                    It would also mean the same for proprietary codebases.

                                    Frankly I think it would kind of rule if we stabbed copyright in the gut that badly, but there's so much vested interest from various copyright holding corporations, I don't think we're likely to get that. Do you?

                                    cwebber@social.coopC trwnh@mastodon.socialT richardjacton@fosstodon.orgR 3 Replies Last reply
                                    0
                                    • cwebber@social.coopC cwebber@social.coop

                                      @bkuhn @ossguy @richardfontana I say "good outcome", and I'm not saying it's an outcome I want, because "what I want" is pretty complicated here. I'm saying, it's the only one where there is the possibility of legal output from these tools that can safely be incorporated into FOSS projects *at all* that is *equitable* for both FOSS and proprietary situations.

                                      And yup, unfortunately, that would mean copyright-laundering of FOSS codebases through LLMs would be possible to strip copyleft.

                                      It would also mean the same for proprietary codebases.

                                      Frankly I think it would kind of rule if we stabbed copyright in the gut that badly, but there's so much vested interest from various copyright holding corporations, I don't think we're likely to get that. Do you?

                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coop
                                      wrote sidst redigeret af
                                      #161

                                      @bkuhn @ossguy @richardfontana So let me summarize:

                                      - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
                                      - The idea of a copyleft-only LLM is a joke and we should not rely on it
                                      - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

                                      That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

                                      But we can't ignore this right now.

                                      We're putting FOSS codebases at risk.

                                      lordcaramac@discordian.socialL jtnystrom@genomic.socialJ jens@social.finkhaeuser.deJ richardfontana@mastodon.socialR fuzzychef@m6n.ioF 8 Replies Last reply
                                      1
                                      0
                                      • cwebber@social.coopC cwebber@social.coop

                                        @bkuhn @ossguy @richardfontana So let me summarize:

                                        - Without knowing the legal status of accepting LLM contributions, we're potentially polluting our codebases with stuff that we are going to have a HELL of a time cleaning up later
                                        - The idea of a copyleft-only LLM is a joke and we should not rely on it
                                        - We really only have two realistic scenarios: either FOSS projects cannot accept LLM based contributions legally from an international perspective, or everything is effectively in the public domain as outputted from these machines, but at least in the latter scenario we get to weaken copyright for everyone.

                                        That's leaving out a lot of other considerations about LLMs and the ethics of using them, which I think most of the other replies were focused on, I largely focused on the copyright implications aspects in this subthread. Because yes, I agree, it can be important to focus a conversation.

                                        But we can't ignore this right now.

                                        We're putting FOSS codebases at risk.

                                        lordcaramac@discordian.socialL This user is from outside of this forum
                                        lordcaramac@discordian.socialL This user is from outside of this forum
                                        lordcaramac@discordian.social
                                        wrote sidst redigeret af
                                        #162

                                        @cwebber @bkuhn @ossguy @richardfontana I think we should just destroy copyright entirely and expand the public domain to contain everything that has ever been published. Intellectual property was a very bad idea in the first place IMHO.

                                        cwebber@social.coopC 1 Reply Last reply
                                        0
                                        • cwebber@social.coopC cwebber@social.coop

                                          @bkuhn @ossguy @richardfontana Part of the problem here is that the AI companies have set the stage themselves. Their presumption is that it's fine to absorb effectively all open and "indie" content, and that this is entirely fair to pull into a model without any legal implications, whereas potentially yes, you may need to "license" something that looks like a Disney character. In the land of code, I also sense that Microsoft is perfectly fine with the idea that you can "copyright launder" a codebase from the GPL to perhaps the public domain, but if someone did that to their own leaked source code, they would be very upset.

                                          Meanwhile, a friend of mine who works in films has said that he keeps hearing rumors that OpenAI would like a cut of stuff made with their stuff. We should presume tthat true.

                                          Regardless, I'm sure everyone on this thread wants an *equitable* situation for proprietary and FOSS licensing. I'll expand on that more in a moment though.

                                          zleap@techhub.socialZ This user is from outside of this forum
                                          zleap@techhub.socialZ This user is from outside of this forum
                                          zleap@techhub.social
                                          wrote sidst redigeret af
                                          #163

                                          @cwebber @bkuhn @ossguy @richardfontana

                                          Indeed, big tech know full well the FLOSS / indie creators don't have the legal funds to defend. Their IP either.

                                          bkuhn@fedi.copyleft.orgB 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper