Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. Free software people: A major goal of free software is for individuals to be able to cause software to behave in the way they want it toLLMs: (enable that)Free software people: Oh no not like that

Free software people: A major goal of free software is for individuals to be able to cause software to behave in the way they want it toLLMs: (enable that)Free software people: Oh no not like that

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
317 Indlæg 120 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • kyle@mastodon.kylerank.inK kyle@mastodon.kylerank.in

    @mjg59 You will get backlash, but you are right.

    Free software folks will have to decide whether what they really wanted was *everyone* to have the freedom to use and modify software, or only that subset of everyone who had the privilege of learning software development.

    There has always been this elitist dividing line in the community between people who contribute code, and people who contribute all the other things FOSS needs to thrive. Now those people can contribute code too.

    zachdecook@social.librem.oneZ This user is from outside of this forum
    zachdecook@social.librem.oneZ This user is from outside of this forum
    zachdecook@social.librem.one
    wrote sidst redigeret af
    #162

    @kyle @mjg59 Proprietary tooling is the reason "Stallman was right" about Bitkeeper, but "everyone was better off for having not listened to him" is the pragmatic side.
    Yes, I want people to benefit from the freedom to modify code, but they will never truly be free if they are using a proprietary LLM to make their modifications.

    1 Reply Last reply
    0
    • chris_evelyn@fedi.chris-evelyn.deC chris_evelyn@fedi.chris-evelyn.de

      @mjg59 Yeah, as soon as there‘s an ethically sourced and trained free LLM that‘s not controlled by very shitty companies I‘m totally on board with you.

      Until then we shouldn’t let that shit near our projects.

      light@noc.socialL This user is from outside of this forum
      light@noc.socialL This user is from outside of this forum
      light@noc.social
      wrote sidst redigeret af
      #163

      @chris_evelyn
      What do you mean by "ethically sourced and trained"?
      @mjg59

      1 Reply Last reply
      0
      • mnl@hachyderm.ioM This user is from outside of this forum
        mnl@hachyderm.ioM This user is from outside of this forum
        mnl@hachyderm.io
        wrote sidst redigeret af
        #164

        @engideer @david_chisnall @mjg59 @ignaloidas I don’t think llms are “rando”. They have randomized elements during training and inference, but they’re not a random number generator. I also would trust a “rando” less than an expert in real life. I wouldn’t trust either blindly either.

        mnl@hachyderm.ioM ignaloidas@not.acu.ltI 2 Replies Last reply
        0
        • promovicz@chaos.socialP promovicz@chaos.social

          @mjg59 What you propose is actually illegal, even if the law doesn’t make much sense. I wonder if you ever had the cops sent after you on a corp-run IP case… maybe it would make you feel different?

          light@noc.socialL This user is from outside of this forum
          light@noc.socialL This user is from outside of this forum
          light@noc.social
          wrote sidst redigeret af
          #165

          @promovicz
          Let's hope the AI lobby will (in any combination of purposely and inadvertently) make that law obsolete.
          @mjg59

          1 Reply Last reply
          0
          • mjg59@nondeterministic.computerM mjg59@nondeterministic.computer

            Free software people: A major goal of free software is for individuals to be able to cause software to behave in the way they want it to
            LLMs: (enable that)
            Free software people: Oh no not like that

            jordan@mastodon.subj.amJ This user is from outside of this forum
            jordan@mastodon.subj.amJ This user is from outside of this forum
            jordan@mastodon.subj.am
            wrote sidst redigeret af
            #166

            @mjg59 I think the issue is more on the forcing of LLMs/AI in *everything* right now, not specifically F/OSS projects. It reeks of dot-com bubble era marketing and in many cases is completely unnecessary.

            1 Reply Last reply
            0
            • mnl@hachyderm.ioM mnl@hachyderm.io

              @engideer @david_chisnall @mjg59 @ignaloidas I don’t think llms are “rando”. They have randomized elements during training and inference, but they’re not a random number generator. I also would trust a “rando” less than an expert in real life. I wouldn’t trust either blindly either.

              mnl@hachyderm.ioM This user is from outside of this forum
              mnl@hachyderm.ioM This user is from outside of this forum
              mnl@hachyderm.io
              wrote sidst redigeret af
              #167

              @engideer @david_chisnall @mjg59 @ignaloidas also I didn’t say anything of what you quoted, and I don’t know where you got it from.

              1 Reply Last reply
              0
              • mnl@hachyderm.ioM mnl@hachyderm.io

                @ignaloidas @mjg59 @david_chisnall @newhinton how did you gain your confidence? How can you call machine learning a bunch of dice? I try to study and build things everyday and yes I don’t trust my code at all, which I think is a healthy attitude to have? I am definitely not able to produce perfect code on the first try.

                ignaloidas@not.acu.ltI This user is from outside of this forum
                ignaloidas@not.acu.ltI This user is from outside of this forum
                ignaloidas@not.acu.lt
                wrote sidst redigeret af
                #168

                @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @newhinton@troet.cafe through repeated checks and knowledge that humans are consistent.


                And like, really, you don't trust your code at all? I, for example, know that the code I wrote is not going to cheat by unit tests, not going to re-implement half of the things from scratch when I'm working on a small feature, nor will it randomly delete files. After working with people for a while, I can be fairly sure that the code they've written can be trusted to the same standards. LLMs can't be trusted with these things, and in fact have been documented to do all of these things.

                It is not a blind, absolute trust, but trust within reason. The fact that I have to explain this to you is honestly embarrassing.

                mnl@hachyderm.ioM 1 Reply Last reply
                0
                • mjg59@nondeterministic.computerM mjg59@nondeterministic.computer

                  Free software people: A major goal of free software is for individuals to be able to cause software to behave in the way they want it to
                  LLMs: (enable that)
                  Free software people: Oh no not like that

                  condret@shitposter.worldC This user is from outside of this forum
                  condret@shitposter.worldC This user is from outside of this forum
                  condret@shitposter.world
                  wrote sidst redigeret af
                  #169
                  @mjg59 i actually like LLMs
                  1 Reply Last reply
                  0
                  • petko@social.petko.meP petko@social.petko.me

                    @mjg59 but wait, there's more

                    What if you're not renowned security expert and open-source celebrity @mjg59 (that currently works at nvidia btw, profiting from the LLM boom, sorry) but just some guy trying to make ends meet doing some coding?...

                    Now you get an LLM mandate from your company that comes with the implication that 'either you boost your productivity with 80% or we fire you and contract a cheap prompter in your place'...

                    lasombra_br@mas.toL This user is from outside of this forum
                    lasombra_br@mas.toL This user is from outside of this forum
                    lasombra_br@mas.to
                    wrote sidst redigeret af
                    #170

                    @petko @mjg59 You can see that there’s no care for any of that. It’s all “like LLMs? Good, go use it, it’s fun”. All your ethical believes go out of the window as soon as your company shares depend on the hype.

                    1 Reply Last reply
                    0
                    • mnl@hachyderm.ioM mnl@hachyderm.io

                      @engideer @david_chisnall @mjg59 @ignaloidas I don’t think llms are “rando”. They have randomized elements during training and inference, but they’re not a random number generator. I also would trust a “rando” less than an expert in real life. I wouldn’t trust either blindly either.

                      ignaloidas@not.acu.ltI This user is from outside of this forum
                      ignaloidas@not.acu.ltI This user is from outside of this forum
                      ignaloidas@not.acu.lt
                      wrote sidst redigeret af
                      #171

                      @mnl@hachyderm.io @engideer@tech.lgbt @david_chisnall@infosec.exchange @mjg59@nondeterministic.computer LLMs are very much a random number generators. The distribution is far, far from uniform, but the whole breakthrough of LLMs was the introduction of "temperature", quite literally random choices, to break them out of monotonous tendencies.

                      mnl@hachyderm.ioM 1 Reply Last reply
                      0
                      • ignaloidas@not.acu.ltI ignaloidas@not.acu.lt

                        @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @newhinton@troet.cafe through repeated checks and knowledge that humans are consistent.


                        And like, really, you don't trust your code at all? I, for example, know that the code I wrote is not going to cheat by unit tests, not going to re-implement half of the things from scratch when I'm working on a small feature, nor will it randomly delete files. After working with people for a while, I can be fairly sure that the code they've written can be trusted to the same standards. LLMs can't be trusted with these things, and in fact have been documented to do all of these things.

                        It is not a blind, absolute trust, but trust within reason. The fact that I have to explain this to you is honestly embarrassing.

                        mnl@hachyderm.ioM This user is from outside of this forum
                        mnl@hachyderm.ioM This user is from outside of this forum
                        mnl@hachyderm.io
                        wrote sidst redigeret af
                        #172

                        @ignaloidas @mjg59 @david_chisnall @newhinton but “fairly sure” is not full trust. I can also be “fairly sure” that something works, but I’m not going to trust my judgment and instead will try to validate it and provide proper guardrails so that if it is misbehaving, it is at least contained. Some things will be just fine even if broken, some less and will make me invest me more of my time. I am not going to try to prove the kernel correct just because I am changing a css color. I don’t see how that is different with llms, and I use them everyday. If anything, they allow me to validate more.

                        ignaloidas@not.acu.ltI 1 Reply Last reply
                        0
                        • radex@social.hackerspace.plR radex@social.hackerspace.pl

                          @mjg59 This doesn't feel right to me. IMO few people actually object to use of LLMs by individuals for tinkering on personal stuff.

                          The criticism as I see it is primarily that:
                          1) there are huge societal/political impacts - uncompensated use of copyrighted material; benefits of it accruing primarily to a few big players; energy use; layoffs; perceived misallocation of massive amounts of capital
                          2) the output quality of LLMs is t r a s h, unsuitable for professional use

                          condret@shitposter.worldC This user is from outside of this forum
                          condret@shitposter.worldC This user is from outside of this forum
                          condret@shitposter.world
                          wrote sidst redigeret af
                          #173
                          @radex @mjg59 2) is not true. glm-5 produces actually good code most of the time. sure you need to do a few adjustments here and there from time to time, but it isn't trash
                          toiletpaper@shitposter.worldT 1 Reply Last reply
                          0
                          • mnl@hachyderm.ioM mnl@hachyderm.io

                            @ignaloidas @mjg59 @david_chisnall @newhinton but “fairly sure” is not full trust. I can also be “fairly sure” that something works, but I’m not going to trust my judgment and instead will try to validate it and provide proper guardrails so that if it is misbehaving, it is at least contained. Some things will be just fine even if broken, some less and will make me invest me more of my time. I am not going to try to prove the kernel correct just because I am changing a css color. I don’t see how that is different with llms, and I use them everyday. If anything, they allow me to validate more.

                            ignaloidas@not.acu.ltI This user is from outside of this forum
                            ignaloidas@not.acu.ltI This user is from outside of this forum
                            ignaloidas@not.acu.lt
                            wrote sidst redigeret af
                            #174

                            @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @newhinton@troet.cafe you are falling down the cryptocurrency fallacy, assuming that you cannot trust anyone and as such have to build stuff assuming everyone is looking to get one over you.

                            This is tiresome, and I do not care to discuss with you on this any longer, if you cannot understand that there are levels between "no trust" and "absolute trust", there is nothing more to discuss.

                            mnl@hachyderm.ioM 1 Reply Last reply
                            0
                            • ignaloidas@not.acu.ltI ignaloidas@not.acu.lt

                              @mnl@hachyderm.io @engideer@tech.lgbt @david_chisnall@infosec.exchange @mjg59@nondeterministic.computer LLMs are very much a random number generators. The distribution is far, far from uniform, but the whole breakthrough of LLMs was the introduction of "temperature", quite literally random choices, to break them out of monotonous tendencies.

                              mnl@hachyderm.ioM This user is from outside of this forum
                              mnl@hachyderm.ioM This user is from outside of this forum
                              mnl@hachyderm.io
                              wrote sidst redigeret af
                              #175

                              @ignaloidas @mjg59 @david_chisnall @engideer temperature based sampling is just one of the many sampling modalities. Nucleus sampling, top-k, frequency penalties, all of these introduce controlled randomness to improve the performance of llms as measured by a wide variety of benchmarks.

                              A random sampling of tokens would actually be uniformly distributed… and obviously grammatically correct sentences is a clear sign that we are not randomly sampling tokens.

                              Are we talking about the same thing?

                              ignaloidas@not.acu.ltI 1 Reply Last reply
                              0
                              • condret@shitposter.worldC condret@shitposter.world
                                @radex @mjg59 2) is not true. glm-5 produces actually good code most of the time. sure you need to do a few adjustments here and there from time to time, but it isn't trash
                                toiletpaper@shitposter.worldT This user is from outside of this forum
                                toiletpaper@shitposter.worldT This user is from outside of this forum
                                toiletpaper@shitposter.world
                                wrote sidst redigeret af
                                #176
                                @condret @radex @mjg59

                                The most useful pattern for using AI code assistance in my experience is test-driven-development. As long as you make sure the tests are robust and have good coverage, the rest is pretty much hands-free. That's not all there is to it, but it's the biggest bang for buck IME.
                                1 Reply Last reply
                                0
                                • ignaloidas@not.acu.ltI ignaloidas@not.acu.lt

                                  @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @newhinton@troet.cafe you are falling down the cryptocurrency fallacy, assuming that you cannot trust anyone and as such have to build stuff assuming everyone is looking to get one over you.

                                  This is tiresome, and I do not care to discuss with you on this any longer, if you cannot understand that there are levels between "no trust" and "absolute trust", there is nothing more to discuss.

                                  mnl@hachyderm.ioM This user is from outside of this forum
                                  mnl@hachyderm.ioM This user is from outside of this forum
                                  mnl@hachyderm.io
                                  wrote sidst redigeret af
                                  #177

                                  @ignaloidas @mjg59 @david_chisnall @newhinton I think you are misreading what I am saying. That is exactly what I am saying. I never fully trust my code, not a single line of it, partly because every line of my code usually requires billions of lines of code I haven’t written to run. I can apply methods and use my experience to trust it enough to run it.

                                  alerque@mastodon.socialA 1 Reply Last reply
                                  0
                                  • mnl@hachyderm.ioM mnl@hachyderm.io

                                    @ignaloidas @mjg59 @david_chisnall @engideer temperature based sampling is just one of the many sampling modalities. Nucleus sampling, top-k, frequency penalties, all of these introduce controlled randomness to improve the performance of llms as measured by a wide variety of benchmarks.

                                    A random sampling of tokens would actually be uniformly distributed… and obviously grammatically correct sentences is a clear sign that we are not randomly sampling tokens.

                                    Are we talking about the same thing?

                                    ignaloidas@not.acu.ltI This user is from outside of this forum
                                    ignaloidas@not.acu.ltI This user is from outside of this forum
                                    ignaloidas@not.acu.lt
                                    wrote sidst redigeret af
                                    #178

                                    @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @engideer@tech.lgbt the fact that something is random does not mean that it has a uniform distribution. "controlled randomness" is still randomness. Taking random points in a unit circle by taking two random numbers for distance and direction will not result in a uniform distribution, but it's still random.

                                    like, do you even read what you're writing? I'm starting to understand why you don't trust the code you wrote

                                    mnl@hachyderm.ioM 1 Reply Last reply
                                    0
                                    • ignaloidas@not.acu.ltI ignaloidas@not.acu.lt

                                      @mnl@hachyderm.io @mjg59@nondeterministic.computer @david_chisnall@infosec.exchange @engideer@tech.lgbt the fact that something is random does not mean that it has a uniform distribution. "controlled randomness" is still randomness. Taking random points in a unit circle by taking two random numbers for distance and direction will not result in a uniform distribution, but it's still random.

                                      like, do you even read what you're writing? I'm starting to understand why you don't trust the code you wrote

                                      mnl@hachyderm.ioM This user is from outside of this forum
                                      mnl@hachyderm.ioM This user is from outside of this forum
                                      mnl@hachyderm.io
                                      wrote sidst redigeret af
                                      #179

                                      @ignaloidas @mjg59 @david_chisnall @engideer now you are talking about absolute trust. I do think we are indeed talking about different things. Do you use LLMs? Do you assign the same level of trust to qwen-3.6 than to gpt-2? because I do not, partly based on benchmarks, partly on personal experience, partly on my (admittedly perfunctory) theoretical understanding of its training and inference setup.

                                      1 Reply Last reply
                                      0
                                      • mjg59@nondeterministic.computerM mjg59@nondeterministic.computer

                                        Look, coders, we are not writers. There's no way to turn "increment this variable" into life changing prose. The creativity exists outside the code. It always has done and it always will do. Let it go.

                                        jens@social.finkhaeuser.deJ This user is from outside of this forum
                                        jens@social.finkhaeuser.deJ This user is from outside of this forum
                                        jens@social.finkhaeuser.de
                                        wrote sidst redigeret af
                                        #180

                                        @mjg59 Indeed.

                                        This is why code generation is not a solution to the problem.

                                        Which problem? People will phrase it differently, but the basic idea is to outsource *the hard part*, which is analysis and phrasing requirements to guide the LLM.

                                        LLMs suck at dealing with shitty specs. They even suck at dealing with good specs. They even suck at dealing with specs they themselves suggested.

                                        https://finkhaeuser.de/2026-04-10-outsourcing-thought-is-going-great/

                                        So using LLMs isn't solving the problem, which is that thinking is hard.

                                        1 Reply Last reply
                                        0
                                        • petko@social.petko.meP petko@social.petko.me

                                          @mjg59 but wait, there's more

                                          What if you're not renowned security expert and open-source celebrity @mjg59 (that currently works at nvidia btw, profiting from the LLM boom, sorry) but just some guy trying to make ends meet doing some coding?...

                                          Now you get an LLM mandate from your company that comes with the implication that 'either you boost your productivity with 80% or we fire you and contract a cheap prompter in your place'...

                                          S This user is from outside of this forum
                                          S This user is from outside of this forum
                                          seanfurey@mas.to
                                          wrote sidst redigeret af
                                          #181

                                          @petko @mjg59

                                          If the cheap prompter can produce the same results, what are the arguments against this?

                                          - copyright violation in the training material
                                          - excessively high use of the world's resources for training and inference

                                          If both of those were handled (that's a big if. Maybe someday, maybe not) what were the arguments be against choosing the cheap Proctor?

                                          petko@social.petko.meP 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper