Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
59 Indlæg 46 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • cwebber@social.coopC cwebber@social.coop

    I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

    Noooooooooo
    Noooooooooooooooooooooooooooo

    LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

    And they *certainly* can't be part of a reproducible pipeline

    thomasjwebb@mastodon.socialT This user is from outside of this forum
    thomasjwebb@mastodon.socialT This user is from outside of this forum
    thomasjwebb@mastodon.social
    wrote sidst redigeret af
    #43

    @cwebber It's pretty simple. If it's like a compiler, then why do you check in the output? And with all the work put into making compilers more efficient (not just making the *output* more efficient), why does it take so long and require an internet connection?

    1 Reply Last reply
    0
    • thomasjwebb@mastodon.socialT This user is from outside of this forum
      thomasjwebb@mastodon.socialT This user is from outside of this forum
      thomasjwebb@mastodon.social
      wrote sidst redigeret af
      #44

      @ansuz @cwebber @joeyh the reproducibility will also get pulled out as the model you used gets sunset. Unless all you check in is a series of prompts and a bunch of tests and simply assume future models will do a better job.

      It could even be a problem where future generations want a "vintage AI" look for whatever reason and unlike so many past generations of tech, they simply won't be able to because it was a cloud service and the company is long gone.

      ansuz@gts.cryptography.dogA 1 Reply Last reply
      0
      • cwebber@social.coopC cwebber@social.coop

        I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

        Noooooooooo
        Noooooooooooooooooooooooooooo

        LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

        And they *certainly* can't be part of a reproducible pipeline

        elrohir@mastodon.galE This user is from outside of this forum
        elrohir@mastodon.galE This user is from outside of this forum
        elrohir@mastodon.gal
        wrote sidst redigeret af
        #45

        @cwebber I'm only going to say that if natural human language was suitable for expressing expected response results in a predictable and well defined manner, we wouldn't have spent the last 50 years memorizing rulebooks that say "MUST means that the definition is an absolute requirement of the specification."

        At this point my rage almost goes beyond whether it's a LLM or a Witch's Cauldron taking the prompts. I want to scream at people NATURAL LANGUAGE IS NOT A RECOMMENDABLE INPUT FORMAT.

        1 Reply Last reply
        0
        • cwebber@social.coopC cwebber@social.coop

          I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

          Noooooooooo
          Noooooooooooooooooooooooooooo

          LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

          And they *certainly* can't be part of a reproducible pipeline

          millihertz@oldbytes.spaceM This user is from outside of this forum
          millihertz@oldbytes.spaceM This user is from outside of this forum
          millihertz@oldbytes.space
          wrote sidst redigeret af
          #46

          @cwebber well, it was until C99 anyway... 😕

          1 Reply Last reply
          0
          • fogti@chaos.socialF This user is from outside of this forum
            fogti@chaos.socialF This user is from outside of this forum
            fogti@chaos.social
            wrote sidst redigeret af
            #47

            @natty @cwebber Java2K (wait, that's more like stochastic interpreter)

            1 Reply Last reply
            0
            • thomasjwebb@mastodon.socialT thomasjwebb@mastodon.social

              @ansuz @cwebber @joeyh the reproducibility will also get pulled out as the model you used gets sunset. Unless all you check in is a series of prompts and a bunch of tests and simply assume future models will do a better job.

              It could even be a problem where future generations want a "vintage AI" look for whatever reason and unlike so many past generations of tech, they simply won't be able to because it was a cloud service and the company is long gone.

              ansuz@gts.cryptography.dogA This user is from outside of this forum
              ansuz@gts.cryptography.dogA This user is from outside of this forum
              ansuz@gts.cryptography.dog
              wrote sidst redigeret af
              #48

              @thomasjwebb @cwebber @joeyh 💯​

              Local models like llama could be reworked to accept a seed for their RNG. There'd be less risk of them becoming unavailable, and they'd be both deterministic and reproducible, but they'd still be terrible for all the other reasons that LLMs are terrible .

              "Sovereign" and reproducible slop is still just slop 🤷

              thomasjwebb@mastodon.socialT 1 Reply Last reply
              0
              • cwebber@social.coopC cwebber@social.coop

                I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                Noooooooooo
                Noooooooooooooooooooooooooooo

                LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                And they *certainly* can't be part of a reproducible pipeline

                deech@mastodon.socialD This user is from outside of this forum
                deech@mastodon.socialD This user is from outside of this forum
                deech@mastodon.social
                wrote sidst redigeret af
                #49

                @cwebber I think we can compromise and call them really shitty compilers.

                1 Reply Last reply
                0
                • cwebber@social.coopC cwebber@social.coop

                  I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                  Noooooooooo
                  Noooooooooooooooooooooooooooo

                  LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                  And they *certainly* can't be part of a reproducible pipeline

                  alienghic@timeloop.cafeA This user is from outside of this forum
                  alienghic@timeloop.cafeA This user is from outside of this forum
                  alienghic@timeloop.cafe
                  wrote sidst redigeret af
                  #50

                  @cwebber

                  I was thinking LLMs are like ouiji boards or tarot readings.

                  Semi random noise where meaning is imposed by the participating humans.

                  1 Reply Last reply
                  0
                  • cwebber@social.coopC cwebber@social.coop

                    I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                    Noooooooooo
                    Noooooooooooooooooooooooooooo

                    LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                    And they *certainly* can't be part of a reproducible pipeline

                    sherwoodinc@floss.socialS This user is from outside of this forum
                    sherwoodinc@floss.socialS This user is from outside of this forum
                    sherwoodinc@floss.social
                    wrote sidst redigeret af
                    #51

                    @cwebber gamified transpilers at best

                    1 Reply Last reply
                    0
                    • mcc@mastodon.socialM mcc@mastodon.social

                      @mntmn @cwebber I think the single interesting thing LLMs have revealed is that there is a substantial market segment who has an active desire for natural language interfaces to the computer and who will flip from "do not engage to the computer" to "engage with the computer" if a natural language interface became available.

                      I do not personally want a natural language interface to the computer. I also do not believe the thing LLM vendors have built is a natural language interface to the computer

                      foolishowl@social.coopF This user is from outside of this forum
                      foolishowl@social.coopF This user is from outside of this forum
                      foolishowl@social.coop
                      wrote sidst redigeret af
                      #52

                      @mcc @mntmn @cwebber Do you remember AskJeeves? A friend of mine worked for them, and told me that their whole thing had been natural language Web searches, but after a few years, their internal research showed that almost all their users were doing searches for literal text, or literal text connected with Boolean operators, just the way they used the other search engines. It wasn't that "natural language search" didn't work, it's that no one wanted to use it.

                      When I was looking up how to disable Google Assistant on my phone, a few of the articles I read opened with some claim that it was the primary reason to use an Android phone to begin with. But outside TV shows, I've rarely heard anyone trying to use it.

                      Corporations were trying to market GUI desktops for the Commodore 64.

                      I'm suspicious that there's really that much demand for natural language interfaces and skeumorphism. We've been using tools for two million years that usually don't much resemble the human body.

                      1 Reply Last reply
                      0
                      • cwebber@social.coopC cwebber@social.coop

                        I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                        Noooooooooo
                        Noooooooooooooooooooooooooooo

                        LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                        And they *certainly* can't be part of a reproducible pipeline

                        grumble209@kolektiva.socialG This user is from outside of this forum
                        grumble209@kolektiva.socialG This user is from outside of this forum
                        grumble209@kolektiva.social
                        wrote sidst redigeret af
                        #53

                        @cwebber "LLMs are compilers for prompts" says a lot more about someone's ignorance about compilers than about their knowledge of LLMs.

                        It's so stupid, it's almost wearing coconut shells on your ears and yelling into a stick and hoping that pallets of food start falling from the sky.

                        1 Reply Last reply
                        0
                        • ansuz@gts.cryptography.dogA ansuz@gts.cryptography.dog

                          @thomasjwebb @cwebber @joeyh 💯​

                          Local models like llama could be reworked to accept a seed for their RNG. There'd be less risk of them becoming unavailable, and they'd be both deterministic and reproducible, but they'd still be terrible for all the other reasons that LLMs are terrible .

                          "Sovereign" and reproducible slop is still just slop 🤷

                          thomasjwebb@mastodon.socialT This user is from outside of this forum
                          thomasjwebb@mastodon.socialT This user is from outside of this forum
                          thomasjwebb@mastodon.social
                          wrote sidst redigeret af
                          #54

                          @ansuz @cwebber @joeyh I do want to play around with llama but that goes so against my instincts of always trying to make development put less strain on my computer (like I really hated how it feels vscode really bloated up). And while yeah, having the model and source code is certainly an improvement, my experience with getting AI/GPU stuff from the past up and running again is... not fun. Having to resurrect a 10 year old version of a model would definitely suck.

                          1 Reply Last reply
                          0
                          • cwebber@social.coopC cwebber@social.coop

                            I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                            Noooooooooo
                            Noooooooooooooooooooooooooooo

                            LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                            And they *certainly* can't be part of a reproducible pipeline

                            murodegrizeco@toad.socialM This user is from outside of this forum
                            murodegrizeco@toad.socialM This user is from outside of this forum
                            murodegrizeco@toad.social
                            wrote sidst redigeret af
                            #55

                            @cwebber

                            LLM maybe may be "dissembler".

                            1 Reply Last reply
                            0
                            • cwebber@social.coopC cwebber@social.coop

                              I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                              Noooooooooo
                              Noooooooooooooooooooooooooooo

                              LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                              And they *certainly* can't be part of a reproducible pipeline

                              m_22@universeodon.comM This user is from outside of this forum
                              m_22@universeodon.comM This user is from outside of this forum
                              m_22@universeodon.com
                              wrote sidst redigeret af
                              #56

                              @cwebber the methods used to prepare the data are similar (preprocessing, encoding, tokenization). If you turned the temperature on an LLM to 0 then it can be used to deterministically output the word with the highest probability at every step. People aren’t talking about that in this case, though.

                              m_22@universeodon.comM 1 Reply Last reply
                              0
                              • m_22@universeodon.comM m_22@universeodon.com

                                @cwebber the methods used to prepare the data are similar (preprocessing, encoding, tokenization). If you turned the temperature on an LLM to 0 then it can be used to deterministically output the word with the highest probability at every step. People aren’t talking about that in this case, though.

                                m_22@universeodon.comM This user is from outside of this forum
                                m_22@universeodon.comM This user is from outside of this forum
                                m_22@universeodon.com
                                wrote sidst redigeret af
                                #57

                                @cwebber even if it was set to be deterministic, it still wouldn’t reliably produce correct output.

                                1 Reply Last reply
                                0
                                • cwebber@social.coopC cwebber@social.coop

                                  I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                                  Noooooooooo
                                  Noooooooooooooooooooooooooooo

                                  LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                                  And they *certainly* can't be part of a reproducible pipeline

                                  bit101@mstdn.socialB This user is from outside of this forum
                                  bit101@mstdn.socialB This user is from outside of this forum
                                  bit101@mstdn.social
                                  wrote sidst redigeret af
                                  #58

                                  @cwebber I've seen this basic message about non-determinism at least 3 times in the past week. I'm glad to see it more. It's an important point.

                                  1 Reply Last reply
                                  0
                                  • cwebber@social.coopC cwebber@social.coop

                                    I keep seeing lots of people saying "LLMs are like compilers/assemblers for prompts"

                                    Noooooooooo
                                    Noooooooooooooooooooooooooooo

                                    LLMs are not compilers, and they're not assemblers. Determinism is a key aspect to assemblers and compilers.

                                    And they *certainly* can't be part of a reproducible pipeline

                                    patrick_h_lauke@mastodon.socialP This user is from outside of this forum
                                    patrick_h_lauke@mastodon.socialP This user is from outside of this forum
                                    patrick_h_lauke@mastodon.social
                                    wrote sidst redigeret af
                                    #59

                                    @cwebber hah, had similar reaction to exactly that misguided point a few weeks ago https://mastodon.social/@patrick_h_lauke/116023804209257572

                                    1 Reply Last reply
                                    0
                                    • pelle@veganism.socialP pelle@veganism.social shared this topic
                                    Svar
                                    • Svar som emne
                                    Login for at svare
                                    • Ældste til nyeste
                                    • Nyeste til ældste
                                    • Most Votes


                                    • Log ind

                                    • Har du ikke en konto? Tilmeld

                                    • Login or register to search.
                                    Powered by NodeBB Contributors
                                    Graciously hosted by data.coop
                                    • First post
                                      Last post
                                    0
                                    • Hjem
                                    • Seneste
                                    • Etiketter
                                    • Populære
                                    • Verden
                                    • Bruger
                                    • Grupper