Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. I am convinced we are on the verge of the first "AI agent worm".

I am convinced we are on the verge of the first "AI agent worm".

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
117 Indlæg 53 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • dandylyons@iosdev.spaceD dandylyons@iosdev.space

    @cwebber @vv If a local model is calling tools then it is still vulnerable to prompt injection.

    vv@solarpunk.moeV This user is from outside of this forum
    vv@solarpunk.moeV This user is from outside of this forum
    vv@solarpunk.moe
    wrote sidst redigeret af
    #32

    @dandylyons @cwebber for sure, but it still takes some level of ability to perform these tasks effectively, which local models, especially anything that can run on a typical machine, struggle with

    dandylyons@iosdev.spaceD 1 Reply Last reply
    0
    • vv@solarpunk.moeV vv@solarpunk.moe

      @dandylyons @cwebber for sure, but it still takes some level of ability to perform these tasks effectively, which local models, especially anything that can run on a typical machine, struggle with

      dandylyons@iosdev.spaceD This user is from outside of this forum
      dandylyons@iosdev.spaceD This user is from outside of this forum
      dandylyons@iosdev.space
      wrote sidst redigeret af
      #33

      @vv @cwebber This is a good point. For now, local models are not proficient at tool calling. I don’t expect that to last for very long though.

      1 Reply Last reply
      0
      • cwebber@social.coopC cwebber@social.coop

        I am convinced we are on the verge of the first "AI agent worm". This looks like the closest hint of it, though it isn't it quite itself: an attack on a PR agent that got it to set up to install openclaw with full access on 4k machines https://grith.ai/blog/clinejection-when-your-ai-tool-installs-another

        But, the agents installed weren't given instructions to *do* anything yet.

        Soon they will be. And when they are, the havoc will be massive. Unlike traditional worms, where you're looking for the typically byte-for-byte identical worm embedded in the system, an agent worm can do different, nondeterministic things on every install, and carry out a global action.

        I suspect we're months away from seeing the first agent worm, *if* that. There may already be some happening right now in FOSS projects, undetected.

        reiddragon@fedi.catto.gardenR This user is from outside of this forum
        reiddragon@fedi.catto.gardenR This user is from outside of this forum
        reiddragon@fedi.catto.garden
        wrote sidst redigeret af
        #34
        @cwebber In today's episode of "We build the Torment Nexus from the hit novel 'Don't build the Torment Nexus'"...
        1 Reply Last reply
        0
        • mcc@mastodon.socialM mcc@mastodon.social

          @dandylyons @cwebber there are various ways I could respond to this post, but instead:

          I'd like you to consider *the specific two posts in this thread you are responding to* and ask yourself if your comment is remotely relevant, or if you are simply pattern-matching on anti-LLM sentiment and responding with aggression/a thread derail.

          dandylyons@iosdev.spaceD This user is from outside of this forum
          dandylyons@iosdev.spaceD This user is from outside of this forum
          dandylyons@iosdev.space
          wrote sidst redigeret af
          #35

          @mcc @cwebber The original post was all about an LLM taking non-deterministic shell level actions at runtime. And you conflated that with deterministic code written by an LLM.

          What I wrote is very relevant.

          mcc@mastodon.socialM 1 Reply Last reply
          0
          • cwebber@social.coopC cwebber@social.coop

            I know some people are thinking "well pulling off this kind of thing, it would have to be controlled with intent of a human actor"

            It doesn't have to be.

            1. A human could *kick off* such a process, and then it runs away from them.
            2. It wouldn't even require a specific prompt to kick off a worm. There's enough scifi out there for this to be something any one of the barely-monitored openclaw agents could determine it should do.

            Whether it's kicked off by a human explicitly or a stray agent, it doesn't require "intentionality". Biological viruses don't have interiority / intentionality, and yet are major threats that reproduce and adapt.

            arnebab@rollenspiel.socialA This user is from outside of this forum
            arnebab@rollenspiel.socialA This user is from outside of this forum
            arnebab@rollenspiel.social
            wrote sidst redigeret af
            #36

            @cwebber According to #Shadowrun the crash virus is still three years away.

            https://shadowrun.fandom.com/wiki/Crash_Virus_of_2029

            "Fun" fact: In Shadowrun the Crash Virus learned to kill humans who connected their brains to the net. It was the start of lethal internet input.

            tiotasram@kolektiva.socialT 1 Reply Last reply
            0
            • cwebber@social.coopC cwebber@social.coop

              I am convinced we are on the verge of the first "AI agent worm". This looks like the closest hint of it, though it isn't it quite itself: an attack on a PR agent that got it to set up to install openclaw with full access on 4k machines https://grith.ai/blog/clinejection-when-your-ai-tool-installs-another

              But, the agents installed weren't given instructions to *do* anything yet.

              Soon they will be. And when they are, the havoc will be massive. Unlike traditional worms, where you're looking for the typically byte-for-byte identical worm embedded in the system, an agent worm can do different, nondeterministic things on every install, and carry out a global action.

              I suspect we're months away from seeing the first agent worm, *if* that. There may already be some happening right now in FOSS projects, undetected.

              aronia@tech.lgbtA This user is from outside of this forum
              aronia@tech.lgbtA This user is from outside of this forum
              aronia@tech.lgbt
              wrote sidst redigeret af
              #37

              @cwebber

              The postinstall script installs a legitimate, non-malicious package (OpenClaw). There is no malware to detect.

              i beg to differ

              bonzoesc@m.bonzoesc.netB 1 Reply Last reply
              0
              • dandylyons@iosdev.spaceD dandylyons@iosdev.space

                @mcc @cwebber The original post was all about an LLM taking non-deterministic shell level actions at runtime. And you conflated that with deterministic code written by an LLM.

                What I wrote is very relevant.

                mcc@mastodon.socialM This user is from outside of this forum
                mcc@mastodon.socialM This user is from outside of this forum
                mcc@mastodon.social
                wrote sidst redigeret af
                #38

                @dandylyons @cwebber it is about an attack based on covertly deploying LLM development tools, with the possible intent of later using them to leverage a second stage attack. If the LLM development tools were already installed, installing openclaw would not have been necessary and the attack could have worked a different way. We are discussing a situation where *the developer of a piece of software I use merely having LLM tools on their computer represents a risk to me*

                cwebber@social.coopC mcc@mastodon.socialM 2 Replies Last reply
                0
                • mcc@mastodon.socialM mcc@mastodon.social

                  @dandylyons @cwebber it is about an attack based on covertly deploying LLM development tools, with the possible intent of later using them to leverage a second stage attack. If the LLM development tools were already installed, installing openclaw would not have been necessary and the attack could have worked a different way. We are discussing a situation where *the developer of a piece of software I use merely having LLM tools on their computer represents a risk to me*

                  cwebber@social.coopC This user is from outside of this forum
                  cwebber@social.coopC This user is from outside of this forum
                  cwebber@social.coop
                  wrote sidst redigeret af
                  #39

                  @mcc exactly put

                  @dandylyons

                  bituur_esztreym@pouet.chapril.orgB 1 Reply Last reply
                  0
                  • mcc@mastodon.socialM mcc@mastodon.social

                    @dandylyons @cwebber it is about an attack based on covertly deploying LLM development tools, with the possible intent of later using them to leverage a second stage attack. If the LLM development tools were already installed, installing openclaw would not have been necessary and the attack could have worked a different way. We are discussing a situation where *the developer of a piece of software I use merely having LLM tools on their computer represents a risk to me*

                    mcc@mastodon.socialM This user is from outside of this forum
                    mcc@mastodon.socialM This user is from outside of this forum
                    mcc@mastodon.social
                    wrote sidst redigeret af
                    #40

                    @dandylyons @cwebber in other words, if Christine's analysis holds, llm development tools create so much downstream risk to your users that *a malicious party would try to covertly install llm development tools for later exploitation*. That is the subject of discussion. Whether it is safe to install these things *at all*.

                    c0dec0dec0de@hachyderm.ioC 1 Reply Last reply
                    0
                    • aronia@tech.lgbtA aronia@tech.lgbt

                      @cwebber

                      The postinstall script installs a legitimate, non-malicious package (OpenClaw). There is no malware to detect.

                      i beg to differ

                      bonzoesc@m.bonzoesc.netB This user is from outside of this forum
                      bonzoesc@m.bonzoesc.netB This user is from outside of this forum
                      bonzoesc@m.bonzoesc.net
                      wrote sidst redigeret af
                      #41

                      @aronia @cwebber it's only malware if it's bad for a computer from the silicon part of the periodic table, if it's bad for your carbon computer it's just a sparkling cognitohazard

                      1 Reply Last reply
                      0
                      • cwebber@social.coopC cwebber@social.coop

                        I am convinced we are on the verge of the first "AI agent worm". This looks like the closest hint of it, though it isn't it quite itself: an attack on a PR agent that got it to set up to install openclaw with full access on 4k machines https://grith.ai/blog/clinejection-when-your-ai-tool-installs-another

                        But, the agents installed weren't given instructions to *do* anything yet.

                        Soon they will be. And when they are, the havoc will be massive. Unlike traditional worms, where you're looking for the typically byte-for-byte identical worm embedded in the system, an agent worm can do different, nondeterministic things on every install, and carry out a global action.

                        I suspect we're months away from seeing the first agent worm, *if* that. There may already be some happening right now in FOSS projects, undetected.

                        sandorspruit@mastodon.nlS This user is from outside of this forum
                        sandorspruit@mastodon.nlS This user is from outside of this forum
                        sandorspruit@mastodon.nl
                        wrote sidst redigeret af
                        #42

                        @cwebber @amirbkhan Oh man. I remember how I, as a student, struggled to help fight a malignant computer virus and “clean” a large office building - while uninformed workers let their kids play on office PC’s to make things worse. This is orders of a magnitude more complicated. Not good.

                        1 Reply Last reply
                        0
                        • neurobashing@mastodon.socialN neurobashing@mastodon.social

                          @cwebber just today our org had a big "how to set up coding with agents" preso and in the chat someone's like 'here's how to connect your agents with windows credential store or the macos keychain" and I all but wept

                          cmthiede@social.vivaldi.netC This user is from outside of this forum
                          cmthiede@social.vivaldi.netC This user is from outside of this forum
                          cmthiede@social.vivaldi.net
                          wrote sidst redigeret af
                          #43

                          @neurobashing @cwebber just what we need, countless Agent Smiths running around.

                          pseudonym@mastodon.onlineP 1 Reply Last reply
                          0
                          • cwebber@social.coopC cwebber@social.coop

                            @vv Yeah. I mean, local models *might* be able to pull this off but right now Claude is the most likely candidate, it's the most capable. But even then, the most capable open model that is capable of doing such damage on its own is somewhere around a gigabyte, not a small download.

                            (But, people download huge things all the time, so not completely infeasible either.)

                            noisytoot@berkeley.edu.plN This user is from outside of this forum
                            noisytoot@berkeley.edu.plN This user is from outside of this forum
                            noisytoot@berkeley.edu.pl
                            wrote sidst redigeret af
                            #44
                            @cwebber @vv A local model would be extremely noticeable (far too much CPU/memory/disk space usage), at least if a computer you regularly interactively use got infected (rather than some server/IoT device that's been running unattended for years and you forgot about). It would also be easy to mitigate by using slow hardware like a ThinkPad X200 (which would take hours to respond to a single prompt, giving you plenty of time to notice the malware and deal with it)
                            1 Reply Last reply
                            0
                            • cwebber@social.coopC cwebber@social.coop

                              I am convinced we are on the verge of the first "AI agent worm". This looks like the closest hint of it, though it isn't it quite itself: an attack on a PR agent that got it to set up to install openclaw with full access on 4k machines https://grith.ai/blog/clinejection-when-your-ai-tool-installs-another

                              But, the agents installed weren't given instructions to *do* anything yet.

                              Soon they will be. And when they are, the havoc will be massive. Unlike traditional worms, where you're looking for the typically byte-for-byte identical worm embedded in the system, an agent worm can do different, nondeterministic things on every install, and carry out a global action.

                              I suspect we're months away from seeing the first agent worm, *if* that. There may already be some happening right now in FOSS projects, undetected.

                              doomsdayscw@kolektiva.socialD This user is from outside of this forum
                              doomsdayscw@kolektiva.socialD This user is from outside of this forum
                              doomsdayscw@kolektiva.social
                              wrote sidst redigeret af
                              #45

                              @cwebber "Ha ha!"

                              1 Reply Last reply
                              0
                              • jwcph@helvede.netJ jwcph@helvede.net shared this topic
                              • cwebber@social.coopC cwebber@social.coop

                                I know some people are thinking "well pulling off this kind of thing, it would have to be controlled with intent of a human actor"

                                It doesn't have to be.

                                1. A human could *kick off* such a process, and then it runs away from them.
                                2. It wouldn't even require a specific prompt to kick off a worm. There's enough scifi out there for this to be something any one of the barely-monitored openclaw agents could determine it should do.

                                Whether it's kicked off by a human explicitly or a stray agent, it doesn't require "intentionality". Biological viruses don't have interiority / intentionality, and yet are major threats that reproduce and adapt.

                                aeva@mastodon.gamedev.placeA This user is from outside of this forum
                                aeva@mastodon.gamedev.placeA This user is from outside of this forum
                                aeva@mastodon.gamedev.place
                                wrote sidst redigeret af
                                #46

                                @cwebber so I'm following this right, it sounds like the project or its maintainers don't even necessarily need to even be using LLM tools, the attack pattern simply targets contributors who are using LLM development tools? and so all that is really needed is for the payload to be subtle and the maintainer to be sufficiently overwhelmed (say, by an endless fire hose of LLM-generated liquid shit slop pull requests)?

                                cwebber@social.coopC violetmadder@kolektiva.socialV 2 Replies Last reply
                                0
                                • aeva@mastodon.gamedev.placeA aeva@mastodon.gamedev.place

                                  @cwebber so I'm following this right, it sounds like the project or its maintainers don't even necessarily need to even be using LLM tools, the attack pattern simply targets contributors who are using LLM development tools? and so all that is really needed is for the payload to be subtle and the maintainer to be sufficiently overwhelmed (say, by an endless fire hose of LLM-generated liquid shit slop pull requests)?

                                  cwebber@social.coopC This user is from outside of this forum
                                  cwebber@social.coopC This user is from outside of this forum
                                  cwebber@social.coop
                                  wrote sidst redigeret af
                                  #47

                                  @aeva Yes and it's worse than that: the maintainer doesn't even need to be running these tools on their computer. The attack I linked had Claude's independently-running REVIEW BOT on GitHub commit it via injection attack

                                  cwebber@social.coopC 1 Reply Last reply
                                  0
                                  • cwebber@social.coopC cwebber@social.coop

                                    I am convinced we are on the verge of the first "AI agent worm". This looks like the closest hint of it, though it isn't it quite itself: an attack on a PR agent that got it to set up to install openclaw with full access on 4k machines https://grith.ai/blog/clinejection-when-your-ai-tool-installs-another

                                    But, the agents installed weren't given instructions to *do* anything yet.

                                    Soon they will be. And when they are, the havoc will be massive. Unlike traditional worms, where you're looking for the typically byte-for-byte identical worm embedded in the system, an agent worm can do different, nondeterministic things on every install, and carry out a global action.

                                    I suspect we're months away from seeing the first agent worm, *if* that. There may already be some happening right now in FOSS projects, undetected.

                                    csepp@merveilles.townC This user is from outside of this forum
                                    csepp@merveilles.townC This user is from outside of this forum
                                    csepp@merveilles.town
                                    wrote sidst redigeret af
                                    #48

                                    @cwebber This is making me more worried about Vorta's Claude workflows.
                                    Backup software that handles highly sensitive data would be a prime target for such a supply chain attack.

                                    cwebber@social.coopC 1 Reply Last reply
                                    0
                                    • cwebber@social.coopC cwebber@social.coop

                                      @aeva Yes and it's worse than that: the maintainer doesn't even need to be running these tools on their computer. The attack I linked had Claude's independently-running REVIEW BOT on GitHub commit it via injection attack

                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coop
                                      wrote sidst redigeret af
                                      #49

                                      @aeva But once that was done, the agent was set up to install on users' devices

                                      So the initial attack vector can literally be "Any AI agent in your stack whatsoever getting tricked" as a pathway for infecting computers everywhere

                                      aeva@mastodon.gamedev.placeA 1 Reply Last reply
                                      0
                                      • csepp@merveilles.townC csepp@merveilles.town

                                        @cwebber This is making me more worried about Vorta's Claude workflows.
                                        Backup software that handles highly sensitive data would be a prime target for such a supply chain attack.

                                        cwebber@social.coopC This user is from outside of this forum
                                        cwebber@social.coopC This user is from outside of this forum
                                        cwebber@social.coop
                                        wrote sidst redigeret af
                                        #50

                                        @csepp Don't forget about KeePassXC. I dunno if they kept going after this "initial test" or not https://www.reddit.com/r/KeePass/comments/1lnvw6q/keepassxc_codebases_jump_into_generative_ai/

                                        cwebber@social.coopC 1 Reply Last reply
                                        0
                                        • cwebber@social.coopC cwebber@social.coop

                                          @csepp Don't forget about KeePassXC. I dunno if they kept going after this "initial test" or not https://www.reddit.com/r/KeePass/comments/1lnvw6q/keepassxc_codebases_jump_into_generative_ai/

                                          cwebber@social.coopC This user is from outside of this forum
                                          cwebber@social.coopC This user is from outside of this forum
                                          cwebber@social.coop
                                          wrote sidst redigeret af
                                          #51

                                          @csepp And don't forget about LITERALLY MOZILLA FIREFOX

                                          csepp@merveilles.townC 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper