Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. Problem: LLMs can't defend against prompt injection.

Problem: LLMs can't defend against prompt injection.

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
9 Indlæg 6 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • mttaggart@infosec.exchangeM This user is from outside of this forum
    mttaggart@infosec.exchangeM This user is from outside of this forum
    mttaggart@infosec.exchange
    wrote sidst redigeret af
    #1

    Problem: LLMs can't defend against prompt injection.

    Solution: A specialized filtering model that detects prompt injections.

    Problem: That too is susceptible to bypass and prompt injection.

    Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

    Problem: If you over-specialize, the LLM won't understand the instructions.

    Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

    Problem: We just reinvented the CLI.

    ai6yr@m.ai6yr.orgA cr0w@infosec.exchangeC mttaggart@infosec.exchangeM bluegreenandfree@mastodon.energyB proton_xor@infosec.exchangeP 5 Replies Last reply
    1
    0
    • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

      Problem: LLMs can't defend against prompt injection.

      Solution: A specialized filtering model that detects prompt injections.

      Problem: That too is susceptible to bypass and prompt injection.

      Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

      Problem: If you over-specialize, the LLM won't understand the instructions.

      Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

      Problem: We just reinvented the CLI.

      ai6yr@m.ai6yr.orgA This user is from outside of this forum
      ai6yr@m.ai6yr.orgA This user is from outside of this forum
      ai6yr@m.ai6yr.org
      wrote sidst redigeret af
      #2

      @mttaggart 😂

      1 Reply Last reply
      0
      • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

        Problem: LLMs can't defend against prompt injection.

        Solution: A specialized filtering model that detects prompt injections.

        Problem: That too is susceptible to bypass and prompt injection.

        Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

        Problem: If you over-specialize, the LLM won't understand the instructions.

        Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

        Problem: We just reinvented the CLI.

        cr0w@infosec.exchangeC This user is from outside of this forum
        cr0w@infosec.exchangeC This user is from outside of this forum
        cr0w@infosec.exchange
        wrote sidst redigeret af
        #3

        @mttaggart It'll never work. Unless you allow it to connect to the Internet.

        mttaggart@infosec.exchangeM 1 Reply Last reply
        0
        • cr0w@infosec.exchangeC cr0w@infosec.exchange

          @mttaggart It'll never work. Unless you allow it to connect to the Internet.

          mttaggart@infosec.exchangeM This user is from outside of this forum
          mttaggart@infosec.exchangeM This user is from outside of this forum
          mttaggart@infosec.exchange
          wrote sidst redigeret af
          #4

          @cR0w That's really where all the troubles began, isn't it

          cr0w@infosec.exchangeC 1 Reply Last reply
          0
          • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

            @cR0w That's really where all the troubles began, isn't it

            cr0w@infosec.exchangeC This user is from outside of this forum
            cr0w@infosec.exchangeC This user is from outside of this forum
            cr0w@infosec.exchange
            wrote sidst redigeret af
            #5

            @mttaggart The Internet was a mistake.

            1 Reply Last reply
            0
            • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

              Problem: LLMs can't defend against prompt injection.

              Solution: A specialized filtering model that detects prompt injections.

              Problem: That too is susceptible to bypass and prompt injection.

              Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

              Problem: If you over-specialize, the LLM won't understand the instructions.

              Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

              Problem: We just reinvented the CLI.

              mttaggart@infosec.exchangeM This user is from outside of this forum
              mttaggart@infosec.exchangeM This user is from outside of this forum
              mttaggart@infosec.exchange
              wrote sidst redigeret af
              #6

              What are we doing with our time on this earth

              https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
              https://www.varonis.com/blog/reprompt

              theorangetheme@en.osm.townT 1 Reply Last reply
              0
              • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                Problem: LLMs can't defend against prompt injection.

                Solution: A specialized filtering model that detects prompt injections.

                Problem: That too is susceptible to bypass and prompt injection.

                Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                Problem: If you over-specialize, the LLM won't understand the instructions.

                Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                Problem: We just reinvented the CLI.

                bluegreenandfree@mastodon.energyB This user is from outside of this forum
                bluegreenandfree@mastodon.energyB This user is from outside of this forum
                bluegreenandfree@mastodon.energy
                wrote sidst redigeret af
                #7

                @mttaggart That will make sure nobody uses it! Problem solved.

                1 Reply Last reply
                0
                • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                  What are we doing with our time on this earth

                  https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
                  https://www.varonis.com/blog/reprompt

                  theorangetheme@en.osm.townT This user is from outside of this forum
                  theorangetheme@en.osm.townT This user is from outside of this forum
                  theorangetheme@en.osm.town
                  wrote sidst redigeret af
                  #8

                  @mttaggart Man, I'm old enough to remember when computers did exactly what you told them to do, and you didn't have to grovel. (You *did* have to learn something like C or BASIC, but that's more like consensual masochism.)

                  1 Reply Last reply
                  0
                  • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                    Problem: LLMs can't defend against prompt injection.

                    Solution: A specialized filtering model that detects prompt injections.

                    Problem: That too is susceptible to bypass and prompt injection.

                    Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                    Problem: If you over-specialize, the LLM won't understand the instructions.

                    Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                    Problem: We just reinvented the CLI.

                    proton_xor@infosec.exchangeP This user is from outside of this forum
                    proton_xor@infosec.exchangeP This user is from outside of this forum
                    proton_xor@infosec.exchange
                    wrote sidst redigeret af
                    #9

                    @mttaggart muahahaha

                    1 Reply Last reply
                    0
                    • jwcph@helvede.netJ jwcph@helvede.net shared this topic
                    Svar
                    • Svar som emne
                    Login for at svare
                    • Ældste til nyeste
                    • Nyeste til ældste
                    • Most Votes


                    • Log ind

                    • Har du ikke en konto? Tilmeld

                    • Login or register to search.
                    Powered by NodeBB Contributors
                    Graciously hosted by data.coop
                    • First post
                      Last post
                    0
                    • Hjem
                    • Seneste
                    • Etiketter
                    • Populære
                    • Verden
                    • Bruger
                    • Grupper