Mastodon Skip to content
  • Home
  • Aktuell
  • Tags
  • Über dieses Forum
Einklappen
Grafik mit zwei überlappenden Sprechblasen, eine grün und eine lila.
Abspeckgeflüster – Forum für Menschen mit Gewicht(ung)

Kostenlos. Werbefrei. Menschlich. Dein Abnehmforum.

  1. Home
  2. Uncategorized
  3. Problem: LLMs can't defend against prompt injection.

Problem: LLMs can't defend against prompt injection.

Geplant Angeheftet Gesperrt Verschoben Uncategorized
19 Beiträge 16 Kommentatoren 0 Aufrufe
  • Älteste zuerst
  • Neuste zuerst
  • Meiste Stimmen
Antworten
  • In einem neuen Thema antworten
Anmelden zum Antworten
Dieses Thema wurde gelöscht. Nur Nutzer mit entsprechenden Rechten können es sehen.
  • mttaggart@infosec.exchangeM This user is from outside of this forum
    mttaggart@infosec.exchangeM This user is from outside of this forum
    mttaggart@infosec.exchange
    schrieb zuletzt editiert von
    #1

    Problem: LLMs can't defend against prompt injection.

    Solution: A specialized filtering model that detects prompt injections.

    Problem: That too is susceptible to bypass and prompt injection.

    Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

    Problem: If you over-specialize, the LLM won't understand the instructions.

    Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

    Problem: We just reinvented the CLI.

    ai6yr@m.ai6yr.orgA cr0w@infosec.exchangeC mttaggart@infosec.exchangeM bluegreenandfree@mastodon.energyB proton_xor@infosec.exchangeP 11 Antworten Letzte Antwort
    1
    0
    • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

      Problem: LLMs can't defend against prompt injection.

      Solution: A specialized filtering model that detects prompt injections.

      Problem: That too is susceptible to bypass and prompt injection.

      Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

      Problem: If you over-specialize, the LLM won't understand the instructions.

      Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

      Problem: We just reinvented the CLI.

      ai6yr@m.ai6yr.orgA This user is from outside of this forum
      ai6yr@m.ai6yr.orgA This user is from outside of this forum
      ai6yr@m.ai6yr.org
      schrieb zuletzt editiert von
      #2

      @mttaggart 😂

      1 Antwort Letzte Antwort
      0
      • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

        Problem: LLMs can't defend against prompt injection.

        Solution: A specialized filtering model that detects prompt injections.

        Problem: That too is susceptible to bypass and prompt injection.

        Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

        Problem: If you over-specialize, the LLM won't understand the instructions.

        Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

        Problem: We just reinvented the CLI.

        cr0w@infosec.exchangeC This user is from outside of this forum
        cr0w@infosec.exchangeC This user is from outside of this forum
        cr0w@infosec.exchange
        schrieb zuletzt editiert von
        #3

        @mttaggart It'll never work. Unless you allow it to connect to the Internet.

        mttaggart@infosec.exchangeM 1 Antwort Letzte Antwort
        0
        • cr0w@infosec.exchangeC cr0w@infosec.exchange

          @mttaggart It'll never work. Unless you allow it to connect to the Internet.

          mttaggart@infosec.exchangeM This user is from outside of this forum
          mttaggart@infosec.exchangeM This user is from outside of this forum
          mttaggart@infosec.exchange
          schrieb zuletzt editiert von
          #4

          @cR0w That's really where all the troubles began, isn't it

          cr0w@infosec.exchangeC 1 Antwort Letzte Antwort
          0
          • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

            @cR0w That's really where all the troubles began, isn't it

            cr0w@infosec.exchangeC This user is from outside of this forum
            cr0w@infosec.exchangeC This user is from outside of this forum
            cr0w@infosec.exchange
            schrieb zuletzt editiert von
            #5

            @mttaggart The Internet was a mistake.

            1 Antwort Letzte Antwort
            0
            • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

              Problem: LLMs can't defend against prompt injection.

              Solution: A specialized filtering model that detects prompt injections.

              Problem: That too is susceptible to bypass and prompt injection.

              Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

              Problem: If you over-specialize, the LLM won't understand the instructions.

              Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

              Problem: We just reinvented the CLI.

              mttaggart@infosec.exchangeM This user is from outside of this forum
              mttaggart@infosec.exchangeM This user is from outside of this forum
              mttaggart@infosec.exchange
              schrieb zuletzt editiert von
              #6

              What are we doing with our time on this earth

              https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
              https://www.varonis.com/blog/reprompt

              theorangetheme@en.osm.townT enobacon@urbanists.socialE thechris@norden.socialT 3 Antworten Letzte Antwort
              0
              • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                Problem: LLMs can't defend against prompt injection.

                Solution: A specialized filtering model that detects prompt injections.

                Problem: That too is susceptible to bypass and prompt injection.

                Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                Problem: If you over-specialize, the LLM won't understand the instructions.

                Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                Problem: We just reinvented the CLI.

                bluegreenandfree@mastodon.energyB This user is from outside of this forum
                bluegreenandfree@mastodon.energyB This user is from outside of this forum
                bluegreenandfree@mastodon.energy
                schrieb zuletzt editiert von
                #7

                @mttaggart That will make sure nobody uses it! Problem solved.

                1 Antwort Letzte Antwort
                0
                • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                  What are we doing with our time on this earth

                  https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
                  https://www.varonis.com/blog/reprompt

                  theorangetheme@en.osm.townT This user is from outside of this forum
                  theorangetheme@en.osm.townT This user is from outside of this forum
                  theorangetheme@en.osm.town
                  schrieb zuletzt editiert von
                  #8

                  @mttaggart Man, I'm old enough to remember when computers did exactly what you told them to do, and you didn't have to grovel. (You *did* have to learn something like C or BASIC, but that's more like consensual masochism.)

                  1 Antwort Letzte Antwort
                  0
                  • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                    Problem: LLMs can't defend against prompt injection.

                    Solution: A specialized filtering model that detects prompt injections.

                    Problem: That too is susceptible to bypass and prompt injection.

                    Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                    Problem: If you over-specialize, the LLM won't understand the instructions.

                    Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                    Problem: We just reinvented the CLI.

                    proton_xor@infosec.exchangeP This user is from outside of this forum
                    proton_xor@infosec.exchangeP This user is from outside of this forum
                    proton_xor@infosec.exchange
                    schrieb zuletzt editiert von
                    #9

                    @mttaggart muahahaha

                    1 Antwort Letzte Antwort
                    0
                    • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                      What are we doing with our time on this earth

                      https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
                      https://www.varonis.com/blog/reprompt

                      enobacon@urbanists.socialE This user is from outside of this forum
                      enobacon@urbanists.socialE This user is from outside of this forum
                      enobacon@urbanists.social
                      schrieb zuletzt editiert von
                      #10

                      @mttaggart it's 2026 and we're still doing "users don't know what a text editor is" exploits, now powered by like 100x more fossil fuels 🤷

                      1 Antwort Letzte Antwort
                      0
                      • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                        Problem: LLMs can't defend against prompt injection.

                        Solution: A specialized filtering model that detects prompt injections.

                        Problem: That too is susceptible to bypass and prompt injection.

                        Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                        Problem: If you over-specialize, the LLM won't understand the instructions.

                        Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                        Problem: We just reinvented the CLI.

                        valthonis@dice.campV This user is from outside of this forum
                        valthonis@dice.campV This user is from outside of this forum
                        valthonis@dice.camp
                        schrieb zuletzt editiert von
                        #11

                        @mttaggart Almost like treating *any* technology as though it were a panacea leads to a lot of wasted effort duplicating what's already possible But Now WIth New.

                        msbellows@c.imM 1 Antwort Letzte Antwort
                        0
                        • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                          What are we doing with our time on this earth

                          https://www.promptarmor.com/resources/claude-cowork-exfiltrates-files
                          https://www.varonis.com/blog/reprompt

                          thechris@norden.socialT This user is from outside of this forum
                          thechris@norden.socialT This user is from outside of this forum
                          thechris@norden.social
                          schrieb zuletzt editiert von
                          #12

                          @mttaggart Somebody needs to create a big Excel file and write down all the measurable costs this "AI" bullshit has created, and compare that to the actually measurable and objective positive effects (OpenAI being valued at one trillion dollars does NOT count).

                          I have a feeling the net result will be a very big number with a very big minus sign in front.

                          And then make a list of things we as a society could have done with that.

                          fcalva@cyberplace.socialF 1 Antwort Letzte Antwort
                          0
                          • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                            Problem: LLMs can't defend against prompt injection.

                            Solution: A specialized filtering model that detects prompt injections.

                            Problem: That too is susceptible to bypass and prompt injection.

                            Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                            Problem: If you over-specialize, the LLM won't understand the instructions.

                            Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                            Problem: We just reinvented the CLI.

                            huynhjl@infosec.exchangeH This user is from outside of this forum
                            huynhjl@infosec.exchangeH This user is from outside of this forum
                            huynhjl@infosec.exchange
                            schrieb zuletzt editiert von
                            #13

                            @mttaggart Keeping data and code separate, that is good secure design. The thing with an LLM is that there is no code. It's all data. You give it data, and it gives you back more data. You give it code? It turns out into data and give you more data. You can't turn it back into code. I don't think you can fix that.

                            1 Antwort Letzte Antwort
                            0
                            • valthonis@dice.campV valthonis@dice.camp

                              @mttaggart Almost like treating *any* technology as though it were a panacea leads to a lot of wasted effort duplicating what's already possible But Now WIth New.

                              msbellows@c.imM This user is from outside of this forum
                              msbellows@c.imM This user is from outside of this forum
                              msbellows@c.im
                              schrieb zuletzt editiert von
                              #14

                              @valthonis @mttaggart As my father used to say: newer is only better if it's better.

                              1 Antwort Letzte Antwort
                              0
                              • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                                Problem: LLMs can't defend against prompt injection.

                                Solution: A specialized filtering model that detects prompt injections.

                                Problem: That too is susceptible to bypass and prompt injection.

                                Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                                Problem: If you over-specialize, the LLM won't understand the instructions.

                                Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                                Problem: We just reinvented the CLI.

                                louis@theforkiverse.comL This user is from outside of this forum
                                louis@theforkiverse.comL This user is from outside of this forum
                                louis@theforkiverse.com
                                schrieb zuletzt editiert von
                                #15

                                @mttaggart this is like how billionaires keep accidentally reinventing trains

                                1 Antwort Letzte Antwort
                                0
                                • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                                  Problem: LLMs can't defend against prompt injection.

                                  Solution: A specialized filtering model that detects prompt injections.

                                  Problem: That too is susceptible to bypass and prompt injection.

                                  Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                                  Problem: If you over-specialize, the LLM won't understand the instructions.

                                  Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                                  Problem: We just reinvented the CLI.

                                  sjaveed@mastodon.socialS This user is from outside of this forum
                                  sjaveed@mastodon.socialS This user is from outside of this forum
                                  sjaveed@mastodon.social
                                  schrieb zuletzt editiert von
                                  #16

                                  @mttaggart no, no, it's more like a magic spell. But then, so is the stuff I type at the bash prompt. Come to think of it, I do feel like a sorcerer!

                                  I put on my robe and wizard hat...

                                  1 Antwort Letzte Antwort
                                  0
                                  • thechris@norden.socialT thechris@norden.social

                                    @mttaggart Somebody needs to create a big Excel file and write down all the measurable costs this "AI" bullshit has created, and compare that to the actually measurable and objective positive effects (OpenAI being valued at one trillion dollars does NOT count).

                                    I have a feeling the net result will be a very big number with a very big minus sign in front.

                                    And then make a list of things we as a society could have done with that.

                                    fcalva@cyberplace.socialF This user is from outside of this forum
                                    fcalva@cyberplace.socialF This user is from outside of this forum
                                    fcalva@cyberplace.social
                                    schrieb zuletzt editiert von
                                    #17

                                    @thechris @mttaggart Globally, i heard figures of around 50B USD. With 500B to come from the Trump ai plan. In total that's more than twice the Apollo program, adjusted for inflation.

                                    1 Antwort Letzte Antwort
                                    0
                                    • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                                      Problem: LLMs can't defend against prompt injection.

                                      Solution: A specialized filtering model that detects prompt injections.

                                      Problem: That too is susceptible to bypass and prompt injection.

                                      Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                                      Problem: If you over-specialize, the LLM won't understand the instructions.

                                      Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                                      Problem: We just reinvented the CLI.

                                      zappes@mastodon.onlineZ This user is from outside of this forum
                                      zappes@mastodon.onlineZ This user is from outside of this forum
                                      zappes@mastodon.online
                                      schrieb zuletzt editiert von
                                      #18

                                      @mttaggart When I was in college learning computer science, we were tought that some briliant British boffin had found out that a program can't really find out if another program is fine, to put it in the broadest simplified terms. I wonder if somebody bothered telling that to those AI wankers.

                                      1 Antwort Letzte Antwort
                                      0
                                      • mttaggart@infosec.exchangeM mttaggart@infosec.exchange

                                        Problem: LLMs can't defend against prompt injection.

                                        Solution: A specialized filtering model that detects prompt injections.

                                        Problem: That too is susceptible to bypass and prompt injection.

                                        Solution: We reduce the set of acceptable instructions to a more predictable space and filter out anything that doesn't match.

                                        Problem: If you over-specialize, the LLM won't understand the instructions.

                                        Solution: We define a domain-specific language in the system prompt, with all allowable commands and parameters. Anything else is ignored.

                                        Problem: We just reinvented the CLI.

                                        boggo@mastodon.gamedev.placeB This user is from outside of this forum
                                        boggo@mastodon.gamedev.placeB This user is from outside of this forum
                                        boggo@mastodon.gamedev.place
                                        schrieb zuletzt editiert von
                                        #19

                                        @mttaggart

                                        Me 10 years ago: OMG AI will be incredible, it will be like in sci-fi, I want this

                                        Now: pls, I just want to CLI

                                        1 Antwort Letzte Antwort
                                        0
                                        • jaddy@friend.enby-box.deJ jaddy@friend.enby-box.de shared this topic
                                        Antworten
                                        • In einem neuen Thema antworten
                                        Anmelden zum Antworten
                                        • Älteste zuerst
                                        • Neuste zuerst
                                        • Meiste Stimmen



                                        Copyright (c) 2025 abSpecktrum (@abspecklog@fedimonster.de)

                                        Erstellt mit Schlaflosigkeit, Kaffee, Brokkoli & ♥

                                        Impressum | Datenschutzerklärung | Nutzungsbedingungen

                                        • Anmelden

                                        • Du hast noch kein Konto? Registrieren

                                        • Anmelden oder registrieren, um zu suchen
                                        • Erster Beitrag
                                          Letzter Beitrag
                                        0
                                        • Home
                                        • Aktuell
                                        • Tags
                                        • Über dieses Forum