Mastodon Skip to content
  • Home
  • Aktuell
  • Tags
  • Über dieses Forum
Einklappen
Grafik mit zwei überlappenden Sprechblasen, eine grün und eine lila.
Abspeckgeflüster – Forum für Menschen mit Gewicht(ung)

Kostenlos. Werbefrei. Menschlich. Dein Abnehmforum.

  1. Home
  2. Uncategorized
  3. Letting AI agents run your life is like handing the car keys to your 5-year-old.

Letting AI agents run your life is like handing the car keys to your 5-year-old.

Geplant Angeheftet Gesperrt Verschoben Uncategorized
24 Beiträge 20 Kommentatoren 0 Aufrufe
  • Älteste zuerst
  • Neuste zuerst
  • Meiste Stimmen
Antworten
  • In einem neuen Thema antworten
Anmelden zum Antworten
Dieses Thema wurde gelöscht. Nur Nutzer mit entsprechenden Rechten können es sehen.
  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

    royal@theres.lifeR This user is from outside of this forum
    royal@theres.lifeR This user is from outside of this forum
    royal@theres.life
    schrieb zuletzt editiert von
    #2

    @briankrebs In the future, programming will consist of setting loose a random action generator and defining its behavior by a series of prohibitions.

    jonas@social.jonaskoeritz.deJ 1 Antwort Letzte Antwort
    0
    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

      tknarr@mstdn.socialT This user is from outside of this forum
      tknarr@mstdn.socialT This user is from outside of this forum
      tknarr@mstdn.social
      schrieb zuletzt editiert von
      #3

      @briankrebs Sounds like "Do not look into laser with remaining eye."

      1 Antwort Letzte Antwort
      0
      • royal@theres.lifeR royal@theres.life

        @briankrebs In the future, programming will consist of setting loose a random action generator and defining its behavior by a series of prohibitions.

        jonas@social.jonaskoeritz.deJ This user is from outside of this forum
        jonas@social.jonaskoeritz.deJ This user is from outside of this forum
        jonas@social.jonaskoeritz.de
        schrieb zuletzt editiert von
        #4

        @royal @briankrebs I faintly remember "Chaos Monkey" being a legit tool that Netflix (?) used internally to just cause random outages in their systems to build resilience. This reminds me of that.

        royal@theres.lifeR 1 Antwort Letzte Antwort
        0
        • jonas@social.jonaskoeritz.deJ jonas@social.jonaskoeritz.de

          @royal @briankrebs I faintly remember "Chaos Monkey" being a legit tool that Netflix (?) used internally to just cause random outages in their systems to build resilience. This reminds me of that.

          royal@theres.lifeR This user is from outside of this forum
          royal@theres.lifeR This user is from outside of this forum
          royal@theres.life
          schrieb zuletzt editiert von
          #5

          @jonas @briankrebs Yes, it was Netflix, and I was thinking of something similar.

          1 Antwort Letzte Antwort
          0
          • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

            Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

            I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

            "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

            https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

            noplasticshower@infosec.exchangeN This user is from outside of this forum
            noplasticshower@infosec.exchangeN This user is from outside of this forum
            noplasticshower@infosec.exchange
            schrieb zuletzt editiert von
            #6

            @briankrebs old style thinking about new stuff

            If you start working on this ping me and we can talk

            briankrebs@infosec.exchangeB 1 Antwort Letzte Antwort
            0
            • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

              Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

              I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

              "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

              https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

              womble@infosec.exchangeW This user is from outside of this forum
              womble@infosec.exchangeW This user is from outside of this forum
              womble@infosec.exchange
              schrieb zuletzt editiert von
              #7

              @briankrebs if there's one thing that the entire history of humanity has taught us, it's that as a species we are great at making nuanced, highly context-specific decisions with incomplete information. I foresee that this is going to turn out great.

              1 Antwort Letzte Antwort
              0
              • noplasticshower@infosec.exchangeN noplasticshower@infosec.exchange

                @briankrebs old style thinking about new stuff

                If you start working on this ping me and we can talk

                briankrebs@infosec.exchangeB This user is from outside of this forum
                briankrebs@infosec.exchangeB This user is from outside of this forum
                briankrebs@infosec.exchange
                schrieb zuletzt editiert von
                #8

                @noplasticshower Are you letting AI agents manage your life?

                noplasticshower@infosec.exchangeN 1 Antwort Letzte Antwort
                0
                • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                  Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                  I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                  "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                  https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                  hellpie@raru.reH This user is from outside of this forum
                  hellpie@raru.reH This user is from outside of this forum
                  hellpie@raru.re
                  schrieb zuletzt editiert von
                  #9

                  @briankrebs I am going to start saying some stuff like "back in my day APTs didn't forget shit all the time" and "when spyware was made the old fashioned way it just worked without needing to ask you anything"

                  1 Antwort Letzte Antwort
                  0
                  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                    leahprice@hcommons.socialL This user is from outside of this forum
                    leahprice@hcommons.socialL This user is from outside of this forum
                    leahprice@hcommons.social
                    schrieb zuletzt editiert von
                    #10

                    @briankrebs Makes you nostalgic for 1860, when machine-generated text meant a new magazine. (from PUNCH)

                    1 Antwort Letzte Antwort
                    0
                    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                      slyborg@vmst.ioS This user is from outside of this forum
                      slyborg@vmst.ioS This user is from outside of this forum
                      slyborg@vmst.io
                      schrieb zuletzt editiert von
                      #11

                      @briankrebs Can’t wait for the slew of stories / TikTok shorts about ‘how my lifehack AI agent phone buddy converted my life savings to Monero and sent it to North Korea’ and that nobody will learn a thing from them.

                      1 Antwort Letzte Antwort
                      0
                      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                        @noplasticshower Are you letting AI agents manage your life?

                        noplasticshower@infosec.exchangeN This user is from outside of this forum
                        noplasticshower@infosec.exchangeN This user is from outside of this forum
                        noplasticshower@infosec.exchange
                        schrieb zuletzt editiert von
                        #12

                        @briankrebs of course!

                        The "let's secure one (or maybe 5) agent(s) at a time" security thing is cute. If I read about another A&A framework approach to this stuff I am going to start using agents to run my life.

                        toriver@mas.toT 1 Antwort Letzte Antwort
                        0
                        • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                          Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                          I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                          "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                          https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                          zl2tod@mastodon.onlineZ This user is from outside of this forum
                          zl2tod@mastodon.onlineZ This user is from outside of this forum
                          zl2tod@mastodon.online
                          schrieb zuletzt editiert von
                          #13

                          @briankrebs

                          Thread re #clawdbot :

                          https://infosec.exchange/@munin/115963686278347109

                          1 Antwort Letzte Antwort
                          0
                          • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                            Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                            I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                            "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                            https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                            C This user is from outside of this forum
                            C This user is from outside of this forum
                            crouton@aus.social
                            schrieb zuletzt editiert von
                            #14

                            @briankrebs This sounds like the optimally worst implementation of a digital assistant. I was looking for a distant variant of this, where I set the (mostly deterministic) rules and actions, etc. Kinda like HA/node-red but aimed at being an assistant rather than controlling a house.
                            Giving it a blank cheque and hooking up to a LLM is insane.

                            1 Antwort Letzte Antwort
                            0
                            • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                              Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                              I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                              "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                              https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                              hal8999@infosec.exchangeH This user is from outside of this forum
                              hal8999@infosec.exchangeH This user is from outside of this forum
                              hal8999@infosec.exchange
                              schrieb zuletzt editiert von
                              #15

                              @briankrebs The stories are writing themselves: My A.I. agent took over my finances, framed me for sex trafficking, then unlocked the doors and turned on lights for the police.

                              1 Antwort Letzte Antwort
                              0
                              • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                carsten@minnesotasocial.netC This user is from outside of this forum
                                carsten@minnesotasocial.netC This user is from outside of this forum
                                carsten@minnesotasocial.net
                                schrieb zuletzt editiert von
                                #16

                                @briankrebs and if the bot "touches" something it was not allowed to touch? "Sorry, me bad, won't do it again, probably" - does it again.

                                1 Antwort Letzte Antwort
                                0
                                • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                  Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                  I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                  "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                  https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                  zompetto@mastodon.artZ This user is from outside of this forum
                                  zompetto@mastodon.artZ This user is from outside of this forum
                                  zompetto@mastodon.art
                                  schrieb zuletzt editiert von
                                  #17

                                  @briankrebs it would be funny if it wasn't so sad. And scary.

                                  1 Antwort Letzte Antwort
                                  0
                                  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                    theorangetheme@en.osm.townT This user is from outside of this forum
                                    theorangetheme@en.osm.townT This user is from outside of this forum
                                    theorangetheme@en.osm.town
                                    schrieb zuletzt editiert von
                                    #18

                                    @briankrebs It's fucking bonkerstown. You might as well just wire up a random number generator and every 1/10 times it just deletes your home directory. At least that achieves the same result but with a fraction of the electricity and human rights abuses.

                                    1 Antwort Letzte Antwort
                                    0
                                    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                      S This user is from outside of this forum
                                      S This user is from outside of this forum
                                      spacelifeform@infosec.exchange
                                      schrieb zuletzt editiert von
                                      #19

                                      @briankrebs

                                      So sayeth the bot.

                                      It is smarter not to talk to them in the first place.

                                      1 Antwort Letzte Antwort
                                      0
                                      • noplasticshower@infosec.exchangeN noplasticshower@infosec.exchange

                                        @briankrebs of course!

                                        The "let's secure one (or maybe 5) agent(s) at a time" security thing is cute. If I read about another A&A framework approach to this stuff I am going to start using agents to run my life.

                                        toriver@mas.toT This user is from outside of this forum
                                        toriver@mas.toT This user is from outside of this forum
                                        toriver@mas.to
                                        schrieb zuletzt editiert von
                                        #20

                                        @noplasticshower @briankrebs
                                        1) Configure agent with guardrails
                                        2) Agent runs into guardrails
                                        3) Agent spins up secondary agent without guardrails
                                        4) Oh no.

                                        noplasticshower@infosec.exchangeN 1 Antwort Letzte Antwort
                                        0
                                        • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                          Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                          I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                          "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                          https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                          craignicol@glasgow.socialC This user is from outside of this forum
                                          craignicol@glasgow.socialC This user is from outside of this forum
                                          craignicol@glasgow.social
                                          schrieb zuletzt editiert von
                                          #21

                                          @briankrebs ok child, you can have the scissors, but only if you promise not to run with them

                                          1 Antwort Letzte Antwort
                                          0
                                          Antworten
                                          • In einem neuen Thema antworten
                                          Anmelden zum Antworten
                                          • Älteste zuerst
                                          • Neuste zuerst
                                          • Meiste Stimmen



                                          Copyright (c) 2025 abSpecktrum (@abspecklog@fedimonster.de)

                                          Erstellt mit Schlaflosigkeit, Kaffee, Brokkoli & ♥

                                          Impressum | Datenschutzerklärung | Nutzungsbedingungen

                                          • Anmelden

                                          • Du hast noch kein Konto? Registrieren

                                          • Anmelden oder registrieren, um zu suchen
                                          • Erster Beitrag
                                            Letzter Beitrag
                                          0
                                          • Home
                                          • Aktuell
                                          • Tags
                                          • Über dieses Forum