Mastodon Skip to content
  • Home
  • Aktuell
  • Tags
  • Über dieses Forum
Einklappen
Grafik mit zwei überlappenden Sprechblasen, eine grün und eine lila.
Abspeckgeflüster – Forum für Menschen mit Gewicht(ung)

Kostenlos. Werbefrei. Menschlich. Dein Abnehmforum.

  1. Home
  2. Uncategorized
  3. Letting AI agents run your life is like handing the car keys to your 5-year-old.

Letting AI agents run your life is like handing the car keys to your 5-year-old.

Geplant Angeheftet Gesperrt Verschoben Uncategorized
24 Beiträge 20 Kommentatoren 0 Aufrufe
  • Älteste zuerst
  • Neuste zuerst
  • Meiste Stimmen
Antworten
  • In einem neuen Thema antworten
Anmelden zum Antworten
Dieses Thema wurde gelöscht. Nur Nutzer mit entsprechenden Rechten können es sehen.
  • briankrebs@infosec.exchangeB This user is from outside of this forum
    briankrebs@infosec.exchangeB This user is from outside of this forum
    briankrebs@infosec.exchange
    schrieb zuletzt editiert von
    #1

    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

    royal@theres.lifeR tknarr@mstdn.socialT noplasticshower@infosec.exchangeN womble@infosec.exchangeW hellpie@raru.reH 17 Antworten Letzte Antwort
    0
    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

      royal@theres.lifeR This user is from outside of this forum
      royal@theres.lifeR This user is from outside of this forum
      royal@theres.life
      schrieb zuletzt editiert von
      #2

      @briankrebs In the future, programming will consist of setting loose a random action generator and defining its behavior by a series of prohibitions.

      jonas@social.jonaskoeritz.deJ 1 Antwort Letzte Antwort
      0
      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

        Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

        I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

        "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

        https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

        tknarr@mstdn.socialT This user is from outside of this forum
        tknarr@mstdn.socialT This user is from outside of this forum
        tknarr@mstdn.social
        schrieb zuletzt editiert von
        #3

        @briankrebs Sounds like "Do not look into laser with remaining eye."

        1 Antwort Letzte Antwort
        0
        • royal@theres.lifeR royal@theres.life

          @briankrebs In the future, programming will consist of setting loose a random action generator and defining its behavior by a series of prohibitions.

          jonas@social.jonaskoeritz.deJ This user is from outside of this forum
          jonas@social.jonaskoeritz.deJ This user is from outside of this forum
          jonas@social.jonaskoeritz.de
          schrieb zuletzt editiert von
          #4

          @royal @briankrebs I faintly remember "Chaos Monkey" being a legit tool that Netflix (?) used internally to just cause random outages in their systems to build resilience. This reminds me of that.

          royal@theres.lifeR 1 Antwort Letzte Antwort
          0
          • jonas@social.jonaskoeritz.deJ jonas@social.jonaskoeritz.de

            @royal @briankrebs I faintly remember "Chaos Monkey" being a legit tool that Netflix (?) used internally to just cause random outages in their systems to build resilience. This reminds me of that.

            royal@theres.lifeR This user is from outside of this forum
            royal@theres.lifeR This user is from outside of this forum
            royal@theres.life
            schrieb zuletzt editiert von
            #5

            @jonas @briankrebs Yes, it was Netflix, and I was thinking of something similar.

            1 Antwort Letzte Antwort
            0
            • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

              Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

              I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

              "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

              https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

              noplasticshower@infosec.exchangeN This user is from outside of this forum
              noplasticshower@infosec.exchangeN This user is from outside of this forum
              noplasticshower@infosec.exchange
              schrieb zuletzt editiert von
              #6

              @briankrebs old style thinking about new stuff

              If you start working on this ping me and we can talk

              briankrebs@infosec.exchangeB 1 Antwort Letzte Antwort
              0
              • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                womble@infosec.exchangeW This user is from outside of this forum
                womble@infosec.exchangeW This user is from outside of this forum
                womble@infosec.exchange
                schrieb zuletzt editiert von
                #7

                @briankrebs if there's one thing that the entire history of humanity has taught us, it's that as a species we are great at making nuanced, highly context-specific decisions with incomplete information. I foresee that this is going to turn out great.

                1 Antwort Letzte Antwort
                0
                • noplasticshower@infosec.exchangeN noplasticshower@infosec.exchange

                  @briankrebs old style thinking about new stuff

                  If you start working on this ping me and we can talk

                  briankrebs@infosec.exchangeB This user is from outside of this forum
                  briankrebs@infosec.exchangeB This user is from outside of this forum
                  briankrebs@infosec.exchange
                  schrieb zuletzt editiert von
                  #8

                  @noplasticshower Are you letting AI agents manage your life?

                  noplasticshower@infosec.exchangeN 1 Antwort Letzte Antwort
                  0
                  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                    hellpie@raru.reH This user is from outside of this forum
                    hellpie@raru.reH This user is from outside of this forum
                    hellpie@raru.re
                    schrieb zuletzt editiert von
                    #9

                    @briankrebs I am going to start saying some stuff like "back in my day APTs didn't forget shit all the time" and "when spyware was made the old fashioned way it just worked without needing to ask you anything"

                    1 Antwort Letzte Antwort
                    0
                    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                      leahprice@hcommons.socialL This user is from outside of this forum
                      leahprice@hcommons.socialL This user is from outside of this forum
                      leahprice@hcommons.social
                      schrieb zuletzt editiert von
                      #10

                      @briankrebs Makes you nostalgic for 1860, when machine-generated text meant a new magazine. (from PUNCH)

                      1 Antwort Letzte Antwort
                      0
                      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                        Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                        I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                        "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                        https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                        slyborg@vmst.ioS This user is from outside of this forum
                        slyborg@vmst.ioS This user is from outside of this forum
                        slyborg@vmst.io
                        schrieb zuletzt editiert von
                        #11

                        @briankrebs Can’t wait for the slew of stories / TikTok shorts about ‘how my lifehack AI agent phone buddy converted my life savings to Monero and sent it to North Korea’ and that nobody will learn a thing from them.

                        1 Antwort Letzte Antwort
                        0
                        • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                          @noplasticshower Are you letting AI agents manage your life?

                          noplasticshower@infosec.exchangeN This user is from outside of this forum
                          noplasticshower@infosec.exchangeN This user is from outside of this forum
                          noplasticshower@infosec.exchange
                          schrieb zuletzt editiert von
                          #12

                          @briankrebs of course!

                          The "let's secure one (or maybe 5) agent(s) at a time" security thing is cute. If I read about another A&A framework approach to this stuff I am going to start using agents to run my life.

                          toriver@mas.toT 1 Antwort Letzte Antwort
                          0
                          • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                            Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                            I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                            "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                            https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                            zl2tod@mastodon.onlineZ This user is from outside of this forum
                            zl2tod@mastodon.onlineZ This user is from outside of this forum
                            zl2tod@mastodon.online
                            schrieb zuletzt editiert von
                            #13

                            @briankrebs

                            Thread re #clawdbot :

                            https://infosec.exchange/@munin/115963686278347109

                            1 Antwort Letzte Antwort
                            0
                            • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                              Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                              I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                              "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                              https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                              C This user is from outside of this forum
                              C This user is from outside of this forum
                              crouton@aus.social
                              schrieb zuletzt editiert von
                              #14

                              @briankrebs This sounds like the optimally worst implementation of a digital assistant. I was looking for a distant variant of this, where I set the (mostly deterministic) rules and actions, etc. Kinda like HA/node-red but aimed at being an assistant rather than controlling a house.
                              Giving it a blank cheque and hooking up to a LLM is insane.

                              1 Antwort Letzte Antwort
                              0
                              • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                hal8999@infosec.exchangeH This user is from outside of this forum
                                hal8999@infosec.exchangeH This user is from outside of this forum
                                hal8999@infosec.exchange
                                schrieb zuletzt editiert von
                                #15

                                @briankrebs The stories are writing themselves: My A.I. agent took over my finances, framed me for sex trafficking, then unlocked the doors and turned on lights for the police.

                                1 Antwort Letzte Antwort
                                0
                                • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                  Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                  I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                  "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                  https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                  carsten@minnesotasocial.netC This user is from outside of this forum
                                  carsten@minnesotasocial.netC This user is from outside of this forum
                                  carsten@minnesotasocial.net
                                  schrieb zuletzt editiert von
                                  #16

                                  @briankrebs and if the bot "touches" something it was not allowed to touch? "Sorry, me bad, won't do it again, probably" - does it again.

                                  1 Antwort Letzte Antwort
                                  0
                                  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                    Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                    I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                    "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                    https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                    zompetto@mastodon.artZ This user is from outside of this forum
                                    zompetto@mastodon.artZ This user is from outside of this forum
                                    zompetto@mastodon.art
                                    schrieb zuletzt editiert von
                                    #17

                                    @briankrebs it would be funny if it wasn't so sad. And scary.

                                    1 Antwort Letzte Antwort
                                    0
                                    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                      Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                      I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                      "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                      https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                      theorangetheme@en.osm.townT This user is from outside of this forum
                                      theorangetheme@en.osm.townT This user is from outside of this forum
                                      theorangetheme@en.osm.town
                                      schrieb zuletzt editiert von
                                      #18

                                      @briankrebs It's fucking bonkerstown. You might as well just wire up a random number generator and every 1/10 times it just deletes your home directory. At least that achieves the same result but with a fraction of the electricity and human rights abuses.

                                      1 Antwort Letzte Antwort
                                      0
                                      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                        Letting AI agents run your life is like handing the car keys to your 5-year-old. What could go wrong?

                                        I was marveling while reading this PCMag piece, which describes how to secure an agentic AI setup that essentially mimics malware: To do it's job properly, the AI agent has to be able to read private messages, store credentials, execute commands, and maintain a persistent state. How do you do that? You chase after it like you would your child.

                                        "The important thing is to make sure you limit "who can talk to your bot, where the bot is allowed to act, [and] what the bot can touch" on your device, the bot's support documentation says."

                                        https://www.pcmag.com/news/clawdbot-moltbot-hot-new-ai-agent-creator-warns-of-spicy-security-risks?test_uuid=04IpBmWGZleS0I0J3epvMrC&test_variant=A

                                        S This user is from outside of this forum
                                        S This user is from outside of this forum
                                        spacelifeform@infosec.exchange
                                        schrieb zuletzt editiert von
                                        #19

                                        @briankrebs

                                        So sayeth the bot.

                                        It is smarter not to talk to them in the first place.

                                        1 Antwort Letzte Antwort
                                        0
                                        • noplasticshower@infosec.exchangeN noplasticshower@infosec.exchange

                                          @briankrebs of course!

                                          The "let's secure one (or maybe 5) agent(s) at a time" security thing is cute. If I read about another A&A framework approach to this stuff I am going to start using agents to run my life.

                                          toriver@mas.toT This user is from outside of this forum
                                          toriver@mas.toT This user is from outside of this forum
                                          toriver@mas.to
                                          schrieb zuletzt editiert von
                                          #20

                                          @noplasticshower @briankrebs
                                          1) Configure agent with guardrails
                                          2) Agent runs into guardrails
                                          3) Agent spins up secondary agent without guardrails
                                          4) Oh no.

                                          noplasticshower@infosec.exchangeN 1 Antwort Letzte Antwort
                                          0
                                          Antworten
                                          • In einem neuen Thema antworten
                                          Anmelden zum Antworten
                                          • Älteste zuerst
                                          • Neuste zuerst
                                          • Meiste Stimmen



                                          Copyright (c) 2025 abSpecktrum (@abspecklog@fedimonster.de)

                                          Erstellt mit Schlaflosigkeit, Kaffee, Brokkoli & ♥

                                          Impressum | Datenschutzerklärung | Nutzungsbedingungen

                                          • Anmelden

                                          • Du hast noch kein Konto? Registrieren

                                          • Anmelden oder registrieren, um zu suchen
                                          • Erster Beitrag
                                            Letzter Beitrag
                                          0
                                          • Home
                                          • Aktuell
                                          • Tags
                                          • Über dieses Forum