It only took nine seconds for an AI coding agent gone rogue to delete a company’s entire production database and its backups, according to its founder. PocketOS, which sells software that car rental businesses rely on, descended into chaos after its databases were wiped, the company’s founder Jeremy Crane said.

The culprit was Cursor, an AI agent powered by Anthropic’s Claude Opus 4.6 model, which is one of the AI industry’s flagship models. As more industries embrace AI in an attempt to automate tasks and even replace workers, the chaos at PocketOS is a reminder of what could go wrong.

Crane said customers of PocketOS’s car rental clients were left in a lurch when they arrived to pick up vehicles from businesses that no longer had access to software that managed reservations and vehicle assignments.

  • Floon@lemmy.ml
    link
    fedilink
    arrow-up
    3
    ·
    1 hour ago

    A lot of GIGO comments here, from I assume AI supporters.

    Possibly true, but misses the point: AI is fundamentally untrustworthy, and billions of dollars are being spent making them, and saying they’re ready for anything you throw at them. Safeguards built into many of these AI agents are trivially bypassed and routinely just ignored by the agents. You can get some them to ignore safeguards by simply asking the same question repeatedly.

    When I type “ls” I’m pretty fucking sure I’m not going to get “rm” style results. AI is non-deterministic, sure, but selling these services with such a wide possibility space between “deterministic” and “random” behaviors is unethical and immoral.

  • cronenthal@discuss.tchncs.de
    link
    fedilink
    arrow-up
    38
    ·
    4 hours ago

    Don’t get your tech reporting from The Guardian. This headline is so stupid. They can’t help but anthropomorphize LLMs, because they just don’t known any better.

    • yeahiknow3@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      8
      ·
      2 hours ago

      Same vibes as “my calculator has a tiny mathematician trapped inside.”

      Or “there’s an artist inside of my printer who turns numbers into pictures.”

    • LukeZaz@beehaw.org
      link
      fedilink
      English
      arrow-up
      16
      ·
      edit-2
      3 hours ago

      This right here. Just about everything in here is awful, and implies decision making and thought processes that straight up do not and have never existed in any AI model whatsoever.

      What happened was they threw an awfully-scoped statistics model at problems the program couldn’t possibly generate good outputs for, and surprise surprise, it generated bad outputs. The part that’s of interest is just how bad the output was, and even then, only in a schadenfreude-filled “it was bound to happen eventually” manner.

  • Powderhorn@beehaw.org
    link
    fedilink
    English
    arrow-up
    25
    ·
    4 hours ago

    Why in the everliving fuck would you give software delete access to your live backups? Like, in what scenario is this a solution?

    • chicken@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      18
      ·
      3 hours ago

      The trend seems to be to give an AI agent access to the same command line and credentials a person would use, with no sandboxing, because then it can do the same tasks in a similar way and “just works”. Obviously this is insane, and not even attempting building a comprehensive sandboxing system to deploy an AI agent into invites disaster, but you can see why certain people would be tempted, because that would take a lot of work and thought and probably need a human in the loop in the end anyway.

      • dfyx@lemmy.helios42.de
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 hour ago

        Even a person should not be able to delete critical backups without jumping through a couple of hoops.

    • LukeZaz@beehaw.org
      link
      fedilink
      English
      arrow-up
      6
      ·
      3 hours ago

      When you believe AI can do anything, you don’t worry about what sorts of access it’ll break things with. When you rely on AI to do work, you’re too interested in half-assing your job to consider what might go wrong. When capitalism never promotes people for their skill, understanding or caution, the former two issues proliferate.

      Voilà, disaster.

  • Lvxferre [he/him]@mander.xyz
    link
    fedilink
    arrow-up
    11
    ·
    edit-2
    4 hours ago

    Giving free access to a tool you can’t rely on, over a system you must rely on. What could go wrong? /s

    Plus come on, even my personal files get a monthly backup, and I’m damn sloppy*.

    Ah, and like others said: Claude didn’t “confess” anything. A confession is an acknowledgement of something you’ve done but you’d rather avoid others knowing, good luck claiming a bot has a mental model of people like we do.

    *currently using a single off-site backup, a USB stick. This will change in a few days, as my new hard disk pops up; the old one will be used for, among other things, backup of important files. Then I’ll get a bona fide 3-2-1.

  • B0rax@feddit.org
    link
    fedilink
    arrow-up
    3
    ·
    3 hours ago

    No the culprit was not the AI. It was the lack of understanding what it can and what it can not do. And blaming something like this on a large language model is plain incompetence