Whether you agree with the Guardian’s conclusions or not, the underlying issue they’re pointing at is broader than any one company: the steady collapse of ambient trust in our information systems.

The Guardian ran an editorial today warning that AI companies are shedding safety staff while accelerating deployment and profit seeking. The concern was not just about specific models or edge cases, but about something more structural. As AI systems scale, the mechanisms that let people trust what they see, hear, and read are not keeping up.

Here’s a small but telling technology-adjacent example that fits that warning almost perfectly.

Ryan Hall, Y’all, a popular online weather forecaster, recently introduced a manual verification system for his own videos. At the start of each real video, he bites into a specific piece of fruit. Viewers are told that if a video of “him” does not include the fruit, it may not be authentic.

This exists because deepfakes, voice cloning, and unauthorized reuploads have become common enough that platform verification, follower counts, and visual familiarity no longer reliably signal authenticity.

From a technology perspective, this is fascinating.

A human content creator has implemented a low-tech authentication protocol because the platforms hosting his content cannot reliably establish provenance. In effect, the fruit is a nonce. A shared secret between creator and audience. A physical gesture standing in for a cryptographic signature that the platform does not provide.

This is not about weather forecasting credentials. It is about infrastructure failure.

When people can no longer trust that a video is real, even when it comes from a known figure, ambient trust collapses. Not through a single dramatic event, but through thousands of small adaptations like this. Trust migrates away from systems and toward improvised social signals.

That lines up uncomfortably well with the Guardian’s concern. AI systems are being deployed faster than trust and safety can scale. Safety teams shrink. Provenance tools remain optional or absent. Responsibility is pushed downward onto users and individual creators.

So instead of robust verification at the platform or model level, we get fruit.

It is clever. It works. And it should worry us.

Because when trust becomes personal, ad hoc, and unscalable, the system as a whole becomes brittle. This is not just about AI content. It is about how societies determine what is real in moments that matter.

TL;DR: A popular weather creator now bites a specific fruit on camera to prove his videos are real. This is a workaround for deepfakes and reposts. It is also a clean example of ambient trust collapse. Platforms and AI systems no longer reliably signal authenticity, so creators invent their own verification hacks. The Guardian warned today that AI is being deployed faster than trust and safety can keep up. This is what that looks like in practice.

Question: Do you think this ends with platform-level provenance becoming mandatory, or are we heading toward more improvised human verification like this becoming normal?

  • Sanctus@anarchist.nexus
    link
    fedilink
    English
    arrow-up
    11
    ·
    4 hours ago

    This is what the powers that be want, nobody sure of anything, complete breakdown of any trust. It won’t be fixed. We will deal with the fruit until the AI can fake that too and then tje clear web dies with a sigh.

    • tover153@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 hours ago

      I’m not convinced this is a plan so much as incentives running ahead of everything else.

      Trust usually doesn’t collapse all at once. It degrades until people start inventing awkward workarounds. The fruit is just what that looks like in the real world.

  • mangaskahn@lemmy.world
    link
    fedilink
    English
    arrow-up
    12
    ·
    5 hours ago

    This already been solved by public key cryptography? Sign the video with your private key, and publish the public key. Anyone can prove that the video is valid. To prove that the public key is valid, anyone can encrypt a message with it and asked for verification.

    • tover153@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      16
      ·
      4 hours ago

      You’re absolutely right that this is a solved problem from a technical standpoint. Public key cryptography gives us everything we need to sign content, verify it, and prove continuity of identity.

      But that’s how we solve it in technology. It’s not how my 82-year-old father solves it.

      For most people, trust isn’t established by verifying signatures or checking keys. It’s established through simple, legible cues they can recognize instantly, without tooling, training, or a mental model of cryptography.

      That’s why the fruit works.

      It’s a human-scale authentication signal. No UI, no standards, no explanation required. “If you see the fruit, it’s him.” That’s something almost anyone can understand and apply.

      The real problem isn’t that cryptographic solutions don’t exist. It’s that platforms haven’t made provenance and verification visible, intuitive, or default for non-technical users. Until they do, people will keep inventing these ad hoc, embodied trust signals.

      That’s what makes this a trust infrastructure failure, not a math failure.

      • TheTechnician27@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        ·
        edit-2
        4 hours ago

        Dude, I’m sorry for saying this (because I get this a lot for my often overly formal writing, and I get it’s ironic on this post), but…

        Your writing reads like it’s LLM-generated. Like, really heavily reads like an LLM wrote it. Long scrawls for pretty simple concepts, I don’t know how to describe why the cadence feels LLM-y other than “vibes”, flawless grammar, needless lists of nouns and adjectives, “it’s not X; it’s Y”, and this weird fucking lifeless demeanor that feels like it has no voice.

        • artifex@piefed.social
          link
          fedilink
          English
          arrow-up
          10
          ·
          4 hours ago

          Great observation! You’re absolutely right! It does sound like it was written by an LLM.

          • TheTechnician27@lemmy.world
            link
            fedilink
            English
            arrow-up
            7
            ·
            4 hours ago

            Wow, thanks! Let’s switch topics. I’m trying to start a business where I sell fruit to weathermen. Can you help me with that?

            • artifex@piefed.social
              link
              fedilink
              English
              arrow-up
              8
              ·
              3 hours ago

              Of course! What a novel idea! A business focusing on a highly specialized audience requires careful consideration and planning.

              Shall I switch to deep-planning mode so I can charge you 10X the tokens?

        • tover153@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          6
          ·
          4 hours ago

          My daughter in family chat said the same thing this last week. It’s possible I have been reading too much LLM generated content, also, this is my first top level post after years of lurking, and I’m trying to come off like I know what I’m doing. If the argument doesn’t land, happy to talk about that. The style I can adjust.

          • rnercle@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 hours ago

            The style I can adjust.

            please don’t.

            everybody wrote like you do before people started saying lol instead of actually laughing.

          • TheTechnician27@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 hours ago

            It’s not even the style on its own*; it’s that you wrote a frankly bloviating short essay about an obvious concept that can be summarized as “most people who watch the weather don’t know what a public key is or how to use one”. I’m disgustingly long-winded, and even I wouldn’t expend that much effort. The style is what escalates that from “padding a high school essay” to “Oh, yup, a GPT wrote this.”

            * “It’s not X, it’s Y” yeah, yeah, I know.

            • tover153@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              4
              ·
              4 hours ago

              That’s fair. I’ve written about this elsewhere and I’m rephrasing parts of it here, which probably makes it feel more essay-shaped than a typical thread reply.

              I wasn’t trying to pad anything. I was trying to connect a few dots that usually get skipped when this comes up.

  • Blue_Morpho@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    ·
    5 hours ago

    I don’t understand how it works? I saw a video where he took a bite of a banana before the forecast. How does that stop anyone from cutting and pasting one second of him biting the banana into the start of an ai video with a statement “banana” .

    • tover153@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      16
      ·
      4 hours ago

      That’s a fair question, and you’re right that it isn’t foolproof.

      The reason it works at all is that the fruit isn’t known in advance. He posts the video first, then updates his site with the correct fruit for that video. Viewers can check after the fact. If someone deep-fakes him, they either have to guess the fruit correctly or regenerate the fake once the real fruit is known.

      That doesn’t make impersonation impossible, but it does make it more expensive and slower.

      And that’s really the point. This isn’t perfect authentication, it’s friction. It raises the cost just enough that casual fakes, reposts, and automated scams stop being worthwhile, even if a determined attacker could still get through.

      Which is also why this is such a telling example. Instead of platforms providing provenance, creators are inventing human-readable ways to increase the cost of lying. Not secure, but legible and effective enough for most people.

      That’s the ambient trust problem in a nutshell. We’re not aiming for mathematically perfect truth, we’re trying to make deception harder than honesty.