• Eager Eagle@lemmy.world
    link
    fedilink
    English
    arrow-up
    36
    ·
    14 hours ago

    Waste of energy. It’s like asking a person to estimate a non-trivial angle. Either use a model trained for that task, or don’t bother.

    • Alvaro@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      8
      ·
      9 hours ago

      The point is that:

      1. It is being used for ut, even though it is obviously not capable of giving a reliable and realistic answer
      2. It allows this usage, even though it is dangerous and not within it’s capabilities
      3. Each model gives answers that vary wildly, something that a human wouldn’t do. A human wouldn’t give you answers that are 10x more for the same question randomly.
      • Eager Eagle@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        11 hours ago

        You’d expect the same answer each time. It’s the same photo, the same model, the same question. But you won’t get the same answer.

        I don’t know what ads show that, but anyone who knows the first thing about LLMs knows you don’t get the same answer twice.

        I’d get this expectation 5 years ago when most people weren’t familiar with it, but come on… you don’t need to feed it an image 500 times to see that.

        • Sandbar_Trekker@lemmy.today
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 hours ago

          Technically, you can get the same answer twice from an LLM, but only when you control the full input. When a model is being run, a random seed/hash is applied to the input. If you run the model locally you could force the seed to always be the same so that you would always get the same answer for a given question.

          • Eager Eagle@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            7 hours ago

            Barely. Even with the code and seeds, it’s still a struggle to do that. There’s plenty of questions from people running pytorch and tensorflow models that can’t reproduce results. Maybe you isolate enough variables that consecutive runs actually produce the same output, but the study is about commercial models. You’ll never get deterministic output from those.