• Eager Eagle@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      ·
      17 hours ago

      they are open weight and have a whitepaper, that’s already vastly better than whatever openai and anthropic are doing

    • flamingo_pinyata@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      16
      ·
      17 hours ago

      It’s an open weights model - you can run your own. If you’re lucky enough to have a GPU or two laying around … then it can be it’s a lot cheaper actually.

      • Rioting Pacifist@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        17 hours ago

        Sure but comparing what companies in an industry full of losers make, is pretty meaningless as we don’t know how much companies are losing right now.

        • FauxLiving@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          9 hours ago

          That doesn’t matter.

          We know DeepSeek costs 1/6th as much as the cost that OpenAI/Anthropic are selling their model’s tokens. We know this because we can see their prices online and compare that to the cost of running the DeepSeek model on comparable hardware (because it is open weight).

          If OpenAI/Anthropic are selling at a loss then it means their internal inferencing costs are even higher than the price that they list on their website.

          This means that the DeepSeek model is at least 1/6th the cost, assuming that OpenAI/Anthropic are selling at cost. If they are eating a loss then that means their inferencing is even more expensive so the DeepSeek model is even better than 1/6th the cost.

          • Rioting Pacifist@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            9 hours ago

            compare that to the cost of running the DeepSeek model on comparable hardware (because it is open weight).

            What?

            You can’t compare the cost of running a model at home to the cost of running a model operationally as a business.

            Or at least nobody should take that comparison seriously.

            If OpenAI/Anthropic are selling at a loss

            Why the If, we know they are running at huge losses.

            Honestly this level of poster analysis is why clankers are able to impress people, everyone has just accepted full bullshit all of the time and nobody is willing to admit the unknowns are significant and the numbers we are being presented are largely unverified/unverifiable!

            • FauxLiving@lemmy.world
              link
              fedilink
              English
              arrow-up
              4
              ·
              edit-2
              9 hours ago

              You can’t compare the cost of running a model at home to the cost of running a model operationally as a business.

              I didn’t say running it at home, that’s ridiculous. The Pro model would not run on home hardware.

              I said:

              compare that to the cost of running the DeepSeek model on comparable hardware

              Here, comparable hardware means an NVIDIA H100. A card who’s use has a well-known market price.

              Why the If, we know they are running at huge losses.

              Ok and I addressed that also:

              if they are eating a loss then that means their inferencing is even more expensive so the DeepSeek model is even better than 1/6th the cost.

              They are selling their inferencing at a loss, and therefore their inferencing cost is higher than the amount they charge. The 1/6th number comes from comparing the amount OpenAI charges to the amount it costs to run DeepSeek.

              Since their costs are higher, then the ratio of their price to DeepSeeks is even better than the quoted 1/6th figure.

              Because, in mathematics, if the numerator is fixed (the top number, i.e. the cost to run DeepSeek, which is known because you can run it yourself on the exact same hardware) and you increase the denominator (the bottom number, representing the cost of GPT/Claude) then the ratio becomes smaller.

              Since you agree that they’re losing money on inferencing then that means the bottom number is unknown, but we know it is higher than the price listed on their website. So, the 1/6th ratio represents the upper bounds on the ratio of costs.