• Mark with a Z@suppo.fi
    link
    fedilink
    English
    arrow-up
    19
    ·
    7 days ago

    So, nvidia, a gpu maker, invents a technology that reduces memory needs in exchange for more gpu needs. Bonus points for AI.

    • fallaciousBasis@lemmy.world
      link
      fedilink
      English
      arrow-up
      20
      ·
      7 days ago

      Nvidia has always had strong real time hardware accelerated memory compression.

      Compute is basically a free lunch compared to memory bottlenecks. And individual textures will probably fit in low level caches, which allows the compute to flex.

        • fallaciousBasis@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          5 days ago

          Compute units are significant, but so are the caches and i/o.

          Real world datasets tend to have a lot of sparsity.

          One of the biggest problems is called a page fault. Which is basically when the app needs to go to storage to find data to continue execution. This results in the processor waiting, which isn’t free.

          Generally, I’d say they go hand in hand about 50/50 plus or minus 10%.

          One benchmarks can fit in L1 cache and really stress the cores but most benchmarks you’re using all the levels of cache, ram, IO, etc… which is a hell of a lot compared to just a little bit of processor in that chip. GPUs again are often just massive collaborations of massive throughout and compute. So it can be to really separate.

          And legit. Most data is compressed these days on the Internet. So that’s again compute used to save data in flight. It’s a neverending tradeoff.