Apple has discontinued the Mac Pro – but it’s just the first of the tower computers to go. The rest will follow soon.

Fruit-sniffers extraordaire 9-to-5 Mac got the news yesterday, complete with official confirmation from Apple itself. It’s official and it’s happened, but there have been warning signs for months – in November 2025, Bloomberg’s Matt Gurman said “The Mac Pro is on the back burner.”

The phantom fruit-flingers of Silicon Valley launched the seven-thousand-buck Apple Silicon-based Mac Pro in June 2023, with an M2 Ultra SoC. It sported seven PCIe slots – but the problem was that cash-rich customers couldn’t add the sorts of expansion that normally go into a PCIe slot… to the extent that Apple publishes a page about PCIe cards you can install in your Mac Pro (2023). Notably, the machine did not support add-on GPUs: only the GPU that’s integrated into the CPU complex along with the machine’s RAM and primary flash storage. The machine also had no RAM expansion whatsoever.

Presumably, this limited its appeal for many traditional buyers, and the machine never saw an M3 or M4 model, let alone the M5 SoC that The Register covered shortly before Bloomberg called the Arm64 cheesegrater’s fate.

  • artyom@piefed.social
    link
    fedilink
    English
    arrow-up
    1
    ·
    13 hours ago

    They put them in everything because they’re smaller and more efficient (and thus quieter) and because they’re competitive with PC desktops in performance. And economies of scale doesn’t hurt either.

    • Kairos@lemmy.today
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      13 hours ago

      I get what you mean. What I’m trying to say is that desktop/non integrated CPUs are cheaper and this cost savings continues into a large form factor. Apple doesn’t put a desktop chip in their iMacs because they don’t make one. That’s not what their customer base needs. If they did it’d be 4x faster for the same price.

      And these arm chips are slower than x86. X86 is so much faster at least for single core performance which matters a LOT more for desktop use cases

      • artyom@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        13 hours ago

        Again, it depends on the workload. There are endless comparisons between high end desktops and comparably-priced Mac desktops, and while the PC is often more powerful, that’s not always the case , and the Mac does it while being much quieter, and not turning the room it’s in into a sauna.

        • Kairos@lemmy.today
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          13 hours ago

          Yes as it turns out when your workload is 99999 idle applications, a larger number cores helps more than single core performance. SOCs don’t change that. They just reduce power and space usage at the expense of cost. It makes no sense to point at the special-case computing company and say that their special case will suddenly override a 50 year pattern.

          • artyom@piefed.social
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            13 hours ago

            It’s nothing to do with idle applications, you should really look more into this because what you’re saying is simply misinformed.

            • Kairos@lemmy.today
              link
              fedilink
              arrow-up
              1
              ·
              13 hours ago

              Can you give me an example? The video you linked has a timestamp to something about video encoding.

                • Kairos@lemmy.today
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  edit-2
                  12 hours ago

                  Which again would be cheaper if they put the chips in separate enclosures. Just way bigger and more power usage.

                  Macs are good at video editing because Apple actually givesa shit about hardware encoding. NVENC is the only competitor. Everything else is shit.