• boonhet@sopuli.xyz
    link
    fedilink
    arrow-up
    5
    ·
    15 hours ago

    22B

    There are 700B+ parameter open weight models now. Frontier models are in the trillions.

    And even that model apparently took a supercomputer to train. I don’t have a supercomputer so I can’t train my own models like I can compile my own software. This is not comparable to running Linux where you can just compile your own kernel or even whole operating system (former Gentoo user here).

    I’ve tried running the models my 8 GB card can handle. They’re OK for a quick question, but they won’t be doing anything useful for me.