• felsiq@piefed.zip
    link
    fedilink
    English
    arrow-up
    8
    ·
    1 day ago

    Did you use a heavily quantized version? Those models are much smaller than the state of the art ones to begin with, and if you chop their weights from float16 to float2 or something it reduces their capabilities a lot more