• WalnutLum
    link
    fedilink
    English
    arrow-up
    12
    ·
    5 months ago

    This is interesting but I’ll reserve judgement until I see comparable performance past 8 billion params.

    All sub-4 billion parameter models all seem to have the same performance regardless of quantization nowadays, so 3 billion is a little hard to see potential in.