• Naz@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    ·
    5 months ago

    Try using a 1-bit LLM to test the article’s claim.

    The perplexity loss is staggering. It’s like 75% accuracy lost or more. It turns a 30 billion parameter model into a 7 billion parameter model.

    Highly recommended that you try to replicate their results.