• ☆ Yσɠƚԋσʂ ☆OP
    link
    fedilink
    arrow-up
    7
    arrow-down
    2
    ·
    3 days ago

    there’s some more info with benchmarks here, it does as well and in some cases better than top tier commercial models https://www.analyticsvidhya.com/blog/2024/12/deepseek-v3/

    The trick that makes it possible is the mixture-of-experts approach. While it has 671 billion parameters overall, it only uses 37 billion at a time, making it very efficient. For comparison, Meta’s Llama3.1 uses 405 billion parameters used all at once. It also has 128K token context window means it can process and understand very long documents, and processes text at 60 tokens per second, twice as fast as GPT-4o.