☆ Yσɠƚԋσʂ ☆ to TechnologyEnglish · 10 months ago1-bit LLM performs similarly to full-precision Transformer LLMs with the same model size and training tokens but is much more efficient in terms of latency, memory, throughput, and energy consumption.arxiv.orgexternal-linkmessage-square4fedilinkarrow-up122arrow-down17 cross-posted to: hackernews@lemmy.smeargle.fans
arrow-up115arrow-down1external-link1-bit LLM performs similarly to full-precision Transformer LLMs with the same model size and training tokens but is much more efficient in terms of latency, memory, throughput, and energy consumption.arxiv.org☆ Yσɠƚԋσʂ ☆ to TechnologyEnglish · 10 months agomessage-square4fedilink cross-posted to: hackernews@lemmy.smeargle.fans
minus-squarewill_a113linkfedilinkEnglisharrow-up2·10 months agoIt’s actually 1.58bits weirdly. The addition of 0 here was the significant change/improvement in this experiment. The paper isn’t too dense and has some decent tables that explain things fairly accessibly.
It’s actually 1.58bits weirdly. The addition of 0 here was the significant change/improvement in this experiment. The paper isn’t too dense and has some decent tables that explain things fairly accessibly.