Motaseam Yousef’s Post

View profile for Motaseam Yousef, graphic

Data Scientist @ KABi | SE | DA | DS | ML | DL | CV | NLP | LLM | RS | TS

Hey everyone, I stumbled upon this super cool paper recently: 'The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits'. It's a game-changer! Picture this: we switch up the LLM weights from float64 to ternary [-1, 0, 1]. Boom! Instant improvement in power, latency, and memory usage. It's like upgrading to turbo mode for your AI! So, we go ahead and convert those weights to ternary (Pareto improvement!), by quantization function. That basically means we're simplifying things by only having three options instead of a gazillion. then when we calculate the output by: Sum( weights * input + bias) We're just adding stuff up without all the fancy multiplication. It's become an addition process only. Who knew you could get the same result with less hardware? check the result below paper link:https://lnkd.in/eu6jpzuY #LLM

  • No alternative text description for this image
Kinan Hasan

Machine Learning Student

9mo

Art!

Like
Reply
See more comments

To view or add a comment, sign in

Explore topics