Lior Gazit’s Post

View profile for Lior Gazit, graphic

Machine Learning Group Manager

On-device #LLMs are getting a strong push from Meta: 1.🪶 Llama 3.2 includes a lightweight, text-only models (1B and 3B) that fit onto edge and mobile devices, including pre-trained and instruction-tuned versions. 2.🗣 The Llama 3.2 1B and 3B models support context length of 128K tokens and are state-of-the-art in their class for on-device use cases like summarization, instruction following, and rewriting tasks running locally at the edge. 3.👩🏻💻👨💻 They’re sharing the first official Llama Stack distributions, which will greatly simplify the way developers work with Llama models in different environments, including single-node, on-prem, cloud, and on-device, enabling turnkey deployment of retrieval-augmented generation (RAG) and tooling-enabled applications with integrated safety. 4.🤗 They’re making Llama 3.2 models available for download on llama.com and Hugging Face, as well as available for immediate development on our broad ecosystem of partner platforms, including AMD, AWS, Databricks, Dell, Google Cloud, Groq, IBM, Intel, Microsoft Azure, NVIDIA, Oracle Cloud, Snowflake, and more. Great news to all❗️

Llama 3.2: Revolutionizing edge AI and vision with open, customizable models

Llama 3.2: Revolutionizing edge AI and vision with open, customizable models

ai.meta.com

To view or add a comment, sign in

Explore topics