Yann LeCun’s Post

View profile for Yann LeCun, graphic
Yann LeCun Yann LeCun is an Influencer

🥁 Llama3 is out 🥁 8B and 70B models available today. 8k context length. Trained with 15 trillion tokens on a custom-built 24k GPU cluster. Great performance on various benchmarks, with Llam3-8B doing better than Llama2-70B in some cases. More versions are coming over the next few months. https://lnkd.in/dNZx72FJ

  • No alternative text description for this image
Cesar Garcia

Independent researcher at La Hora Maker

7mo

I am quite happy about the release of Llama 3. The only thing I hoped it would change (and make Llama 3 much more useful) is this clause of the license: You will not use the Llama Materials or any output or results of the Llama Materials to improve any other large language model (excluding Meta Llama 3 or derivative works thereof). So, I thing we should stick to Mistral series to generate any kind of open sourced synthetic datasets.

Karan Sachdeva

IBM AWS Global Strategic Alliance Leader for AI and Data @ IBM

7mo

Just explained llama 3 value prop to a customer-Imagine Llama3 as a highly skilled librarian who has read an enormous amount of books- -1.5 trillion to be exact! This librarian comes in two versions: one can juggle 8 billion books (simpler version) and the other 70 billion books (more complex version) allowing them to pull from a vast amount of knowledge to answer questions. The librarian has a memory that can retain the context of up to 8,000 books at once, helping them remember and connect details over a long conversation.

Thank you Yann . Why there isn't a comparison with GPT 4? I can bet this is going to make new waves in the open-source community and hugging face is soon going to be flooded with its fine-tuned versions. 😃

Someone know how to offload the model downloaded in Hugging Face? The model it's too heavy for my computational resource

Tarek Khalil (TK)

I help sales teams succeed using #WhatsApp 🚀📊📈

7mo

Groq supporting it soon? 🙏🙏🙏

Dean Horak

Sr Software Engineer at Leidos, certified AI/Machine Learning developer

7mo

I’m a bit disappointed as it is only trained on data as late as December 2022 and does not have internet access. That is quite stale data as compared to the leading LLMs.

Anshuman (Ansh) Pandey

Co-founder and CEO | Driving GenAI adoption at Enterprises

7mo

Sire, you are a legend! 🫡 P.S. Llama 3 live on Tune AI

  • No alternative text description for this image
Dmitriy Kalyada

🚀 Founder | 🧠 @ Allessent <-> AI Ecosystem Architect | 👗💻 Digital Fashion Disruptor

7mo

🏎️ The Llama family is a key GenAI driver. Since the first LLama family models, it has also served as a symbol of responsible AI in the industry, being an open-source platform. 🦙🏆 Llama3 has raised the bar for research and development in the LLM-based product domain, pushing efficiency to new heights and setting a new standard for the field.

Konstantine Arkoudas

Chief AI Officer & Architect

7mo

Congratulations, the numbers are remarkable. If I'm reading the benchmarks correctly, the 70B version outperforms GPT-4 on MATH and GPQA. However, a context size of 8K is underwhelming in April 2024. Looking forward to upcoming releases! 

Manish Sainani 🤫

Founder, Engineer & CEO @ hushh 🤫, ex-Google Director of Product - Core Developer & ML, GPU TPU AI/ML Infra, Board Member, Tech Advisor & Investor, ex-Microsoft AI, Splunk AI/ML, E&Y / Capgemini, UBS, ING, Adobe, Sun

7mo

Artificial Intelligence (AI) can significantly enhance data accessibility by automating the classification, cataloging, and quality control of data. This automation streamlines the process of making data easily accessible and usable, reducing errors and improving efficiency. For example, at hushh.ai, AI could automate the organization and retrieval of user data, ensuring it's both accessible and valuable with consent, aligning with your mission to make user-data universally accessible.

See more comments

To view or add a comment, sign in

Explore topics