🥁 Llama3 is out 🥁 8B and 70B models available today. 8k context length. Trained with 15 trillion tokens on a custom-built 24k GPU cluster. Great performance on various benchmarks, with Llam3-8B doing better than Llama2-70B in some cases. More versions are coming over the next few months. https://lnkd.in/dNZx72FJ
Just explained llama 3 value prop to a customer-Imagine Llama3 as a highly skilled librarian who has read an enormous amount of books- -1.5 trillion to be exact! This librarian comes in two versions: one can juggle 8 billion books (simpler version) and the other 70 billion books (more complex version) allowing them to pull from a vast amount of knowledge to answer questions. The librarian has a memory that can retain the context of up to 8,000 books at once, helping them remember and connect details over a long conversation.
Thank you Yann . Why there isn't a comparison with GPT 4? I can bet this is going to make new waves in the open-source community and hugging face is soon going to be flooded with its fine-tuned versions. 😃
Someone know how to offload the model downloaded in Hugging Face? The model it's too heavy for my computational resource
Groq supporting it soon? 🙏🙏🙏
I’m a bit disappointed as it is only trained on data as late as December 2022 and does not have internet access. That is quite stale data as compared to the leading LLMs.
🏎️ The Llama family is a key GenAI driver. Since the first LLama family models, it has also served as a symbol of responsible AI in the industry, being an open-source platform. 🦙🏆 Llama3 has raised the bar for research and development in the LLM-based product domain, pushing efficiency to new heights and setting a new standard for the field.
Congratulations, the numbers are remarkable. If I'm reading the benchmarks correctly, the 70B version outperforms GPT-4 on MATH and GPQA. However, a context size of 8K is underwhelming in April 2024. Looking forward to upcoming releases!
Artificial Intelligence (AI) can significantly enhance data accessibility by automating the classification, cataloging, and quality control of data. This automation streamlines the process of making data easily accessible and usable, reducing errors and improving efficiency. For example, at hushh.ai, AI could automate the organization and retrieval of user data, ensuring it's both accessible and valuable with consent, aligning with your mission to make user-data universally accessible.
Independent researcher at La Hora Maker
7moI am quite happy about the release of Llama 3. The only thing I hoped it would change (and make Llama 3 much more useful) is this clause of the license: You will not use the Llama Materials or any output or results of the Llama Materials to improve any other large language model (excluding Meta Llama 3 or derivative works thereof). So, I thing we should stick to Mistral series to generate any kind of open sourced synthetic datasets.