The July 25th Open-Source LLaMA 3.1 release deserves attention, especially to the investors of close source LLM models. 3 model sizes of 8B, 70B and 405B parameters enables edge to datacenter usage. Expanded context windows to 128k tokens supports much large prompt and multilingual support. The 405B parameter model competes well with GPT-4 and Claude 3.5 and is integrated with various cloud platforms. The LLaMA model can be fine-tuned and adapted for specific tasks without full retraining with techniques like LoRA as full retraining requires 2048 A100 GPUs for 21 days (estimate). The hardware costs alone for the training could reach tens of millions of dollars (painful to VCs) so efficient alternatives to full retraining is key to AI startups leveraging LLaMA. The approaches are growing, but include LoRA, 8-bit precision to reduce memory footprint and fine-tuning on specific datasets or tasks. Why did Meta open-source LLaMA? Meta aims to disrupt the competitive edge of companies with proprietary models such as Google and OpenAI. The traditional open-sourcing benefits of building a large developer base, fostering innovation and adoption. It also helps Meta attract and retain top AI researchers who prefer working with open and accessible technologies. Likely, Meta will offer managed services with specialized hardware in the future.
Well said!
I Help Companies Save Millions on Carrier & Datacenter Costs.
4moLLaMA's open-source release is indeed a game-changer. Efficient fine-tuning techniques are key. Meta's move disrupts AI dominance.