Rajesh Das’ Post

View profile for Rajesh Das, graphic

Senior software engineer (Gen AI) @ Happiest Minds| Ex-Data scientist @ TVS motor | Mathematics | Data science | Quantum Computing

𝐋𝐨𝐑𝐀 𝐋𝐚𝐧𝐝: 310 𝐅𝐢𝐧𝐞-𝐭𝐮𝐧𝐞𝐝 𝐋𝐋𝐌𝐬 𝐭𝐡𝐚𝐭 𝐑𝐢𝐯𝐚𝐥 𝐆𝐏𝐓-4, 𝐀 𝐓𝐞𝐜𝐡𝐧𝐢𝐜𝐚𝐥 𝐑𝐞𝐩𝐨𝐫𝐭 Low Rank Adaptation (LoRA) has emerged as one of the most widely adopted methods for Parameter Efficient Fine-Tuning (PEFT) of Large Language Models (LLMs). LoRA reduces the number of trainable parameters and memory usage while achieving comparable performance to full fine-tuning. The paper shows that 4-bit LoRA fine-tuned models outperform base models by 34 points and GPT-4 by 10 points on average. This paper also introduces LoRAX, an open-source Multi-LoRA inference server that facilitates the deployment of multiple LoRA fine-tuned models on a single GPU using shared base model weights and dynamic adapter loading. 𝐋𝐨𝐑𝐀𝐗 𝐩𝐨𝐰𝐞𝐫𝐬 𝐋𝐨𝐑𝐀 𝐋𝐚𝐧𝐝, a web application that hosts 25 LoRA fine-tuned Mistral-7B LLMs on a single NVIDIA A100 GPU with 80GB memory. 𝐋𝐨𝐑𝐀 𝐋𝐚𝐧𝐝 highlights the quality and cost-effectiveness of employing multiple specialized LLMs over a single, general-purpose LLM. LoRA Land - https://lnkd.in/epmgenNN Happiest Minds Technologies , Happiest Minds Generative AI , Sridhar Mantha , Praveen R P , Srikant Sowmyanarayanan

LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report

LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report

arxiv.org

To view or add a comment, sign in

Explore topics