Aman Chadha’s Post

View profile for Aman Chadha, graphic

GenAI Leadership @ AWS • Stanford AI • Ex-, Amazon Alexa, Nvidia, Qualcomm • EB-1 "Einstein Visa" Recipient/Mentor • EMNLP 2023 Outstanding Paper Award

🌟 Live RAG Comparison Test: Pinecone vs. MongoDB vs. Postgres vs. SingleStore ❓Why It Matters: 🔹The role of Retrieval-Augmented Generation (RAG) has become increasingly significant in today's #GenAI landscape. RAG combines the best of both worlds of generative and retrieval-based models to provide more contextually accurate and rich responses in LLM apps.  🔹Latency is crucial for RAG in LLM apps because it directly impacts the user experience by determining how quickly and interactively the model can retrieve and integrate relevant information to generate accurate responses 🔹Rohit Bhamidipati will offer a hands-on RAG performance evaluation of leading back-end/database vendors such as Pinecone vs. MongoDB vs. Postgres vs. SingleStore, comparing throughput, efficiency, and latency for large datasets. 📚 What You'll Learn: ➡️ The mechanics of RAG and its impact on enhancing language model responses. ➡️ How Vector Databases like Pinecone, MongoDB, PostgreSQL, and SingleStore facilitate the functionality of RAG. ➡️ Comparative analysis showcasing real-time performance metrics of these databases. ➡️ Best practices for integrating these technologies into your AI and ML projects to boost efficiency and accuracy. ⚡️ Can't make it? No worries! All registrants will receive a copy of the webinar recording and additional resources via email post-session ✅ Live demo and code-share session will be offered for a hands-on experience 📅 Event Details: 🔹 Thursday, May 9 @ 10-11 AM PDT 🔹 #Free Registration: https://lnkd.in/g6EvB8NG #artificialintelligence #rag #ad

Mike Burger

Co-Founder HQforAI | Team Builder | Gen AI Coach | Everything Data | Top AI Voice

7mo

Hi Aman Chadha, how do you weigh the importance of an embedding model vs. the vector store itself? It is my understanding that Pinecone (as an example) is a vector database in which you can store data and embeddings…but the embedding model is a separate but critical component of the process depending on the quality, structure, and types of data you are looking to RAG. What’s your strategy for chunking and embedding (before vector DB)?

Like
Reply

To view or add a comment, sign in

Explore topics