This is a good article on using LLMs in the real world. It covers many aspects that we learned at Nurish over the last few months and more. Using LLMs is an inexact science, but has its own needs in processes and tools. While most of the attention has been on the LLMS, their capabilities, and RAG, there are mundane issues that need to be taken care of, issues where the solutions for pre-LLM development are not that well-suited. These include version management, integration, testing, data management, quality assessment, and more. This article covers many of those practical aspects. Eugene Yan, Bryan Bischof, Charles Frye & more https://applied-llms.org/
Rajesh Kandaswamy’s Post
More Relevant Posts
-
Built an AI-powered Summarizer Agent that processes and summarizes documents across formats (PDF, DOCX, XLSX, CSV) using the Groq API. 1. Automates extraction and summarization. 2. Supports large file handling with smart chunking. 3. Saves results directly into a CSV for easy analysis. This project, developed for an AI forensic researcher, is now open-source. Check it out on GitHub: https://lnkd.in/g24-qdmg #AI #Forensics #OpenSource #Automation #AIAgent
GitHub - al-kafi-sohag/summarizer-agent: Summarizer Agent is an AI-powered tool designed to efficiently process and summarize documents. It automates the scanning of files from a designated folder, validates their format, and utlizes Groq API to generate concise, accurate summaries. The summaries are saved in a CSV file, streamlining document analysis for quicker insights.
github.com
To view or add a comment, sign in
-
If you want practical knowledge to build with LLMs (with some theoretical knowledge sprinkled in), this guide is nothing short of amazing. Read, re-read, and keep these writings close. Think: Designing Data-Intensive Applications, but for LLM apps.
All three parts of "What We Learned from a Year of Building with LLMs" are now live on O'Reilly! We share 42 lessons across: • Tactics: https://lnkd.in/gfm4tutJ • Ops: https://lnkd.in/gPNhesmG • Strategy: https://lnkd.in/giyTXWRe Read all 42 lessons here: https://applied-llms.org A sample of what each section discusses: • Tactics: Prompting basics, RAG, workflows, caching, when to finetune, evals, guardrails • Ops: Looking at data, working with models, product and risk, building a team • Strategy: "No GPUs before PMF", "the system not the model", how to iterate, cost trends I learned a lot and had so much fun writing these, and hope you'll get as much out of reading them too. Such an honor to collab with this group of thoughtful practitioners: Bryan Bischof Charles Frye Hamel H. Jason Liu Shreya Shankar
What We’ve Learned From A Year of Building with LLMs – Applied LLMs
applied-llms.org
To view or add a comment, sign in
-
Very nice assessment on RAG and Fine-tuning related trade-offs. There's no single best solution. Rather, there are complementary solutions and selected design depends on application needs. #AITransformation
Language Models, AI Agents, Agentic Applications, Development Frameworks & Data-Centric Productivity Tools
These two approaches are often pitted against each other as to which one is the best. RAG is known for improving accuracy via in-context learning and is very affective where context is important. RAG is easier to implement and often serves as a first foray into implementing LLMs due to RAG’s inspectability, observability and not being as opaque as fine-tuning. With RAG the input token size increases with the prompt size, and the study found that the output token size tend to be more verbose and harder to steer. The study found that fine-tuning offers a precise, succinct output that is attuned to brevity. It is highly effective and presents opportunities to learn new skills in a specific domain. But the initial cost is high in terms of data preparation and running the fine-tuning process. But when it comes to usage, fine-tuning necessitates minimal input token size, making it a more efficient option for handling large data sets. It is evident that additional cost will have to be incurred somewhere; with RAG it is during model inference and not upfront. With fine-tuning the cost upfront is higher, with lower inference cost. Read the full post here: https://lnkd.in/dXzzbuYj
To view or add a comment, sign in
-
I completed a new course in AI: Journey Into Format Control in Prompt Engineering!
CodeSignal Course Certificate of Achievement
learn.codesignal.com
To view or add a comment, sign in
-
🌟 Response Accuracy 🌟 Accuracy is not just hallucination detection - it is a question of whether the response from a selected model generated a complete answer based on all available data. This is where understanding the Token Window (or Context Window) is critical. 🚀✨ Each model has a limit to the amount of data that can be passed through (including the prompt, training data and the answer). In picking which Model to use, you will want to consider the level of detail and amount of data needed to support an accurate response. Azure OpenAI GPT 3.5 has a Token Window of about 10 pages of text - so while it is the cheapest model, it can only pass a small amount of data. (Don't ask it to summarize the quest of your favorite character in a book, rather its better suited to generate a summary of a specific battle scene.) Azure OpenAI GPT-4.0 has a Token Window of about 20 pages - This allows for more data to be passed with the prompt, however GPT4 is about 5X more expensive than Davinci and 50X more expensive than GPT3.5. 🎥 Check out this video to learn more about the Context Window: https://lnkd.in/e8BpA5_P Shortcut --> Timestamp: 07:51 Token / Context Window Being able to move between models on the Ask Sage platform allows users to get optimal responses, ensuring better accuracy, cost efficiency and access to different types of content generation (like creating a image from text). Give us a try today! www.asksage.ai
Training Data & Token Window: A Comprehensive Guide
https://www.youtube.com/
To view or add a comment, sign in
-
Hearing the same question over and over again either in person or on the various networks, I decided to share my thoughts and of course, to get other points of view, suggestions and experiences. "How to introduce LLMs and the rapidly evolving ecosystem to engineers in my company with no background on the subject?", they ask. My quick answer (no presentation or template, sorry) is to try to build it as a workshop and cover the following: 1. Lightweight explanation about the mechanics behind how LLMs work (yes, the transform architecture as well - they're engineers, they should get it. At least high level). There's that cool GPT-2 excel sheet which is a great learning tool. Also mention the various LLMs out there, open vs closed source, and how they are currently benchmarked. 2. Prompt engineering, tactics and methods such as few shot learning, chain of thought, etc. Use a lot of live examples at this point to demonstrate how leveraging different methods improve the outcome. Since it's a workshop you can pause for a challenge aiming to achieve something with an LLM that is usually done by combining those methods. 3. Next I would suggest exposing them to embeddings, vector databases and the RAG concept. It has become such a common and cardinal practice, you shouldn't stop before going over this as well (but you might want to split the session into two). So after a short theoretical explanation about embeddings and vector DBs, followed by examples, introduce RAG and immediately start another hands on practice. There are so many resources online that will guide any developer to implement a system that takes documents, chunks them up, creates embeddings, stores them in a database and finally enables you to query the data, all in under one hour. Tip: Use company-specific data for this, such as your knowledge base, your blog posts, your product guides, history of a public Slack channel, etc. 4. Finally, talk about fine tuning LLMs (I wouldn't go too deep into this one), and if you haven't explained before, that creating embeddings is also done using a model, which can be a fine-tuned foundation model's embedding model. To sum it up, talk about the opportunities for your company/products/services that emerge given such capabilities. Sometimes "newcomers" have the best ideas! What would you add or do different? Have you tried something similar, how did it go?
To view or add a comment, sign in
-
What metrics do you track when evaluating #LLMs? 👀 Here‘s an overview of complex statistical and model-based scorers 💡 Bonus: we also cover the #evaluation #frameworks that help you get started assessing #LargeLanguageModels. https://lnkd.in/dsDxQQng
Evaluating LLMs: complex scorers and evaluation frameworks
symflower.com
To view or add a comment, sign in
-
With the IT community currently focused on the use of LLMs in business, it’s important to highlight the significance of solutions like RAG (Retrieval-Augmented Generation) platforms. As competition in the LLM space intensifies and token costs steadily decline, RAG is becoming the crucial link between business and LLMs and a major budget item for companies. RAG platforms essentially integrate all pre-existing search and ranking methods, fully leveraging their capabilities. I won’t delve into the technical specifics of what RAG entails, as the article below covers this comprehensively. It also discusses related technologies such as vector databases (with a rare, user-friendly explanation of what they are and why they’re needed), LLM prompt engineering, Knowledge Graphs, and more. In summary, RAG is a highly interesting and promising area of development. It represents both an independent class of solutions and a scientifically intensive field where innovations can provide a tangible competitive edge. From a business model perspective, RAG occupies a strategic place in the value chain, encompassing consulting, implementation, and licensing. Additionally, RAG (or its owner/user) determines which LLM will receive budget allocations. The second part of the article makes an intriguing and well-founded claim: as the contextual window of LLMs increases, the need for RAG in many tasks diminishes since all the necessary data can fit within the expanded contextual window. However, RAG remains relevant for an essential reason: rights management and access control. This cannot be managed at the LLM level, as LLMs can be convinced of anything, and this will remain true for a long time (if not always). https://lnkd.in/ehiQsg9Q
AI-Driven Predictive Analytics for Revenue Forecasting in Healthcare
https://www.datasciencecentral.com
To view or add a comment, sign in
-
Curious about how Large Language Models (LLMs) can revolutionize your data workflows? Join Mateusz Hordyński, our Technical Leader, as he explores the practicalities of integrating Large Language Models (LLMs) with relational databases. In this informative session, you’ll gain insights into: 👉 The tangible benefits of linking LLMs with structured data sources 👉 Real-world challenges and strategies to address them 👉 Case studies showcasing the practical applications of this integration Additionally, Mateusz introduces db-ally, our open-source library designed to simplify natural language querying of structured data, enhancing efficiency, security, and reliability: 🚀 https://lnkd.in/dBB7RFxm Expand your knowledge and improve your data interaction capabilities. Watch the full video now on YouTube: https://lnkd.in/dKdaKh4Z #DataScience #MachineLearning #AI #LLM #BigData #OpenSource #DataIntegration #NaturalLanguageProcessing #TechInsights
Bridging LLMs and databases. Lessons learned in production
https://www.youtube.com/
To view or add a comment, sign in
-
Join Mateusz Hordyński, our Technical Leader, as he delves into integrating Large Language Models (LLMs) with relational databases. Learn about the key benefits, address common challenges, and explore practical case studies. Additionally, discover db-ally, our innovative open-source library designed for efficient and secure natural language querying of structured data: https://lnkd.in/dXN2A7X9 Enhance your data strategy and operational efficiency - watch the full video on YouTube: https://lnkd.in/dmVEdNdy #DataScience #MachineLearning #AI #LLM #BigData #OpenSource #DataIntegration #BusinessInnovation
Curious about how Large Language Models (LLMs) can revolutionize your data workflows? Join Mateusz Hordyński, our Technical Leader, as he explores the practicalities of integrating Large Language Models (LLMs) with relational databases. In this informative session, you’ll gain insights into: 👉 The tangible benefits of linking LLMs with structured data sources 👉 Real-world challenges and strategies to address them 👉 Case studies showcasing the practical applications of this integration Additionally, Mateusz introduces db-ally, our open-source library designed to simplify natural language querying of structured data, enhancing efficiency, security, and reliability: 🚀 https://lnkd.in/dBB7RFxm Expand your knowledge and improve your data interaction capabilities. Watch the full video now on YouTube: https://lnkd.in/dKdaKh4Z #DataScience #MachineLearning #AI #LLM #BigData #OpenSource #DataIntegration #NaturalLanguageProcessing #TechInsights
Bridging LLMs and databases. Lessons learned in production
https://www.youtube.com/
To view or add a comment, sign in