Watch IBM’s Martin Keen discuss the importance of selecting the right large language model (LLM) for enterprise use. This video highlights three key metrics: performance, cost-effectiveness, and trustworthiness.
Martin explores the IBM Granite Foundation models designed to meet these requirements, offering transparency, scalability, and efficiency.
When it comes to picking a large language model with support for choice, last time I checked there was something like 700,000 different Ellms or large language models on Hugging Face. Now I'd like to cover just a couple of those, specifically the IBM Granite Foundation models. But first let's consider how to pick an enterprise grade foundation model, meaning an LLM suitable for deployment in an enterprise setting. Something you'd be happy to run your business with O Let's consider that through three different metrics. O, the foundation model, it needs to be performant. That's an important metric, but it also needs to be. Cost. Effective, and it needs to be. Trusted. Those are the three metrics we're going to consider and trusted of course, because you can't scale generative AI with models that you cannot trust O take these one by one now by. Performant we're talking about measurements like latency and throughput. Is the foundation model able to keep U with the seed and enterprise requires it to operate at? Then related to that. Is cost effectiveness. Now, according to the scientific journal Nature, a search that's driven by generative AI will use something like four to five times the amount of energy that's needed to run a conventional web search. So we need a foundation model that can deliver the necessary performance with low inferencing costs, and we need the foundation model. To be trusted and we can gauge that through metrics like hallucination scores, but also a model that offers transparency. So we know what data the model was trained on. And I think in many instances, models are kind of skewed a bit like this. They're highly performant, but they're expensive to run at inference time. And there's a lack of transparency on the training data the model was built with now. With the Granite models, IBM set out to create enterprise grade foundation models that alley an equal weight to all three of these metrics. So it looks more like this. So what should you know about the IBM Granite Foundation models? Well, many of the models are. Open source. You can find them on Hugging Face under the Apache 2.0 license. That enables broad commercial usage. Now these models also have transparency in training data, meaning we actually know the data sources that we use to train the models, and that's quite atypical. Most MLMS are notoriously vague on how their models were trained, so that's a nice change. Now Granite language models are trained on trusted enterprise. Data spanning academic code, legal and finance data sources as such as well the first 13 billion parameter Granite LLM was trained on about 6.5 terabytes of data and that includes 1.8 million scientific papers that were posted on archive. It also includes all US utility patterns granted by the US. DTO, and that's from 1975 all the way through to 2023 and it includes the public domain free law. Which are legal opinions from U.S. Federal and state courts. Essentially, the models have been governed and filtered to only use enterprise safe data sources. The granite models have also been designed to be performant as well. Especially in areas of coding and language tasks, outperforming some models that are actually twice their size and smaller models means also. Then more efficient with less compute requirements and at lower cost of inferencing. Now I keep mentioning the granite models, plural. So which models are we talking about? So granite is actually a family of LLM Foundation models spanning multiple modalities. And you can find many of these on Hugging Face. So let's take a look at some of them and we'll start with granite for language. Now these are decoder models of different parameter sizes O that includes A7B open source model and the B here refers to billions of parameters, so 7 billion parameters. There's also an. 8B model that's designed specifically for Japanese text. There's a couple of 13B models and there is a 20 billion parameter multilingual model that supports English, German, Spanish, French, and Portuguese. Now there's also granite for. Code. And that again comes in different parameter sizes, from 3 billion all the way through to 34 billion parameters. And granite for code is trained on 116 programming languages. Now there's also Granite for. Time series. That's a family of retrain models for time series forecasting. These models are trained on a collection of data sets spanning a range of business and industrial application domains. And these models are optimized to run on pretty much anything, even a laptop. And then finally, there is Granite for Geospatial, which is a partnership between NASA and IBM to create a foundation model for Earth observations using large scale. Satellite and remote sensing data O that's the IBM Granite models, models that are trusted performance and efficient and that can be applied to a wide variety of enterprise use cases.
IBM CTO ♦ Digitally-savvy and Cyber-savvy Board Director ♦ CEO Advisor ♦ Competent Boards Faculty ♦ Making Purpose Real Through Board Excellence ♦ Global Perspective, Digital Transformation, AI, Cybersecurity, ESG Expert
Selecting the right large language model is crucial for driving enterprise success. Martin Keen's insights on performance, cost, and trustworthiness resonate deeply, especially in today's fast-paced tech landscape. IBM has a long history of commitment to transparency and efficiency and now through the Granite Foundation models. Thank you, Jonathan Adashek, for sharing this valuable discussion that highlights the importance of strategic tech choices in fostering innovation and resilience.
#CEO
#KSgems
#GoodTechIBM
Watch IBM’s Martin Keen discuss the importance of selecting the right large language model (LLM) for enterprise use. This video highlights three key metrics: performance, cost-effectiveness, and trustworthiness.
Martin explores the IBM Granite Foundation models designed to meet these requirements, offering transparency, scalability, and efficiency.
Curious about how foundation models are transforming businesses? We’ve created a comprehensive megathread on Reddit, Inc. that dives deep into the family of IBM Granite models.
What you'll learn:
- Foundation models explained: Simplifying complex concepts.
- IBM Granite introduction: Exploring features and benefits.
- Selecting the right model: Tips and best practices for businesses.
Check it out: https://ibm.biz/Bda4GX
Have you heard of Granite ⚡ - IBM's flagship Foundation Model ?
What makes Granite differentiated from the fad and noise is its enterprise-level transparency, governance and performance. Its a foundation model developed by IBM Research - Trained on trusted enterprise data spanning internet, academic, code, legal and finance.
Here is an opportunity to try it out with our newly introduced IBM watsonx chat 🚀
Start chatting here: https://lnkd.in/gH7Q5f_e
Also try with 💬 -
llama-2-13b-chat (Optimised for Speed)
llama-2-70b-chat (Optimised for Accuracy)
Why are IBM Granite foundation models a great tool for automating daily tasks?
Sujatha (Suj) Perepa, Distinguished Engineer, IBM, covers the model's capabilities, including:
✅ Q&A
✅ Text generation
✅ Extraction, summarization, and classification
✅ Multi-language support
✅ Code generation
Watch below 👇
AI evangelist, advocate and enthusiast! Helping companies scale AI to create delightful customer & employee experiences! Empowering Leaders to use EI to build better teams. Inspiring people to lead more fulfilled lives!
Why are IBM Granite foundation models a great tool for automating daily tasks?
Sujatha (Suj) Perepa, Distinguished Engineer, IBM, covers the model's capabilities, including:
✅ Q&A
✅ Text generation
✅ Extraction, summarization, and classification
✅ Multi-language support
✅ Code generation
Watch below 👇
Why are IBM Granite foundation models a great tool for automating daily tasks?
Sujatha (Suj) Perepa, Distinguished Engineer, IBM, covers the model's capabilities, including:
✅ Q&A
✅ Text generation
✅ Extraction, summarization, and classification
✅ Multi-language support
✅ Code generation
Watch below 👇
Thanks to IBM for a clear explanation—few can explain where their large language model (LLM) data comes from and what’s inside.
Ansible Lightspeed, built on IBM Granite 20B models, simplifies Ansible playbook creation.
With the latest Ansible Automation Platform 2.5 release, Red Hat Ansible Lightspeed now supports on-premise deployments, allowing organizations with strict data privacy or air-gapped requirements to securely deploy both Ansible Lightspeed and IBM watsonx Code Assistant on Cloud Pak for Data, ensuring regulatory compliance while leveraging advanced AI-driven automation.
Why are IBM Granite foundation models a great tool for automating daily tasks?
Sujatha (Suj) Perepa, Distinguished Engineer, IBM, covers the model's capabilities, including:
✅ Q&A
✅ Text generation
✅ Extraction, summarization, and classification
✅ Multi-language support
✅ Code generation
Watch below 👇
Always keep in mind that fine-tuning smaller, fit-for-purpose models like Granite enables enterprises to pursue frontier model performance at a fraction of the cost.
Today we launched IBM Granite 3.0 – a set of high-performing, compact LLMs to help businesses build and deploy tailored AI solutions without compromising trust and safety. In line with IBM's larger commitment to open innovation, all the Granite models are open sourced under fully permissive Apache 2.0 licenses.
Models can be aligned to various domains and business tasks, from assistant-like chat to customer support, audience analysis, and more.
Learn more about the Granite models here: https://lnkd.in/eGw68x94
Gen AI Architect | IBM Build for Startups
1moinstead of trying to be the biggest, granite has multiplied to become a large family of small specialized models. great strategy