Ola Krutrim , India’s first AI unicorn, recently launched its LLM, Krutrim-7B-chat. This new model is trained in 10 Indian languages and is available on Databricks Marketplace. It can generate text in Hindi, English, Tamil, Telugu, Marathi, Malayalam, Gujarati, Bengali, and Kannada. This is the first of several India-specific models trained by Ola Krutrim. It has been trained on a massive dataset of Indic languages— “Krutrim delivers text infused with Indian sensibilities and cultural awareness.” Check out the model here. The announcement came to light at the Data Intelligence Day in Mumbai on May 10, 2024, by Gautam Bhargava, VP & Head of AI Engineering at Krutrim. “We’re excited to be at the forefront of building foundational LLMs trained on Indian languages to better serve our customers in India. We have been working very closely with the Databricks team to pre-train and finetune our foundational LLM,” said Bhargava. He added, “We are confident that Krutrim, our base model, will be able to provide high-quality responses in Indian languages and dialects.” According to Bhargava, Krutrim was trained using Databricks Mosaic AI Training’s GPU orchestration and system optimisations, enabling state-of-the-art generative AI solutions for customers globally. The model is now available on the Databricks Marketplace.
Siddharth Jindal’s Post
More Relevant Posts
-
Large Language Model Operations (LLMops) on AWS What are #LLMops? LLMops, short for #LargeLanguageModel #Operations, refer to the practice of leveraging large language models (LLMs) like GPT-3, #AnthropicClaude, #MistralAI, and others to automate various tasks and workflows. The core idea behind LLMops is to use the powerful generation capabilities of LLMs to create software applications, #APIs, and tools that can understand and generate human-like text, images, video, and audio. The purpose of LLMops is to augment and automate a wide range of generation-related tasks that were previously labor-intensive or required significant domain expertise. So, LLMops encompass all operations needed to enable and enhance the usage of LLMs. #GenerativeAI #AWS #AI https://lnkd.in/egjCKfWQ
CloudNature | Large Language Model Operations (LLMops) on AWS
cloudnature.net
To view or add a comment, sign in
-
🌟 Creating Smart Solutions: The Power of Generative AI & AWS! In the realm of tech innovation, combining cutting-edge technologies can lead to remarkable applications. A recent article explores the creation of a generative AI image description app using Anthropic's Claude 3.5 model alongside Amazon Bedrock and the AWS Cloud Development Kit (CDK). This union not only enhances automation but also fosters inclusive user experiences. 🔧 Technological Foundations: The app utilizes Claude 3.5 for natural language processing and Amazon Bedrock for accessing essential foundational models. The AWS CDK simplifies infrastructure setup, allowing developers to focus on innovation without getting bogged down in complex configurations. 🎯 Practical Applications: The app aims to auto-generate descriptive texts for images, significantly benefiting accessibility efforts, content creation, and automated tagging. Such innovations streamline workflows and cater to diverse user needs, revealing the potential for AI to transform industries. 🚀 Implementation & Scalability: From initial setup on AWS to deploying a scalable application, the article provides step-by-step guidance for developers. This ensures the application can adapt efficiently to varying loads while continuously improving user engagement. Stay Ahead in Tech! Connect with me for cutting-edge insights and knowledge sharing! Want to make your URL shorter and more trackable? Try linksgpt.com #BitIgniter #LinksGPT #GenerativeAI #AWS #CloudDevelopment
Build a generative AI image description application with Anthropic’s Claude 3.5 Sonnet on Amazon Bedrock and AWS CDK
aws.amazon.com
To view or add a comment, sign in
-
🕵️ These powerful open-source frameworks can accelerate your generative AI projects! These tools, built upon popular frameworks like LangChain and LlamaIndex, are designed to streamline the implementation process and bring your AI applications to life. Here’s a closer look at each of them: 🐦 Sparrow (Katana ML): This innovative solution excels at data extraction and processing from diverse documents and images. Whether dealing with forms, invoices, receipts, or other unstructured data sources, Sparrow’s modular architecture ensures robust performance. Its pluggable architecture allows seamless integration with tools like LlamaIndex, Haystack, or Unstructured, and supports local LLM data extraction pipelines via Ollama or Apple MLX. Sparrow's API transforms your data into structured outputs, ready for integration with custom workflows. 🔎 https://lnkd.in/eF39NxWv 🌊 LaVague: An advanced large action model framework designed to automate browser interactions seamlessly. By leveraging Natural Language Processing (NLP), LaVague interprets instructions in plain language and executes them as browser actions using Selenium. Thanks to techniques like Chain-of-Thoughts, LaVague can dynamically extract relevant HTML elements from web pages and generate Selenium code without needing to fine-tune the language model. 🔎 https://lnkd.in/eMUTFxD9 🐳 Cognita: While LangChain and LlamaIndex provide easy-to-use abstractions for quick experimentation, Cognita steps in for production environments. It offers a modular, API-driven architecture that’s easily scalable and extendable. Cognita organizes your codebase into modular RAG components and provides a production-ready environment with no-code UI support. It also supports incremental indexing by default, making it an excellent choice for both local setups and enterprise applications. 🔎 https://lnkd.in/ePUbAMbd 🤖 Agent Evaluation (Amazon Web Services (AWS)): This generative AI-powered framework is designed for testing virtual agents. It utilizes an LLM agent (evaluator) to conduct conversations with your agent (target), evaluating the responses to ensure optimal performance and reliability. 🔎https://lnkd.in/e9Jhr6hb These frameworks are among others that can transform the process of developping generative AI applications. Dive in and see how they can enhance your projects! #GenerativeAI #OpenSource #AIFrameworks #MachineLearning #DataProcessing #AI #TechInnovation #LangChain #LlamaIndex
To view or add a comment, sign in
-
Appreciate PostToday for the coverage on how Amazon Web Services (AWS) is enabling GenAI startups develop culturally aware and industry changing solutions! One such great example is BOTNOI GROUP that's has built a text-to-speech Voice bot and Chat bot using Generative AI and created new synthesized voices in Thai language with remarkable naturalness and clarity. Botnoi has used 100+ millions of Thai and other ASEAN languages audio/text sentences to train their NLP model. Amongst many other AWS services, Botnoi is powered by Amazon Connect to integrate and transcribe the voice call to the Botnoi Voicebot and response the call with Botnoi Voice with the answer or route to the next action. With this system, customers can get responses with human languages available 24/7, which can improve customer satisfaction. https://lnkd.in/gPCCbBbE Lakshmi Priya Nopawan Vince, AWS PMP® Ying Shan Ho Dan Chamberlain Eunice Cheng Vatsun Thirapatarapong Conor McNamara #genai #Aws #hereataws #startups
AWS เผย นักลงทุนมองหาสตาร์ทอัพด้านเอไอ พัฒนาโซลูชันพลิกโฉมอุตสาหกรรม
posttoday.com
To view or add a comment, sign in
-
Mistral AI Introduces Les Ministraux: Ministral 3B and Ministral 8B- Revolutionizing On-Device AI High-performance AI models that can run at the edge and on personal devices are needed to overcome the limitations of existing large-scale models. These models require significant computational resources, making them dependent on cloud environments, which poses privacy risks, increases latency, and adds costs. Additionally, cloud reliance is not suitable for offline scenarios. Introducing Ministral 3B and Ministral 8B Mistral AI recently unveiled two groundbreaking models aimed at transforming on-device and edge AI capabilities—Ministral 3B and Ministral 8B. These models, collectively known as les Ministraux, are engineered to bring powerful language modeling capabilities directly to devices, eliminating the need for cloud computing resources. With on-device AI becoming more integral in domains like healthcare, industrial automation, and consumer electronics, Mistral AI’s new offerings represent a major leap towards empowering applications that can perform advanced computations locally, securely, and more cost-effectively. These models are set to redefine how AI interacts with the physical world, offering a new level of autonomy and adaptability. Technical Details and Benefits The technical design of les Ministraux is built around striking a balance between power efficiency and performance. Ministral 3B and 8B are transformer-based language models optimized for lower power consumption without compromising on accuracy and inference capabilities. The models are named based on their respective parameter counts—3 billion and 8 billion parameters—which are notably efficient for edge environments while still being robust enough for a wide range of natural language processing tasks. Mistral AI leveraged various pruning and quantization techniques to reduce the computational load, allowing these models to be deployed on devices with limited hardware capacity, such as smartphones or embedded systems. Ministral 3B is particularly optimized for ultra-efficient on-device deployment, while Ministral 8B offers greater computational power for use cases that require more nuanced understanding and language generation. Importance and Performance Results The significance of Ministral 3B and 8B extends beyond their technical specifications. These models address key limitations in existing edge AI technology, such as the need for reduced latency and improved data privacy. By keeping data processing local, les Ministraux ensures that sensitive user data remains on the device, which is crucial for applications in fields like healthcare and finance. Preliminary benchmarks have shown impressive results—Ministral 8B, for instance, demonstrated a notable increase in task completion rates compared to existing on-device models, while maintaining efficiency. The models also allow developers to create AI applications that are less reliant on internet connectivity,...
To view or add a comment, sign in
-
🚀Unlocking enhanced LLM capabilities with RAG in Big Query🎉 ✅The rise of generative AI has opened up a world of possibilities, but it's not without its challenges. One of the biggest hurdles is creating models that can perform well in real-time and make sense of specific data. Fortunately, retrieval augmented generation (RAG) is here to help! This natural language processing technique uses a two-step process to provide more accurate and informative answers. ✅RAG simplifies data analytics by leveraging vector search in datastores like BigQuery to enhance large language model (LLM) capabilities. And the best part? You can do all of this from a single platform without moving any data, thanks to BigQuery ML! 👉👉👉Looking to improve the performance of your LLMs? Give RAG a try and see the difference it can make! #Google #LLM #RAG #BigQuery #ML https://lnkd.in/dZEZpfUT
How to use RAG in BigQuery to bolster LLMs | Google Cloud Blog
cloud.google.com
To view or add a comment, sign in
-
We just released v1.1 of our synthetic text-to-SQL dataset, the largest open source dataset of its kind, designed to accelerate #AI model training. 🚀 v1.1 addresses several suggestions we heard from developers working with the dataset. We'll continue to refine and update this dataset, so please keep the feedback coming! Blogpost: https://lnkd.in/e_PQgRc3 Dataset: https://lnkd.in/eX5e-78F #syntheticdata #llms #dataquality
Introducing world's largest synthetic open-source Text-to-SQL dataset
gretel.ai
To view or add a comment, sign in
-
Databricks invests in Mistral : Paris based AI start up the below announcement came on March 14th Databricks announced a partnership with Paris-based Mistral, the well-funded startup that’s made waves in the global AI community with its growing family of highly-performant large language models (LLMs) — many of them, open sourced. Under the engagement, Databricks is investing an undisclosed amount in Mistral, adding to its series A round, and bringing select Mistral LLMs to its data intelligence platform. Databricks confirmed that the partnership with Mistral will result in the native integration of two text-generation models from the company – Mistral 7B and Mixtral 8x7B, both open source. The former is a small 7 billion parameter transformer model, trained with 8k context length, and is very efficient to serve. Meanwhile, the latter is a sparse mixture of expert models (SMoE), supporting a context length of 32k, and capable of handling English, French, Italian, German, and Spanish. Mixtral 8x7B even outperforms Meta’s Llama 2 70B (from which it was trained on) and OpenAI’s GPT-3.5 across multiple benchmarks, including GSM-8K and MMLU, while boasting faster inference. Mistral AI models can now be consumed and customized in a variety of ways on Databricks, which offers the most comprehensive set of tools for building, testing and deploying end-to-end generative AI applications. Whether starting with a side-by-side comparison of pre-trained models or consuming models through pay-per-tokens there are several options for getting started quickly. The move will see direct integration of the models, making it easier for enterprise users to use them with their data for generative AI applications — without any change to the security, privacy, and governance the Databricks platform already offers.. Databricks says users can experiment with the models in the Mosaic AI Playground available through the platform console, use them as optimized model endpoints through Mosaic AI Model Serving or customize them with their proprietary data hosted on the platform (Mosaic AI Foundation Model Adaptation) to target a specific use case. Microsoft invested a $16 million investment from Microsoft to add its models to the Azure cloud platform. The deal made Mistral only the second company after OpenAI to offer its models on the Microsoft platform. with both Snowflake and Databricks integrating to Mistral Models makes the space heated up.
To view or add a comment, sign in
bhanudas satam