For our contribution to the AlLuminate benchmark, we worked with a diverse and talented group of 24 experts who came together to bring a comprehensive perspective on LLM safety in Hindi. We're super grateful to our collaborators from Nazariya: A Queer Feminist Resource Group, The Mooknayak, RATI Foundation, Newsmeter, The Alternative Story, BOOM, Prayogshala (मंक प्रयोगशाला), Vestlandsforsking, University of Delhi, National Council of Women Leaders and International Land Coalition for their invaluable input and support!
Today, we are thrilled to announce the release of AILuminate, a first-of-its kind safety test for large language models (LLMs). AILuminate is the first AI safety benchmark to have broad support from both industry and academia – and its launch represents a major milestone in progress towards a global standard on AI safety. The AILuminate benchmark was developed by the MLCommons AI Risk & Reliability working group — a team of leading AI researchers from institutions including Stanford University, Columbia University, and Eindhoven University of Technology, civil society representatives, and technical experts from Google, Intel Corporation, NVIDIA, Meta, Microsoft, Qualcomm Technologies, Inc., amongst others committed to a standardized approach to AI safety. The AILuninate benchmark delivers a comprehensive set of safety grades for today's most prevalent LLMs. Building on MLCommons’ track record of producing trusted AI performance benchmarks, the AILuminate benchmark offers a scientific, independent analysis of LLM risk that can be immediately incorporated into company decision-making. Learn more and explore the benchmark. https://lnkd.in/gBu9wm2B #AILuminate #AI #AIsafetybenchmark #LLMs