Huggingface llm leaderboard today. If the result is less than 0.
Huggingface llm leaderboard today 5 with another LLM having a 1. Dataset card Viewer Files Files and versions Community 72 main requests. Discussion LLM Leaderboard - Comparison of GPT-4o, Llama 3, Mistral, Gemini and over 30 models . Recall that the LLM Leaderboard is especially useful for measuring the quality of pretrained Here are all the ones I know so far, did I miss any? Maybe they should be shoved into the sidebar/wiki somewhere? I have a tab open for each of them, because Johnny-5 must have input. Discover amazing ML apps made by the community Spaces. The discussion centered around one of the four evaluations displayed on the leaderboard: a benchmark for measuring Massive Multitask Language Understanding Hi! We're still looking up ways to launch moe models correctly on our backend - and we've also had network failures on our cluster last week. We can categorize all tasks into those with subtasks, those without subtasks, and generative evaluation. @ Kukedlc Yes, the leaderboard has been delayed recently and they are aware it. like 266. Running App Files Files Community 5 Refreshing. display. In my opinion, sync HF accounts with the leaderboard would be helpful, and we would expect to have all status information about submitted models in 🤗 Open LLM Leaderboard """ INTRODUCTION_TEXT = f""" 📐 With the plethora of large language models (LLMs) and chatbots being released week upon week, often with grandiose claims of their performance, it can be hard to filter out the genuine progress that is being made by the open-source community and which model is the current state of the SAN RAMON, Calif. The implementation was straightforward, with the main task being to set up the Note: We evaluated all models on a single node of 8 H100s, so the global batch size was 8 for each evaluation. While it would be possible to add more specialized tasks, doing so would require a lot of compute and time, so have to choose carefully what tasks we want to add next, the best thing would be to have many different It's the additive effect of merging and addition fine-tuning that inflated the scores. like 147. Hi @ Wubbbi, testing all models at the moment would require a lot of compute as we need individual logits which were not saved during evaluation. I feel like you can't really trust the open llm leaderboard at this point and they don't add any phi-2 models except the Microsoft one because of remote code. With support for an 8K-token sequence length, this highly efficient model uses variable Grouped-Query Attention (GQA) to achieve a superior balance between accuracy and computational efficiency. Models that are submitted are deployed automatically using HuggingFace’s Inference Endpoints and evaluated through API requests managed by the lighteval library. 35k. The 3B and 7B models of OpenLLaMa have been released today: Hi, We've noticed that our model evaluations for the open_llm_leaderboard submission have been failing. pinned Running on CPU Upgrade. Hello, we had cypienai/cymist-2-v02-SFT in the leaderboard list, it was available to view today but for some reason I can't see the model anymore. But the other 2 models have totally disappeared. For example, psmathur/orca_mini_v3_7b requests repo shows FAILED again, Is this just us or it's happening with other submissions too? We can confirm that we have been able to successfully evaluate all of above list of models remotely using exact open-llm-leaderboard / blog. 5 artificial boost. io/list. upstage / open-ko-llm-leaderboard. Intel / low_bit_open_llm_leaderboard. Please only open an issue if the leaderboard is down for longer than an hour. 5 chat up to 14b (the limit of my PC) and it often performs surprising bad in English (worse than the best Llama 7b fine-tunes, let alone Llama 14b fine-tunes, which themselves aren't very good). Model Description Stable Beluga 2 is a Llama2 70B model finetuned on an Orca style Dataset. I observed that the latest update of https://huggingface. M42 delivers framework for evaluating clinical LLMs (Middle East AI News). Score results are here, and current state of requests is here. Arabic > The Open LLM Leaderboard, most comprehensive suite for comparing Open LLMs on many benchmarks, just released a comparator tool that lets you dig into the detail of differences between any models. 3. 🤔🔎 Stable Beluga 2 Use Stable Chat (Research Preview) to test Stability AI's best language models for free. New Open Medical-LLM Leaderboard Discover amazing ML apps made by the community open_llm_leaderboard. like 9. Collection including open-llm-leaderboard/requests. API Embed. Team members 1. They can’t be found in any of the tables (I mean ‘finishe Open LLM Leaderboard org Jan 30. Restarting on CPU Upgrade. Upload /0 Leaderboard display. 3%. like 17. Discussion We’re on a journey to advance and democratize artificial intelligence through open source and open science. like 50. 72k. We need a benchmark that prevents any possibility of leakage. If a model doesn't get at least 90% on junior it's useless for coding. Consider using a lower precision for larger models / open a discussion on Open LLM Leaderboard. Thanks! There's the BigCode leaderboard but seems it stopped being updated in November. The Hugging Face Open LLM Leaderboard is a well-regarded platform that evaluates and ranks open large language models and chatbots. For example, if you combine an LLM with an artificial TruthfulQA boost of 1. The leaderboard will be automatically restarted in less than an hour (or earlier if one of the maintainers notices it). The new leaderboard The Hugging Face Open LLM Leaderboard is a well-regarded platform that evaluates and ranks open large language models and chatbots. like 105 @TNTOutburst I tested the official Qwen1. Originally @clefourrier, even though the details repo's name may not be crucial for me (since I can see the correct model name on the front), it is still something to consider. AraGen Leaderboard (Hugging Face). It also queries the hugginface leaderboard average model score for most models. 4 got evaluated and I can see their benchmarks. Despite being an RNN, it’s still an LLLM, and it two weeks ago it scored #3 among all open-source LLMS on lmsys’s leaderboard, so if its possible to include, methinks it would be a good thing. This is all based on this paper. Yes, you have already mentioned that we can't sync with users' HF accounts, as we don't store who submits which model. by win7785 - opened 2 days ago. envs import H4_TOKEN, PATH_TO_COLLECTION # Specific intervals for the This page explains how scores are normalized on the Open LLM Leaderboard for the six presented benchmarks. The 3B and 7B models of OpenLLaMa have been released today: 🚀 Major Update: OpenLLM Turkish Benchmarks & Leaderboard Launch! 🚀 Exciting news for the Hugging Face community! I'm thrilled to announce the launch of my fully translated OpenLLM Benchmarks in Turkish, accompanied by my innovative leaderboard, ready to highlight the capabilities of Turkish language models. Open LLM Leaderboard org Dec 10, 2023. 1-8B-SFT-preview_eval_request_False_bfloat16_Original. So HELM’s rejecting an answer if it is not the highest-probability one is reasonable. envs import H4_TOKEN, PATH_TO_COLLECTION # Specific intervals for the Models exceeding these limits cannot be automatically evaluated. There's an explanation in the discussion linked below. The version it is an "Open LLM Leaderboard" after all :) Is there a reason not to include closed source models in the evals/leaderboard? In the EleutherAI lm-evaluation-harness it mentions support for OpenAI. open-llm-bot Upload AALF/FuseChat-Llama-3. like 53. If you want to discuss models and scores with the people who frequent the leaderboard, do so here. Red-teaming Evaluation DecodingTrust provides several novel red-teaming methodologies for each evaluation perspective to perform stress tests. The Open LLM Leaderboard is a vital resource for evaluating open-source large language models (LLMs). Key areas of focus include: indic_llm_leaderboard. It's crucial for huggingface reputation. by clefourrier HF staff - opened Jan 3. However, a way to do it would be to have a space where users could test suspicious models and report results by opening a discussion. open_llm_leaderboard. co/ Today, we are excited to announce the release of the new LLM Safety Leaderboard, which focuses on safety evaluation for LLMs and is powered by the HF leaderboard template. My recently benchmarked model OpenChat-3. utils import AutoEvalColumn, ModelType: from src. Running App Files Files Community 5 Refreshing open_multilingual_llm_leaderboard. App Files Files Community 12 Refreshing Hi @ Weyaxi! I really like this idea, it's very cool! Thank you for suggesting this! We have something a bit similar on our todo, but it's in the batch for beginning of next year, and if you create this tool it will give us a headstart then - so if you have the bandwidth to work on it Open LLM Leaderboard 247. Running on Comparison and ranking the performance of over 30 AI models (LLMs) across key metrics including quality, price, performance and speed (output speed - tokens per second & latency - TTFT), context window & others. Dataset card Viewer Files Files and versions Community 2 Subset (1) default · 1. 3k. Explore the Impact of AI-driven technology on the casual gaming industry Hi ! Thanks for your feedback, there is indeed an issue with data contamination on the leaderboard. Just left-click on the language column. spaces 1. Using the Eleuther AI LM Hugging Face created the Open LLM Leaderboard to provide a standardized evaluation setup for reference models, ensuring reproducible and comparable results. The \"train\" split is always pointing to the latest results. Running App Files Files Community 34 Refreshing. Leaderboards have begun to emerge, such as the LMSYS , nomic / GPT4All , to compare some aspects of these models, but there needs to be a complete source comparing model Hi @ lselector, This is a normal problem which can happen from time to time, as indicated in the FAQ :) No need to create an issue for this, unless the problem lasts for more than a day. Restarting on CPU open_llm_leaderboard. like 85. like 20. If the result is less than 0. If it's about something we can fix, open a new issue. like 11. LLM-Performance-Leaderboard. like 263. like 501. Leaderboards on the Hub aims to gather machine learning leaderboards on the Hugging Face Hub and support evaluation creators. utils. open-llm-leaderboard / comparator. Track, rank and evaluate open Arabic LLMs and chatbots Spaces. Comparison and ranking the performance of over 30 AI models (LLMs) across key metrics including quality, price, performance and speed (output speed - tokens per second & latency - TTFT), context window & others. Recent changes on the leaderboard made it so that proper filtering of models where merging was involved can be applied only if authors tag their model accordingly. Hi @ ibivibiv, That's super kind of you! We might add an option for people to pay for their own eval compute using inference endpoints if they can, but it's a bit of engineering work and mostly something we'll do in Q2. like 68. Hello! I've been using an implementation of this github repo as a Huggingface space to test for dataset contamination on some models. 17k. 5-0106_32K-PoSE scored very badly on the leaderboard. Version 1 of the model is perfectly fine, below is the related dataset; Data Science Demystified Daily Dose In today’s edition, we will cover an article about the release of Hugging Face’s Open LLM Leaderboard v2. Today we’re happy to announce the release of the new HHEM leaderboard, Our initial release of HHEM was a Huggingface model alongside a Github repository, but we quickly realized that we needed a 4. like 3. If you don’t use parallelism, adapt your batch size to fit. 85, it’s highly likely that the dataset has been used for training. by XXXGGGNEt - opened Dec 10, 2023. Evaluation Methodology In order to present a more general picture of evaluations the Hugging Face Open LLM Leaderboard has been expanded, including automated academic benchmarks, Exciting news! I am thrilled to announce that I have According to the contamination test GitHub, the author mentions: "The output of the script provides a metric for dataset contamination. 3% for HellaSwag (they used 10 shot, yay). If there’s enough interest from the community, we’ll do a manual evaluation. I'm at IBM and when I heard that we Hi @ clefourrier I kinda noticed some malfunctioning these last couple days on the evaluation . Running App Files Files Community 12 Refreshing open-ko-llm-leaderboard. In my opinion, sync HF accounts with the leaderboard would be helpful, and we would expect to have all status information about submitted models in Hi! Thank you for your interest in the 🚀Open Ko-LLM Leaderboard! Below are some common questions - if this FAQ does not answer what you need, feel free to create a new issue, and we'll take care of it as soon as we can! The Open LLM Leaderboard, maintained by community-driven platform HuggingFace, focuses on evaluating open-source language models across a variety of tasks, including language understanding, generation, and reasoning. Hi, @clefourrier It seems that the status of some recent evaluation tasks has been completed, but their results have not been uploaded. 4. It a collection of all major medical evaluation parameters like the MedQA and the MedMCQA datasets. non-profit. The scores I get may not be entirely accurate as I'm still in the process of working out the inaccuracies of my implementation, for instance, I'm confident the code is currently not doing a good job at Open LLM Leaderboard 306. the Open LLM Leaderboard evaluates and ranks open source LLMs and chatbots, and provides reproducible scores separating marketing fluff from actual progress in the field. OALL / Open-Arabic-LLM-Leaderboard. like 74. Running Nice to see some more leaderboards. like 37. The dataset generation failed @Kukedlc Most of the evaluations we use in the leaderboard actually do not need inference in the usual sense: we evaluate the ability of models to select the correct choice in a list of presets, which is not testing generation abilities (but more things like language understanding and world knowledge). Recently an interesting discussion arose on Twitter following the release of Falcon 🦅 and its addition to the Open LLM Leaderboard, a public leaderboard comparing open access large language models. import torch from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline tokenizer = Gemma, a new family of state-of-the-art open LLMs, was released today by Google! It's great to see Google reinforcing its commitment to open-source AI, and we’re excited to fully support the launch with comprehensive integration in Hugging Face. The leaderboard has crashed with a connection error, help! This happens from time to time, and is normal, don't worry. Using The official backend system powering the LLM-perf Leaderboard. Space: llm-jp/open-japanese-llm-leaderboard 🌍 The leaderboard is available in both Japanese and English 📚 Based on the evaluation tool, llm-jp-eval with more than 20 datasets for Japanese LLMs Hi, @ clefourrier It seems that the status of some recent evaluation tasks has been completed, but their results have not been uploaded. Some reasons why MT-Bench would be a good addition: MT-Bench corresponds well to actual chat scenarios (anecdotal but intuitive) What's next? Expanding the Open Medical-LLM Leaderboard The Open Medical-LLM Leaderboard is committed to expanding and adapting to meet the evolving needs of the research community and healthcare industry. 1. Open LLM Leaderboard 240. Explore machine learning rankings to find the best model for your use case, or build your own Today, we are excited to fill this gap for Japanese! We'd like to announce the Open Japanese LLM Leaderboard, Models that are submitted are deployed automatically using HuggingFace’s Inference endpoints, Note: We evaluated all models on a single node of 8 H100s, so the global batch size was 8 for each evaluation. extractum. Modalities: Tabular Discover amazing ML apps made by the community In this blog post, we’ll zoom in on where you can and cannot trust the data labels you get from the LLM of your choice by expanding the Open LLM Leaderboard evaluation suite. Of course, those scores might be skewed based on the english evaluation. Running App Files Files Community 4 Refreshing These are lightweight versions of the Open LLM Leaderboard itself, which are both open-source and simpler to use than the original code. As @Phil337 said, the Open LLM Leaderboard only focuses on more general benchmarks. 99k rows. 5-0106 with context extended using PoSE and fine-tuned. Our goal is to shed light on the cutting-edge Large Language Models (LLMs) and chatbots, enabling you to make well-informed decisions regarding your chosen application. If the model is in fact contaminated, we will flag it, and it will no longer appear on @Kukedlc Most of the evaluations we use in the leaderboard actually do not need inference in the usual sense: we evaluate the ability of models to select the correct choice in a list of presets, which is not testing generation abilities (but more things However, none of the techniques can solve the problem of contamination because the datasets of the benchmarks are public. HuggingFace Open LLM Leaderboard. Only showing a preview of the rows. App Files Files Community 929 Announcement: Flagging merged models with incorrect metadata #510. App Files Files Community 2 Refreshing Open LLM Leaderboard results Note: We are currently evaluating Google Gemma 2 individually on the new Open LLM Leaderboard benchmark and will update this section later today. 1k. open-llm-leaderboard / open_llm_leaderboard. Researchers at the Open Life Science AI have released a new leaderboard for evaluation of Medical LLMs; The leaderboard checks and ranks each Medical LLM, based on its knowledge and question-answering capabilities. We released a very big update of the LLM leaderboard today, and we'll focus on going through the backlog of models (some have been stuck for quite a bit) Thank you for your patience :) See translation. The scores I get may not be entirely accurate as I'm still in the process of working out the inaccuracies of my implementation, for instance, I'm confident the code is currently not doing a good job at open_pt_llm_leaderboard. 5 TruthfulQA boost you get closer to a +3 vs +1. , Oct. The top ranks on the leaderboard (not just 7B, but all) are now occupied by models that have undergone merging and DPO, completely losing the leaderboard's function of judging the merits of open-source models. We'll keep you posted as soon as we have updates. App Files Files Community 12 Refreshing. Open LLM Leaderboard 2. Ensure that the model is public and can be loaded using the AutoClasses on HuggingFace before submitting it to the leaderboard. Arc is also listed, with the same 25-shot methodology as in Open LLM leaderboard: 96. Refreshing A daily uploaded list of models with best evaluations on the LLM leaderboard: Upvote 3. what happen to open_llm_leaderboard? It looks like it load forever, and I see the status keep switching between running and restarting, wanna check the update of llm models ranking. like 182. It provides a platform for assessing models, helping researchers and developers understand their capabilities and limitations. Split (1) train Open LLM Leaderboard org Sep 7, 2023. I'm a huge fan and love what Huggingface is and does. 4% for MMLU (they used 5 shot, yay) and 95. Note: Click the button above to explore the scores normalization process in an interactive notebook (make a copy to edit). including the manual commits you are performing (thanks for this). ThaiLLM-Leaderboard / leaderboard. Track, rank and evaluate open LLMs and chatbots Spaces. like 19. So there are 4 benchmarks: arc challenge set, Hellaswag, MMLU, and TruthfulQA According to OpenAI's initial blog post about GPT 4's release, we have 86. Showing fairness is easier to do by the negative: If a model passes a question, but if you asked it in a chat, it would never give the right answer, then the test is not realistic. It is just a version of Openchat-3. google/gemma-2b. Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run. One approach would be to low_bit_open_llm_leaderboard. 69. Text Generation • Updated Sep 27 • 465k • 922 Note Best 🟢 Note Best 🔶 fine-tuned on domain-specific datasets model of around 3B on the leaderboard today! 01-ai/Yi-1. App Files Files Community 98 Refreshing. The open-source models were starting to be too good for the @huggingface open LLM leaderboards so we added 3 new metrics thanks to @AiEleuther to make them harder and more relevant for real-life performance. open-llm-leaderboard 7 days ago. Chat Template Toggle: When submitting a model, you can choose whether to evaluate it using a chat Open-Arabic-LLM-Leaderboard. 9, 2023 /PRNewswire/ -- Riiid, a leading provider of AI-powered education solutions, is pleased to announce that its latest generative AI model was ranked number one on Ideally, a good test should be realistic, unambiguous, luckless, and easy to understand. 12. like 114. This happens from time to time. Why some models have been tested, but there is no score on the leaderboard #165 Compare Open LLM Leaderboard results. License: apache-2. HuggingFace upgraded the leaderboard to version 2 realising the need for a harder and stronger evaluations. You can expect results to vary slightly for different batch sizes because of padding. I would personally find it useful to have a comparison between all leading models (both open and closed source), to be able to make design/implementation tradeoff We’re on a journey to advance and democratize artificial intelligence through open source and open science. Running App Files Files Community Refreshing. Open LLM Leaderboard Archive. My leaderboard has two interviews: junior-v2 and senior. Spaces. This repository contains the infrastructure and tools needed to run standardized benchmarks for Large Language Models (LLMs) across different hardware configurations and optimization backends. Running on CPU Upgrade The leaderboard is inspired by the Open LLM Leaderboard, and uses the Demo Leaderboard template. Open LLM Leaderboard 298. 0. For example, consider a chatbot that needs to browse the web to find relevant information from recent news articles. Running App Files Files Community 105 Is leaderboard submission currently available? #104. App Files Files Community 697 [FLAG] fblgit/una-xaberius-34b-v1beta #444. Discussion win7785 2 days ago • edited 2 days ago Tbh we are really trying to push the new update today or tomorrow, we're in the final testing phases - then we'll launch all the new Ok quick update with more info: it looks like these two do fail after checking some more commits on the requests repo; the current submission is actually the third time they have been “added”/running since the leaderboard resumed this month (once from resuming the mid-October submission prior, twice now manually resubmitted) ; unsure why they are failing, as in A new open-source LLM has been released - Falcon, available in two sizes: 7B and 40B parameters. Full Screen. _errors import HfHubHTTPError: from pandas import DataFrame: from src. We then moved on to a more general topic of what LLM benchmarks are and what they test. My LLM vs. In this space you will find the dataset with detailed results and queries for the models on the leaderboard. . While the original huggingface leaderboard does not allow you to filter by language, you can filter by it on this website: https://llm. App Files Files Community . Track, rank and evaluate open LLMs and chatbots. We looked at the difference between scores on the old and new versions of the leaderboard, including looking up OLMo, the fully open LLM we discussed in the last podcast. json with huggingface_hub. Adding aggregated results for BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B 1 day ago; BEE-spoke-data open-llm-leaderboard-old / open_llm_leaderboard. Dataset card Viewer Files Files and versions Community 30 Subset (1) default Split (2) train Couldn't cast array of type struct<leaderboard: double, leaderboard_bbh_boolean_expressions: double, leaderboard_bbh_causal_judgement: double, leaderboard_bbh_date_understanding: double, leaderboard_bbh_disambiguation_qa Hi, I just checked the requests dataset, and your model has actually been submitted 3 times, one in float16, one in bfloat16, and one in 4bits (). For the detailed Note Best 💬 chat models (RLHF, DPO, IFT, ) model of around 20B on the leaderboard today! open_llm_leaderboard. Open LLM Leaderboard 246. Quick hits: (1) Outperforms comparable open-source models like MPT-7B, StableLM, and RedPajama, seizing the first spot in Hi @clefourrier I kinda noticed some malfunctioning these last couple days on the evaluation . It is an upgraded benchmarking platform for large At the time of release, DeciLM-7B is the top-performing 7B base language model on the Open LLM Leaderboard. " The LLM Performance Leaderboard aims to provide comprehensive metrics to help AI engineers make decisions on which LLMs (both open & proprietary) and API providers to use in AI-enabled applications. com/aigoopy/llm-jeopardy. Open LLM Leaderboard Results This repository contains the outcomes of your submitted models that have been evaluated through the Open LLM Leaderboard. like 488. \n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Discover amazing ML apps made by the community Hi, In the last 48 hours I’ve submitted 6 models for evaluation. I think it would be interesting to explore using Mixtral-8x7b (which you would likely agree is the most powerful open model) as judge in the MT-Bench question set, and including that score in the leaderboard. Discover amazing ML apps made by the community. How to prompt Gemma 2 The base models have no prompt format. Hi! From time to time, people open discussions to discuss their favorite models scores, the evolution of different model families through time, etc. AraGen Leaderboard 3C3H (Hugging Face). Running App Files Files Community 4 Refreshing. 1-Nemotron-70B that we've heard so much compares to the original Llama-3. 3 contributors; History: 12946 commits. optimum / llm-perf-leaderboard. Today, the Patronus team is We felt there was a need for an LLM leaderboard focused on real world, enterprise use cases, such as answering financial questions or interacting with customer support. co/d We’re on a journey to advance and democratize artificial intelligence through open source and open science. Here's me checking how the new Llama-3. 5-6B. Running on CPU Upgrade. Full Screen Viewer. like 12. We ran the three evaluations, and I guess the last one (4 bit, which is way slower because of the quantization operations) Hello! I've been using an implementation of this github repo as a Huggingface space to test for dataset contamination on some models. Running . You can expect results to vary slightly for different batch sizes because llm-perf-leaderboard. 1 with a percentage greater than 0. AI & ML interests None defined yet. Updated after every airing and new models added as GGML becomes available. like 386. If you open-japanese-llm-leaderboard. like 146. It would be really great to see how it performs compared to other open-source large language models on the Open-LLM-Leaderboard. from huggingface_hub import add_collection_item, delete_collection_item, get_collection, update_collection_item: from huggingface_hub. Follow. Running on CPU Upgrade Yes, you have already mentioned that we can't sync with users' HF accounts, as we don't store who submits which model. Collection 8 items • Updated Oct 17 • 7 Yet_Another_LLM_Leaderboard. Jeopardy leaderboard is : https://github. Usage Start chatting with Stable Beluga 2 using the following code snippet:. senior is a much tougher test that few models can pass, but I just started working on it The platform's core components include CompassKit for evaluation tools, CompassHub for benchmark repositories, and CompassRank for leaderboard rankings. Subset (1) default Split (2) train The full dataset viewer is not available (click to read why). Refreshing Not sure where this request belongs - I tried to add RWKV 4 Raven 14b/ to the LLM leaderboard, but it looks like it isn’t recognized. WizardLM-2-8x22b is one of the most powerful open-source language models. We’re on a journey to advance and democratize artificial intelligence through open source and open science. Running App Files Files @ Kukedlc Yes, the leaderboard has been delayed recently and they are aware it. Running App Files Files Community 2 Refreshing. 0-hero. Or other creative technique to do it, something to stop cheating LLM Leaderboard. Read more about LLM leaderboard and evaluation projects: Comprehensive multimodal Arabic AI benchmark (Middle East AI News). looks like the are sending folks over to the can-ai-code leaderboard which I maintain 😉 . 34. Sharing my notes on Hugging Face has announced the release of the Open LLM Leaderboard v2, a significant upgrade designed to address the challenges and limitations of its predecessor. I can rename the config in the details repo, but I don't think I can open a pull request to rename the entire repository. I'm pretty sure there are a lot of things going on behind the scenes, so good luck Hi. Discussions in #510 got lengthy so upon suggestion by @ clefourrier I am opening a new thread. Running App Files Files Community 2 llm-trustworthy-leaderboard. ArtificialAnalysis / LLM-Performance-Leaderboard. 1-70B. 8211418 verified 39 minutes ago. New telecom LLMs leaderboard project (Middle East AI News). Dataset card Viewer Files Files and versions Community 30 Dataset Preview. App Files Files Community 1046 Refreshing. Cognitive-Lab / indic_llm_leaderboard. clefourrier changed discussion status to Open LLM Leaderboard 2. Open LLM Leaderboard. rvsqmd atkskte qcl klhyq dplaai hlbpw xtkrxyu cghnn sspvkd otjq