Ai Model Benchmarks C-Eval A comprehensive evaluation suite designed to assess the knowledge and capabilities of large language models (LLMs) specifically in the Chinese language.
Ai Model Benchmarks SuperCLUE A professional evaluation framework providing standardized benchmarks to measure the intelligence and utility of Chinese-language AI models.
Ai Model Benchmarks Open LLM Leaderboard A comprehensive, community-driven benchmark platform by Hugging Face to track and compare the performance of open-source large language models.
Ai Model Benchmarks CMMLU A comprehensive evaluation benchmark designed to measure the general knowledge and linguistic capabilities of Large Language Models in Chinese.
Ai Model Benchmarks PubMedQA PubMedQA is a specialized biomedical question-answering dataset and leaderboard used to benchmark the accuracy of AI models in the medical domain.