AGI-Eval is a specialized evaluation community designed to benchmark the capabilities and performance of various AI large language models.
An advanced evaluation system by H2O.ai that utilizes Elo rating methodologies to benchmark and rank Large Language Models (LLMs).