AI 모델 벤치마크 매직아레나 매직아레나 is a competitive benchmarking platform designed to evaluate and rank visual generative AI models through side-by-side human comparison.
AI 모델 벤치마크 MMBench MMBench is a comprehensive evaluation framework designed to measure the capabilities of multimodal large language models across a wide array of visual and textual tasks.
AI 모델 벤치마크 Open LLM Leaderboard A comprehensive, community-driven benchmark platform by Hugging Face to track and compare the performance of open-source large language models.
AI 모델 Scale AI A comprehensive data engine for AI 개발, specializing in high-quality data labeling, RLHF, and model evaluation for enterprise machine learning.