У нас вы можете посмотреть бесплатно Community Evals: Because We’re Done Trusting Black-Box Leaderboards Over the Community или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
To address the discrepancy between saturated benchmark metrics and actual model reliability, Hugging Face has introduced "Community Evals," a decentralized framework designed to democratize and transparently report AI performance. This system enables benchmark dataset repositories to function as dynamic leaderboards that aggregate evaluation scores directly from model repositories, where results are stored in standardized YAML files adhering to Inspect AI specifications. By permitting the broader community to submit evaluation results via pull requests and maintaining a Git-based history of these contributions, the initiative establishes a verifiable and reproducible ecosystem that captures both model author and independent community data. While this open approach does not immediately resolve issues such as test-set contamination or the plateauing of scores on established tests like GSM8K, it aims to illuminate the "who, how, and when" of evaluations, fostering a more rigorous environment for developing and tracking the next generation of model capabilities. https://huggingface.co/blog/community...