Codabench: Flexible, easy-to-use, and reproducible meta-benchmark platform

Zhen Xu*, Sergio Escalera, Adrien Pavão, Magali Richard, Wei Wei Tu, Quanming Yao, Huan Zhao, Isabelle Guyon

*Kontaktforfatter

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningpeer review

4 Citationer (Scopus)
17 Downloads (Pure)

Abstract

Obtaining a standardized benchmark of computational methods is a major issue in data-science communities. Dedicated frameworks enabling fair benchmarking in a unified environment are yet to be developed. Here, we introduce Codabench, a meta-benchmark platform that is open sourced and community driven for benchmarking algorithms or software agents versus datasets or tasks. A public instance of Codabench is open to everyone free of charge and allows benchmark organizers to fairly compare submissions under the same setting (software, hardware, data, algorithms), with custom protocols and data formats. Codabench has unique features facilitating easy organization of flexible and reproducible benchmarks, such as the possibility of reusing templates of benchmarks and supplying compute resources on demand. Codabench has been used internally and externally on various applications, receiving more than 130 users and 2,500 submissions. As illustrative use cases, we introduce four diverse benchmarks covering graph machine learning, cancer heterogeneity, clinical diagnosis, and reinforcement learning.

OriginalsprogEngelsk
Artikelnummer100543
TidsskriftPatterns
Vol/bind3
Udgave nummer7
DOI
StatusUdgivet - 8 jul. 2022
Udgivet eksterntJa

Bibliografisk note

Publisher Copyright:
© 2022 The Authors

© 2022 The Authors.

Fingeraftryk

Dyk ned i forskningsemnerne om 'Codabench: Flexible, easy-to-use, and reproducible meta-benchmark platform'. Sammen danner de et unikt fingeraftryk.

Citationsformater