Skip to Main content Skip to Navigation
Journal articles

Improving Reproducibility in Machine Learning Research (A Report from the NeurIPS 2019 Reproducibility Program)

Abstract : One of the challenges in machine learning research is to ensure that presented and published results are sound and reliable. Reproducibility, that is obtaining similar results as presented in a paper or talk, using the same code and data (when available), is a necessary step to verify the reliability of research findings. Reproducibility is also an important step to promote open and accessible research, thereby allowing the scientific community to quickly integrate new findings and convert ideas to practice. Reproducibility also promotes the use of robust experimental workflows, which potentially reduce unintentional errors. In 2019, the Neural Information Processing Systems (NeurIPS) conference, the premier international conference for research in machine learning, introduced a reproducibility program, designed to improve the standards across the community for how we conduct, communicate, and evaluate machine learning research. The program contained three components: a code submission policy, a community-wide reproducibility challenge, and the inclusion of the
Complete list of metadata

https://hal.archives-ouvertes.fr/hal-03341560
Contributor : Florence d'Alché-Buc Connect in order to contact the contributor
Submitted on : Saturday, September 11, 2021 - 8:08:07 AM
Last modification on : Tuesday, October 19, 2021 - 11:16:31 AM

File

20-303.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-03341560, version 1

Collections

Citation

Joelle Pineau, Philippe Vincent-Lamarre, Koustuv Sinha, Vincent Larivière, Alina Beygelzimer, et al.. Improving Reproducibility in Machine Learning Research (A Report from the NeurIPS 2019 Reproducibility Program). Journal of Machine Learning Research, 2021, 22. ⟨hal-03341560⟩

Share

Metrics

Record views

52

Files downloads

36