ADHS'21 will feature a Repeatability Evaluation (RE). All accepted papers with computational artifacts will be invited to submit a Repeatability Evaluation Package (REP) for evaluation against ease of reproducibility and usability. The evaluation will be conducted by the RE committee who will try to reproduce the results and examine the usability. The aim of the RE is not only to find issues with reproducibility, but also to identify the effort required to reproduce the computational results.
Papers with RE packages that pass the repeatability evaluation will be listed online and in the final proceedings. Furthermore, such papers will also receive a repeatability badge that will be included on the first page of the published version. A paper with a RE package that is not able to pass the repeatability evaluation criteria, will be treated the same as a paper which did not submit a RE package.
It is optional to submit an REP, and if submitted, the evaluation result will not affect the review process of its corresponding paper. For the submission date, please see the Call for Papers. The repeatability evaluation effort at ADHS is inspired by the HSCC repeatability evaluation and borrows significantly from it.
The RE committee will review and provide feedback on the ease and quality of reproducibility of the results, which will further strengthen the quality of published papers with computational elements. The authors can use the feedback to improve the reproducibility and usability of their software. Another goal of the ADHS repeatability evaluation process is to provide authors feedback on the quality of the software.
The website for submission of the Repeatability Evaluation Package (REP) is EasyChair.
The REP should have:
We strongly recommend that the authors do a dry run on the final version of the REP. This is best done by letting someone else try installing the REP on another machine and check the accuracy of the included installation and running instructions.
The submissions will be evaluated based on the below criterion:
Reproducibility is evaluated in order to answer the question: can non-developers with the REP software and the provided instructions reproduce the claims and results in the paper (figures, tables, etc)? Specifically, this criteria checks for:
Usability addresses the ease of use and extension, and evaluates the REP on:
Like initial paper submissions, REPs are also considered confidential material. The committee members agree not to share REP contents, and to delete them after evaluation. REPs remain the property of the authors, and there is no requirement to post them publicly (although encouraged).
The repeatability evaluation process uses anonymous reviews so as to solicit honest feedback. Authors of REPs should make a genuine effort to avoid learning the identity of the reviewers. This effort may require turning off analytics, or only using systems with high enough traffic that REC accesses will not be apparent. In all cases where tracing is unavoidable the authors should provide warnings in the documentation so that reviewers can take necessary precautions to maintain anonymity.