Skip to main content

AD/AE Process & Badges

This committee’s objective is to foster and advance reproducible research in the SC community. Consequently, our aim is to aid SC authors in submitting the required documentation describing their artifact, facilitating our evaluation and subsequent badge assignment for the artifact.


Reproducibility Chair

Sascha Hunold, Technical University of Vienna

Reproducibility Vice Chair
Le Mai Weakley, Indiana University Bloomington

Reproducibility Vice Chair
Emmanuel Agullo, INRIA Bordeaux – Sud-Ouest

Getting Started

Why is reproducibility important?

You will be making it easy for other researchers to compare with your work, to adopt and extend your research. This instantly means more recognition directly visible through badges for your work and higher impact. As described in this SC20 survey, thirty-five percent (35%) of the respondents have used the appendices information from papers in the SC conference proceedings in their research.

What Is An Artifact/Research Object?

We use the terms artifact and research object as synonyms. A paper consists of several computational artifacts that extend beyond the submitted article itself: software, datasets, environment configuration, mechanized proofs, benchmarks, test suites with scripts, etc. A complete artifact package must contain (1) the computational artifacts, and (2) instructions/documentation describing the contents and how to use it. Further details and guidance about artifacts and research objects are provided all through this page.

AD/AE Process

The review process will take place in two stages: The first stage is mandatory and involves providing an Artifact Description (AD) that will be checked for completion and accessibility. The second stage, the Artifact Evaluation (AE), is optional and will be carried out for accepted papers that applied for badges.

Stage 1: Artifact Description (AD) – Mandatory

The artifact description provides information about a paper’s artifacts. All SC papers must provide (1) an artifact description or (2) a reason why an artifact description can not be provided. The committee may provide feedback to authors in a single-anonymous arrangement. The AD/AE Committee will not share any information with the Program Committee other than to confirm whether artifacts meet the criteria.

APR 20, 2024: Submission of AD Appendix – mandatory

Stage 2: Artifact Evaluation (AE) – optional

Based on the reviews of the AD, an updated version of the AD needs to be submitted. At this phase, it is optional to submit the Artifact Evaluation, and all the computational artifacts required for the reproducibility of experiments (e.g., any code / data artifacts used). When the AE is submitted, authors need to apply for any of the reproducibility badges available. Then, the AD/AE committee starts to evaluate the artifact. This step relies on cooperation between paper authors and the committee. We will use the SC Conference Submission System for messaging between the two parties. Via this communication, the committee may ask for access to special hardware, ask for missing components, provide failure/error messages, and generally keep the author’s posted on their evaluation status. Phase 2 finishes with the artifact freeze. Authors need to assign a DOI (What is a DOI?) to their artifacts to guarantee no further modifications are possible to their artifacts.

JUN 28, 2024: Submission of Revised AD Appendix and AE Appendix + Application for Badges.

AUG 16, 2024: Artifact freeze, authors assign a DOI to their Artifact. Artifact freeze means that the artifact must not be changed after this time.

AUG 23, 2024: Artifact badge decision.


AD/AE Submissions

Stage 1: Artifacts are submitted as PDF via the “Artifact Description” submission form. 

Stage 2: Artifacts are submitted as PDF via the “Badges” submission form. 

Details on how to compile the AD/AE Appendices, including LaTeX templates, can be found on the AD/AE Appendices page.


AE Report

The artifact reviewers will provide a report that details the Artifact Evaluation and includes a justification for awarding each badge. These reports will accompany the respective papers in the conference proceedings.

AD/AE Organization & Evaluation

Artifact Decription (AD) – Mandatory

The Artifact Description is a mandatory step for all submitted papers. All authors must provide descriptions of their artifacts. 

The Artifact Identification summarizes the main contributions of the article and details which role each computational artifact(s) plays for supporting these contributions. A paper typically makes n distinct contributions. In the Artifact Identification, restate these contributions and denote them as C1, C2,…, Cn. Each contribution shall be briefly described to elucidate its scientific or technical significance. Concurrent to the article’s contributions, there exist m computational artifacts denoted as A1, A2,…, Am. Each artifact shall be clearly described. For each contribution Ci, specify which computational artifact(s) Aj serves to substantiate or validate Ci. Clearly and explicitly delineate the link between contributions and their supporting artifacts.

An AD Appendix should provide all necessary information (e.g., input datasets, links to source code, collection of software dependencies, hardware dependencies, etc.) that are needed to reproduce the listed computational artifacts.

If you are unable to provide an artifact description (e.g., due to proprietary reasons), please provide detailed reasons why you are not able to do so. Failure to provide detailed description will lead to further questions from the AD/AE Committee. The AD/AE Committee will provide feedback to the SC Technical Program Committee, and inadequate explanations will have a negative effect on the paper’s overall review.

Artifact Evaluation (AE) – Optional

The Artifact Evaluation is an optional step for all accepted papers. It allows authors to apply for a reproducibility badge. If you wish to acquire badges for your paper, you must choose the requested badges in AD/AE form. The Artifact Evaluation (AE) will be included after the AD in the same appendix. If a paper only provides an AD Appendix, authors may still request the ARTIFACTS AVAILABLE badge. However, all badge requests are handled in the 2nd stage of the AD/AE Process.

The AE Appendix complements the AD Appendix by enriching the meta information provided in the AD Appendix with concrete instructions on how to reproduce the computational results shown in the paper. Therefore, the AD Appendix can be considered as a necessary checklist for providing all required information. The AE Appendix, on the contrary, contains the step-by-step instructions on how to reproduce the results that support the main finding of the paper.

Guidelines for artifact providers

Please carefully read the guidelines when compiling the AE Appendix.

Important Disclaimer

The Reproducibility Initiative works in a best effort manner. Please note that personnel and computational resources are limited. Sometimes it is simply not possible to run a given experiment at a certain time, due to unavailability of specialized hardware such as GPUs during the artifact evaluation period, unpredictable maintenance times of compute centers, or individual time constraints of reviewers (time zones, etc.). Nonetheless, the Reproducibility Initiative attempts to fairly judge the computational artifacts of SC papers that demand reproducibility badges.

AE Time Budget

Each AE Appendix is allocated an 8-hour review budget. This means artifact reviewers are expected to dedicate this amount of time to result reproduction. When preparing their computational artifacts, artifact providers should be mindful of this time constraint. It is crucial to note that the 8-hour budget pertains to the net time required for reproducing results. For instance, if artifact reviewers have to wait several hours for access to a shared hardware infrastructure, like Chameleon Cloud, these waiting periods are not counted within the allocated budget.

Computational Resources

In order to evaluate the computational artifacts provided by the authors, reviewers rely on the availability of computational resources, such as large-scale compute clusters or dedicated GPUs. 

Chameleon Cloud supports the SC24 Reproducibility Initiative by providing computational resources for authors and reviewers. 

In principal, there are three distinct scenarios for the AE:

  1. (Preferred) The authors (artifact providers) create the computational artifacts in a downscaled manner on Chameleon Cloud, e.g., use less compute nodes or simplified datasets/instances to reduce reproduction times. Now, authors and reviewers will have access to the same computational resources, and thus, an artifact evaluation has the highest chance of being successful. It is important to note that while a downscaled experiment, i.e., reducing the number of resources and computational time required, allows reviewers to assess the functionality of an artifact, a RESULTS REPLICATED badge cannot be granted if the principal results in the paper were achieved from experiments requiring an order of magnitude greater in resources and time.
  2. Computational artifacts cannot or should not be executed on Chameleon Cloud, as specialized hardware or software is required. Now, there are two sub-cases:
    (a) The authors can provide access to the specific computational resources for the reviewers. In this case, reviewers should be actively assisted to get accounts on the respective machines in a timely manner.
    (b) The authors do not or cannot provide access to computational resources needed for the artifact evaluation. Under these circumstances, the artifact reviewers will attempt to meet the requirements. However, the likelihood of obtaining the RESULTS REPLICATED badge will be diminished.
  3. (Optional) This year’s initiative proposes the optional use of Guix, a software tool designed to support reproducible software deployment. Guix allows deployment of the exact same software environment on different machines and at different points in time, while still retaining complete provenance info. By eliminating almost entirely variability induced by the software environment, Guix gives authors and reviewers more confidence into the results of computational experiments. Interested authors will find more information at (AE) Alternative with Guix.

Please note that deployments on commercial cloud infrastructures or any other platform requiring reviewers to incur monetary costs for reproducing experiments are not allowed, as such practices can create barriers for potential reviewers and for future third-party researchers interested in the artifacts. If you submit artifacts that violate this requirement, the only badge you may apply for is ARTIFACTS AVAILABLE.

Granularity of AE Description

Since public, open software and data repositories such as Zenodo, DataPort, Dryad, FigShare, or Harvard Dataverse are explicitly encouraged to be used for archiving artifacts, we would like to clarify where to put the information on how to build and execute experiments, i.e., should the information be put into the AE Appendix or into READMEs on data repositories. 

We recommend that authors put all instructions required to reproduce experiments in the AE Appendix, i.e., simply referring to READMEs on external repositories such as Zenodo should be avoided.

For the sake of readability, the detailed nature of build and run instructions might be very excessive for an AE Appendix. Please think about the right level of granularity. For example, it may suffice to state that a script like “compile.sh” needs to be called on machine X, while “compile.sh” itself may contain hundreds of individual steps.

If authors would like to provide a README in the data repositories (e.g., Zenodo), we recommend to provide only one single, comprehensive top-level README file, which should encompass all necessary information that needs to be conveyed.

Computational Artifacts

The computational artifacts of a paper include all the elements that support the reproducibility of its experiments, such as software, datasets, environment configuration, mechanized proofs, benchmarks, test suites with scripts, etc. Authors can choose any version-controlled software and data repositories to share their artifacts, such as Zenodo, FigShare, Dryad, Software Heritage, GitHub, or GitLab.

The AD/AE, in addition to documenting the computational artifacts, will also include links to the required repositories. If needed, README files can also be attached to the computational artifacts, either containing the same information as in the AD/AE or complementing it, for instance, providing further and more detailed instructions and documentation. As a general rule, authors should try to do their best to simplify the reproducibility process, to save committee members the burden of reverse-engineering the authors’ intentions. For example, a tool without a quick tutorial is generally very difficult to use. Similarly, a dataset is useless without some explanation on how to browse the data. For software artifacts, the AD/AE and the README should—at a minimum—provide instructions for installing and running the software on relevant inputs. For other types of artifacts, describe your artifact and detail how to “use” it in a meaningful way.

Importantly, make your claims about your artifacts concrete. This is especially important if you think that these claims differ from the expectations set up by your paper. The AD/AE Committee is still going to evaluate your artifacts relative to your paper, but your explanation can help to set expectations up front, especially in cases that might frustrate the evaluators without prior notice. For example, tell the AD/AE Committee about difficulties they might encounter in using the artifact, or its maturity relative to the content of the paper.

preparation sources

There are several sources of good advice about preparing artifacts for evaluation:

AE & Computational Artifacts Evaluation

During the Artifact Evaluation Stage, all the computational artifacts associated with the paper, such as software, datasets, or environment configuration required to reproduce the experiments are assessed. The goal of Artifact Evaluation is to award badges to artifacts of accepted papers. We base all badges on the NISO Reproducibility Badging and Definitions Standard. In 2024, the assigned badges will be per ACM Reproducibility Standard.

Authors of papers must choose to apply for a badge a priori during the AE phase. Authors can apply for one or more of the three kinds of badges that we offer. The badges available are Artifacts Available, Artifacts Evaluated-Functional, and Results Replicated. Please, note that they are incremental: If one applies for Artifacts Evaluated Functional, this also includes Artifacts Available. If one applies for Results Replicated, this also includes the other two badges. The type of badge and the criteria for each badge is explained next. To start the Reproducibility Evaluation Process, authors must provide links to their computational artifacts.

After the evaluation process, artifacts must be frozen to guarantee their persistence and immutability. An artifact must be accessible via a permanently persistent and publicly shareable DOI (What is a DOI?) on a hosting platform that supports persistent DOIs and versioning (for example, DataPort, Dryad, FigShare, Harvard Dataverse, or Zenodo). Authors should not provide links or zipped files hosted through personal webpages or shared collaboration platforms, such as Next Cloud, Google Drive, or Dropbox. 

Zenodo and FigShare provide an integration with GitHub to automatically generate DOIs from Git tags. Therefore, it is possible to host code using version control provided by GitHub and describe the artifact using Zenodo or FigShare. Please, note that Git itself (or any other control versioning software) does not generate a DOI, and it needs to be paired with Zenodo or FigShare.

artifacts available

The following are necessary to receive this badge:

  • Assigned DOI to your research object by the Artifact Freeze deadline. DOIs can be acquired via Zenodo, FigShare, Dryad, Software Heritage. Zenodo provides an integration with Github to automatically generate DOIs from Git tags.
  • Links to code and data repositories on a hosting platform that supports versioning: GitHub, or GitLab. Please do NOT provide Dropbox links or gzipped files hosted through personal webpages.

Note that, for physical objects relevant to the research, the metadata about the object should be made available.

What do we mean by accessible? Artifacts used in the research (including data and code) are permanently archived in a public repository that assigns a global identifier and guarantees persistence, and are made available via standard open licenses that maximize artifact availability.

Artifacts Evaluated-Functional

The criteria for the Artifacts Evaluated-Functional badge require an AD/AE committee member to agree whether the artifact provides enough details to exercise the artifact of components in the paper. For example, is it possible to compile the artifact, use a Makefile, or perform a small run? If the artifact runs on a large cluster—can it be compiled on a single machine? Can analysis be run on a small scale? Does the artifact describe the components to nurture future use of this artifact?

The reviewer will assess the details of the research artifact based on the following criteria:

  • Documentation: Are the artifacts sufficiently documented to enable them to be exercised by readers of the paper?
  • Completeness: Do the submitted artifacts include all of the key components described in the paper?
  • Exercisability: Do the submitted artifacts include the scripts and data needed to run the experiments described in the paper, and can the software be successfully executed?

We encourage authors to describe their (i) workflow underlying the paper, (ii) describing some of the black boxes, or a white box (e.g., source, configuration files, build environment), (iii) input data: either the process to generate the input data should be made available, or when the data is not generated, the actual data itself or a link to the data should be provided, (iv) environment (system configuration and initialization, scripts, workload, measurement protocol) used to produce the raw experimental data, and (v) the scripts needed to transform the raw data into the graphs included in the paper.

Results Replicated

The evaluators successfully reproduced the key computational results using the author-created research objects, methods, code, and conditions of analysis. Note we do not aim to recreate the exact or identical results, especially hardware-based results. However, we do aim to:

  • Reproduce Behavior: This is of specific importance where results are hardware-dependent. Bitwise reproducibility is not our goal. If we get access to the same hardware as used by experiments, we will aim to reproduce the results on that hardware. If not, we aim to work with authors to determine the equivalent or approximate behavior on available hardware. For example, if results are about response time, our objective will be to check if a given algorithm is significantly faster than another one, or that a given parameter affects negatively or positively the behavior of a system.
  • Reproduce the Central Results and Claims of the Paper: We do not aim to reproduce all results and claims shown in the paper. The AD/AE committee will evaluate all artifacts that the authors have declared as supporting the central results of the accepted paper.
SC speaker

reproducibility initiative

SC has been a leader in tangible progress towards scientific rigor, through its pioneering practice of enhanced reproducibility of accepted papers. 

Back To Top Button