Artifact Evaluation for Computer Systems' Research
We work with the community and ACM to improve methodology and tools for reproducible experimentation, artifact submission / reviewing and open challenges!
About Artifacts Committee Submission Reviewing FAQ Prior AE

Updated guide is now available here



This guide (V20151015) should help you describe and submit your artifacts for a review. It gradually evolves based on our past Artifact Evaluations and your feedback (see this presentation with an outcome of the past PPoPP/CGO'15 AE). It should also help you prepare your artifacts for a possible public release, if you plan to do so (for example as an auxiliary material in a Digital Library or on your personal web page).

Navigation:

What to expect

We are trying to make artifact submission as simple as possible. You just need to pack your artifact (code and data) using any publicly avialable tool you prefer or arrange a remote access to machine with pre-installed software (exceptional cases when rare hardware or proprietary software is used).

Then, you need to prepare a small and informal guide for reviewers using our AE LaTeX template (see below) to explain what are your artifacts, how to access and use them, and what is the expected result (we currently discuss with the conference chairs and ACM to let you keep it as Appendix in your paper if your artifact passes evaluation).

At least two reviewers will follow your guide to replicate your results (for example, exact output match) or reproduce them (for example, varying performance numbers or scalability on a different machine), and will then send you a report with the following overall assessment of your artifact based on our reviewing guidelines:

where "met expectations" score or above means that your artifact successfully passed evaluation and will receive a stamp of approval (added to the paper itself):
The highest ranked artifact (usually not only reproducible but also customizable and reusable) will also receive a "distinguished artifact" award announced at the joint CGO-PPoPP'16 AE discussion session. This section is also used as a discussion forum with the community about how to improve AE.

Since our eventual goal is to promote artifact validation and sharing (rather than naming and shaming problematic artifacts), you will be able to address raised issues during the rebuttal. Furthermore, we allow a small amount of communication between reviewers and authors whenever there are installation/usage problems. In such cases, AE chairs will serve as a proxy to avoid revealing reviewers' identity (the review is blind, i.e. your identity is known to reviewers since your paper is already accepted, but not vice versa).

Preparing artifacts for submission

You just need to perform the following 4 steps to submit your artifact:
  1. Pack your artifact (code and data) or provide an easy access to them using any publicly available and free tool you prefer or strictly require.

    For example, you can use the following:
    • Virtual Box to pack all code and data including OS (typical images are around 2..3GB. we strongly recommend to avoid images larger than 10GB).
    • Docker to pack only touched code and data during experiment.
    • Standard zip or tar with all related code and data, particularly when artifact should be rebuilt on a reviewers machine (for example to have a non-virtualized access to a specific hardware).
    • Private or public GIT or SVN.
    • Arrange a remote access to machine with pre-installed software (exceptional cases when rare hardware or proprietary software is used or the VM image is too large)) - you will need to privately send the access information to the AE chairs. Also, please avoid making any changes to the remote machine during evaluation (unless explicitly agreed with AE chairs) - you can do it during rebuttal phase, if needed!
    You can also check and use other public tools available for artifact sharing.

    Note that after experiencing various sharing and reproducibility issues during past Artifact Evaluation, we started developing new and open-source technology which involves the community to solve them. If you are interested to check it out, we will be happy to help you use this technology or get your feedback to improve it for the future AE:

  2. Write a brief artifact abstract to informally describe your artifact including minimal hardware and software requirements, how it supports your paper, how it can be validated and what is the expected result. It will be used to select appropriate reviewers.

  3. Fill in and append AE template (download here) to the PDF of your accepted paper. Though it should be relatively intuitive, you can check out extra notes about this template based on our past AE experience.

  4. Submit artifact abstract and new PDF at this EasyChair website for the joint PPoPP/CGO 2016 AE.
If you encounter problems, find some ambiguities or have any questions, do not hesitate to contact AE chairs privately or publicly via LinkedIn group and this mailing list.

If accepted

You can now add the following stamp to the final camera-ready version of your paper: While there are no strict formatting rules for the stamp, please add it anywhere close to the title. For example, see PPoPP'15 article together with this LaTeX example. You can change \hspace and \raisebox parameters to better fit stamp to your paper.

We strongly encourage you to submit your AE appendix as an auxiliary material for Digital Library (while removing all unnecessary or confidential information) along with the final variant of your paper. This will help readers better understand what was evaluated. We currently discuss procedures with ACM DL colleagues - please stay tuned!

Though you are not obliged to publicly release your artifacts (in fact, it is sometimes impossible due to various limitations), we also strongly encourage you to share them with the community (even if they are not open-source). You can release them as an auxiliary material in Digital Libraries together with your AE appendix or use your institutional repository and various public services for code and data sharing.

Even accepted artifacts may have some unforeseen behavior and limitations discovered during evaluation. Now you have a chance to add related notes to your paper as a future work (if you wish)..

Note, that we are developing open-source Collective Knowledge framework to help researchers share their artifacts as unified, reusable and customizable components together with experimental workflows and interactive articles. If you are interested to know more, please check some online examples and recent DATE'16 paper.

A few examples of accepted artifacts from the past conferences, workshops and journals

Methodology archive

We keep track of all past versions of submission/reviewing methodology to let readers understand which one was used in papers with evaluated artifacts.

Thank you for participating in Artifact Evaluation!

This guide was prepared by Grigori Fursin.
Maintained by
cTuning foundation (non-profit R&D organization)
and volunteers!
          
Powered by Collective Knowledge
                     
  
  
  
           Locations of visitors to this page