The 1st International Fuzzing Workshop (FUZZING) 2022 welcomes all researchers, scientists, engineers and practitioners to present their latest research findings, empirical analyses, techniques, and applications in the area of fuzzing and software testing for automated bug finding.
The workshop will be organized as Phase 1 in a 2-phase preregistration-based publication process. All research papers will be peer-reviewed on the basis of a full-length preregistered report, and acceptance will be based on (i) the significance and novelty of the hypotheses or techniques, and (ii) the soundness and reproducibility of the methodology specified to validate the claims or hypotheses -- but explicitly not based on the strength of the (preliminary) results. More details about the benefits of this process can be found in this blog post co-authored by the workshop organizers: http://fuzzbench.com/blog/2021/04/22/special-issue/
Update (8th June): We are excited to announce that ACM TOSEM has agreed to host the registered reports that have been accepted in Phase 1 of FUZZING'22 in their new Registered Reports track! Please submit the revision of your accepted FUZZING'22 paper under https://mc.manuscriptcentral.com/tosem until 15 June 2022 (chosing "Registered Report" as paper type).
Along with the submission of the accepted FUZZING'22 paper to TOSEM, we also facilitate an Artifact Evaluation (AE). Please checkout our AE page with more information and the submission instructions: fuzzingworkshop.github.io/artifacts.html
Dissecting American Fuzzy Lop - A FuzzBench Evaluation
Andrea Fioraldi, Alessandro Mantovani (EURECOM), Dominik Maier (TU Berlin), Davide Balzarotti (EURECOM)
NSFuzz: Towards Efficient and State-Aware Network Service Fuzzing
Shisong Qin (Tsinghua University), Fan Hu (State Key Laboratory of Mathematical Engineering and Advanced Computing), Bodong Zhao, Tingting Yin, Chao Zhang (Tsinghua University)
Fuzzing Configurations of Program Options
Zenong Zhang (University of Texas at Dallas), George Klees (University of Maryland), Eric Wang (Poolesville High School), Michael Hicks (University of Maryland), Shiyi Wei (University of Texas at Dallas)
Generating Test Suites for GPU Instruction Sets through Mutation and Equivalence Checking
Shoham Shitrit, Sreepathi Pai (University of Rochester)
First, Fuzz the Mutants
Alex Groce, Goutamkumar Kalburgi (Northern Arizona Univeristy), Claire Le Goues, Kush Jain (Carnegie Mellon University), Rahul Gopinath (Saarland University)
Fine-Grained Coverage-Based Fuzzing
Bernard Nongpoh, Marwan Nour, Michaël Marcozzi, Sébastien Bardin (Université Paris Saclay)
datAFLow: Towards a Data-Flow-Guided Fuzzer
Adrian Herrera (Australian National University), Mathias Payer (EPFL), Antony Hosking (Australian National University)
Submissions are solicited in, but not limited to, the following areas:
The workshop solicits registered reports drafts. A registered report is a full paper sans the evaluation or experiments. Each draft will be reviewed by at least three members of the program committee according to the review criteria mentioned above with the key objective of providing constructive feedback. Accepted drafts are made available to all participants. These drafts will be presented and discussed in detail at the workshop in order for the authors to receive further constructive feedback. After incorporating this feedback, the authors can submit final versions of the Registered Reports for review. Notably, accepted Registered Reports will be invited as full articles in a Special Issue in one of the premier software engineering journals. This invitation is equivalent to an in-principle acceptance in the general pre-registration process. For the final journal article, the authors are expected to conduct the experiments, evaluation, or study as specified in their registered report.
We are currently discussing with top journals regarding the publication of the final papers. Therefore, the deadlines for the submission of registered reports and full journal papers might be delayed, depending on the constraints from the journal.
Submitted report drafts are expected to be a full technical paper sans the full evaluation/experiment results. To assess the feasibility of the experiments, however, we expect some preliminary results in small-scale experiments.
Submitted report drafts should include no more than 8 pages, excluding references. There are no page limits on the references. Papers must be formatted according to the NDSS requirements. Templates are available at https://www.ndss-symposium.org/ndss2022/call-for-papers .
Depending on the authors’ preference, accepted registered report drafts will be published in the workshop proceedings of NDSS through the Internet Society. The proceedings will be submitted for publication in IEEE Xplore.
Update: FUZZING'22 will employ a double-anonymous policy for all submissions. Please ensure that the authors remain anonymous in your submission. For more details on this policy, please email us or consult this excellent FAQ. (Due to the short notice, we will not desk-reject a submission in violation of this policy. Instead, we will ask the authors to update the submission with an anonymized revision.)
Report drafts can be submitted at: https://easychair.org/conferences/?conf=fuzzing22
Q: Accepting papes without considering the outcome of the experiments
sounds like “lowering the bar” for publication. Won’t this lead to
A: No. In fact, we strongly believe that the process will lead to higher quality papers with a stronger focus on the significance/novelty of the proposed approach and the soundness and reproducibility of the evaluation. Specifically, moving the in-principle acceptance to a time *before* the evaluation is conducted i) improves the soundness of the evaluation (based on early reviewer feedback) and ii) ensures that the evaluation is free of bias (e.g., HARKing).
Q: Does this mean that you can publish negative results too? What is the
point of that?
A: Yes. Compared to the existing publication model, our preregistration-based model will also allow the publication of negative results, i.e., results that show that a proposed approach does indeed not work or a reasonable hypothesis does not hold. However, given that reviewers deemed the investigation to be worthwhile and the evaluation methodology to be sound, we are convinced that the publication even of negative results will not only avoid redundant efforts across the community but also enrich our understanding of the problem under investigation. Like for positive results, we ask that authors thoroughly analyze the underlying reasons for negative results and provide an interpretation.
Q: Can I modify the idea or the experimental protocol after in-principle
acceptance in Stage-1? If yes, how much can be changed?
A: Yes. The registered report serves as an agreement on the minimal experiment and any deviation must be clearly justified. You are allowed to change the paper and the experimental protocol within reason. For instance, if the results point to an optimization opportunity for a proposed technique, you are welcome to implement and evaluate the benefit of the optimization. If the results call for a deeper investigation of certain aspects of the proposal, you are welcome to add further research questions. Any such deviation from the experimental protocol must be explained in a Summary of Changes which will be subject to review in Stage-2. For larger deviations from the evaluation protocol or when in doubt, you are welcome to request permission to follow an alternative evaluation protocol.
Q: What if my idea gets “scooped” after my registered report is
A: The registered report is published and “active”, i.e., the community knows that you are working on the project laid out in the registered report. The “scoop” would be pretty obvious. Even if related work is published after the in-principle acceptance is granted in Stage-1, this will not influence the final acceptance decision for your paper in Stage-2. The accepted registered report is a joint commitment by the authors and the reviewers.
Q: What if I cannot finish the experiments and finalize the paper before
the final submission deadline?
A: There is indeed a deadline for submitting the final version of your paper. If you need an extension, please provide an explanation for this extension anda suggestion for a new deadline. Extensions are normally granted but require a reasonable justification. If a deadline expires without submission, the registered report becomes “inactive” and the in-principle acceptance is withdrawn. The paper is not considered “under review” anymore and can be submitted elsewhere.
Q: What if my final paper gets rejected? Can I submit to other venues?
What if the other venue rejects the paper claiming that the idea has already
been published in the registered report?
A: Unless inactive, withdrawn, or rejected/accepted in Stage-2, an accepted registered report is considered as a full paper currently “under review” and cannot be submitted elsewhere. Once the report is withdrawn, rejected, or inactive, you can submit the final paper elsewhere. You can think of the registered report as a workshop paper without evidence or only preliminary results. Conferences or journals often publish extensions of workshop papers. There should be no difficulty submitting the full paper despite the publication of the registered report.
Q: Stage 1 is double-anonymous: Reviewers do not know the authors and vice
versa. Yet, Stage 2 is single-anonymous: Reviewers know about the authors
for each submission. How does this influence the final decision?
A: Stage 2 merely serves as a confirmation that the agreed experimental protocol has been followed and that deviations are explained. Since this judgment is objective, we believe that the risk of reviewer bias is sufficiently low in Stage 2.
|Design by Mike Pierce||© Conference Organizers|