A System-Level Analysis of Conference Peer Review

πŸ“… 2022-07-12
πŸ›οΈ ACM Conference on Economics and Computation
πŸ“ˆ Citations: 8
✨ Influential: 1
πŸ“„ PDF
πŸ€– AI Summary
In conference peer review, misaligned incentives among authors, conferences, and reviewers stem from inherent noise in paper quality assessment. Method: We formulate a Stackelberg game between authors and conferences, introducing the novel concept of β€œresubmission gap,” and analyze the dynamic trade-offs among acceptance thresholds, author resubmission behavior, and reviewer load via agent-based simulation, noise-aware modeling, and parameter estimation from historical data. Contributions/Results: (1) Raising the acceptance threshold reduces reviewer burden while preserving accepted paper quality; (2) Author self-selection induces a counterintuitive effect: stricter review increases the acceptance rate of high-quality papers; (3) A small number of high-quality reviews combined with a high threshold outperforms a large volume of low-quality reviews, and reusing prior reviews significantly alleviates load without compromising quality. We quantify how key parameters affect system performance, providing both theoretical foundations and empirical support for conference policy design.
πŸ“ Abstract
We undertake a system-level analysis of the conference peer review process. The process involves three constituencies with different objectives: authors want their papers accepted at prestigious venues (and quickly), conferences want to present a program with many high-quality and few low-quality papers, and reviewers want to avoid being overburdened by reviews. These objectives are far from aligned; the key obstacle is that the evaluation of the merits of a submission (both by the authors and the reviewers) is inherently noisy. Over the years, conferences have experimented with numerous policies and innovations to navigate the tradeoffs. These experiments include setting various bars for acceptance, varying the number of reviews per submission, requiring prior reviews to be included with resubmissions, and others. The purpose of the present work is to investigate, both analytically and using agent-based simulations, how well various policies work, and more importantly, why they do or do not work. We model the conference-author interactions as a Stackelberg game in which a prestigious conference commits to a threshold acceptance policy which will be applied to the (noisy) reviews of each submitted paper; the authors best-respond by submitting or not submitting to the conference, the alternative being a "sure accept" (such as arXiv or a lightly refereed venue). Our findings include: observing that the conference should typically set a higher acceptance threshold than the actual desired quality, which we call the resubmission gap and quantify in terms of various parameters; observing that the reviewing load is heavily driven by resubmissions of borderline papers --- therefore, a judicious choice of acceptance threshold may lead to fewer reviews while incurring an acceptable loss in quality; observing that depending on the paper quality distribution, stricter reviewing may lead to higher or lower acceptance rates --- the former is the result of self selection by the authors. As a rule of thumb, a relatively small number of reviews per paper, coupled with a strict acceptance policy, tends to do well in trading off these two objectives; finding that a relatively small increase in review quality or in self assessment by the authors is much more effective for conference quality control (without a large increase in review burden) than increases in the quantity of reviews per paper.; showing that keeping track of past reviews of papers can help reduce the review burden without a decrease in conference quality. For robustness, we consider different models of paper quality and learn some of the parameters from real data.
Problem

Research questions and friction points this paper is trying to address.

Analyzing conference peer review policies through game theory models
Investigating acceptance thresholds and resubmission impacts on quality
Evaluating how reviewer accuracy affects conference review efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Modeled conference review as Stackelberg game
Introduced resubmission gap acceptance threshold
Analyzed reviewer load through borderline submissions
πŸ”Ž Similar Papers
No similar papers found.
Y
Yichi Zhang
University of Michigan
Fang-Yi Yu
Fang-Yi Yu
George Mason University
G
G. Schoenebeck
University of Michigan
D
D. Kempe
University of Southern California