[D] Many times I feel additional experiments during the rebuttal make my paper worse

Reddit r/MachineLearning / 3/28/2026

💬 OpinionIdeas & Deep Analysis

Key Points

  • The author argues that stricter enforcement of review quality has led to more reviewers feeling compelled to find flaws and request extra experiments, even for papers that are otherwise acceptable.
  • They report that rebuttal-stage “what-if” experiments (e.g., changing backbone, dataset, task, or settings) often worsen papers by creating detrimental results and giving reviewers “gotcha” moments.
  • As both an author and a reviewer, they say they have had to intervene in discussion to deprioritize additional experiments suggested by reviewers when they don’t meaningfully affect the paper’s core claims.
  • The piece emphasizes that experiment requirements should be sufficient to support the main claims, not exhaustive coverage of every loosely related case.
  • It concludes that reviewers should feel comfortable stating that a paper meets the bar while expressing curiosity questions that do not impact their evaluation.

Back in the days when I just started to review for major conferences, it was common to give and receive reviews saying "I don't have major concerns".

In the past 3-5 years, the field has spent significant effort cracking down on low-quality reviews, which is great. But a side effect is that we don't see these kinds of "easy" reviews anymore. It feels like the reviewers are obliged to find something wrong with the paper to show they are doing their job. Even on papers where all reviewers are accepting, it's common for the author to be requested 5-10 additional numbers/plots during rebuttal.

Many times, these experiments are detrimental. Most of them are "what ifs". How about a different backbone, task, dataset, or a specific setting? And whenever something doesn't work (especially during the rebuttal timeframe), the reviewer is having a good "gotcha" moment. I'm not only complaining as an author but also as a reviewer. Several times, I had to step in during the discussion: "I don't think X experiment suggested by Reviewer Y is important," And every time the AC sided with me.

The requirement for experiments should always be "sufficient to support the core claims," not "exhaustively examine every single barely applicable case." Folks, it's OK to say "the paper passes the bar, but I have curiosity questions that do not affect my rating" (I have written this line many times in my reviews).

submitted by /u/AffectionateLife5693
[link] [comments]
広告