Andreas Zeller is faculty at the CISPA Helmholtz Center for Information Security and professor for Software Engineering at Saarland University. His research on automated debugging, mining software archives, specification mining, and security testing has proven highly influential. Zeller is one of the few researchers to have received two ERC Advanced Grants, most recently for his S3 project. Zeller is an ACM Fellow and holds an ACM SIGSOFT Outstanding Research Award.
Mail: andreas.zeller@cispa.de
Phone: +49 681 87083-1001
Bluesky: @andreaszeller.bsky.social
Mastodon: @AndreasZeller@mastodon.social
Linkedin: andreaszeller
GitHub: andreas-zeller
Hosted on GitHub Pages — Theme by orderedlist
by Andreas Zeller
In 2012, I attended a physical meeting of the program committee responsible for selecting the best scientific papers for the ESEC/FSE 2013 conference in Saint Petersburg, Russia. This meeting was particularly memorable because a few hours in, we discovered that one PC member had apparently given all submissions with Russian authors the highest possible grade, and all other submissions the lowest grade. He was excluded from the meeting on the spot, and all of us spent the night re-reviewing the papers originally assigned to him.
Today, such manipulation would be much more difficult. For one, our conferences routinely employ double-blind reviewing, so reviewers are unaware of the authors’ identities. Second, such a behavior is easily detectable as a statistical anomaly. If you want to manipulate the system today, you have to be more clever. Recently, I again heard rumors of a collusion ring in Software Engineering research – a group of authors and reviewers who agree to favorably review each other’s papers. I have no evidence that such a ring actually exists, but if I were to set up one, here’s how:
So, what could be wrong with this? The important thing is that there is an implicit understanding that all group members will review each other’s papers favorably. Of course, this would never admitted or discussed in my group: First rule of Fight Club is you do not talk about Fight Club. If someone were to check the group, it’s just a regular chat group of XYZ lovers, nothing to see here. And if someone were to find out that all members of my group tend to favorably review numerous “XYZ” submissions, I’d declare that is easily explained by all of us in the group working in the area and appreciating it (note: the area, not the members). The concept of plausible deniability is crucial for maintaining appearances – it is a standard feature of organized crime, which often operates behind legitimate-looking businesses. Go and prove that I’m doing wrong!
Any participation in such a scheme would be highly unethical for any researcher. The research community and their research institutions face severe sanctions against such cheaters: their reputation and career will be over, plain and simple. However, this requires that such behavior be detected in the first place. As recent research has shown, it is practically impossible to detect such collusion rings solely by observing the review process. Consequently, such rings may already be operating under the radar.
Should we now go and search for such collusion rings? Not unless we have actual evidence for their existence! The Software Engineering research community highly values its culture of mutual trust and respect, and we should do our very best not to endanger it with an unmotivated witch hunt. However, there are a few measures we can take to prevent and mitigate such collusion in the first place.
The above measures will require some sophistication to work around, but such sophistication may also leave more potential evidence. However, I tend to believe that the alleged Software Engineering collusion ring is much less sophisticated, and that its members simply exchange their paper numbers and conference names. Since authors are the only ones to know their paper numbers, a single screenshot would show they’re complicit. Such evidence can be used against them today – or in twenty years, at the height of their career. What kind of pressure must these researchers endure to run such risks?
Acknowledgments. Marcel Böhme, Lars Grunske, Giancarlo Pellegrino, Michael Pradel, and Ben Stock provided valuable insights, lessons learned, and important feedback on earlier versions of this post. Thanks a lot!
tags: popular