Who's Reviewing for Natural Language Processing Conferences?

an analysis using public data

Introduction

We all complain about reviewers – and if you go by Twitter complaints – ML-adjacent conference reviewing is getting worse. When reviews are released, Twitter is full of (often justified) complaints and screenshots of the terrible things reviewers are saying.

There are a number of problems in reviewing (poor quality reviews, GPT-generated reviews, reviewers who don’t reply to rebuttals, etc.) but a consistent theme has been that the number of papers submitted to conferences is outstripping the number of qualified reviewers/ACs/SACs. And this makes intuitive sense: as the field grows in popularity and attracts a larger audience it will take time for newcomers to be qualified to review.

This blog post focuses on the question “who is reviewing?” Although I can’t answer this question for all ML venues, I can answer it for the Natural Language Processing (NLP) community which reliably publishes a list of reviewers in their proceedings. Through this, I can gather fine-grained information for how many times individuals are reviewing compared to how many papers they publish.

Data

I scraped all PDF proceedings from *CL conferences from the ACL Anthology website. Each proceedings contains a list of reviewers who reviewed at the conference. Note: this is not the number of papers they reviewed, only whether they reviewed/AC’d/SAC’d at all for *CL conference**.

An example of the reviewer list in EACL 2023's proceedings PDF

I then matched these names to the ACL Anthology using their name matching code (resolve_name). Finally, I gathered the list of all papers that they published in the ACL Anthology, again using the ACL Anthology code.

However, data from 15 years ago in NLP conferences is not quite as interesting, as the explosion of growth in our field happened about 5 years ago. Thus, I discard all reviewing and publication data before 2018 to focus on the last 5 years (2018-2023). The final analysis is thus done for the period of 2018 - June 2023 (including ACL 2023). I only considered these venues and their Findings papers: AACL, EACL, NAACL, ACL, and EMNLP (excluding workshops).

I also manually annotated some of the data to provide deeper insight that could not be obtained automatically. I personally annotated individual’s h-index, their region (Asia, North America, Europe, etc.),There was only one person I randomly sampled who was from Australia so I added them to the "Asia" category even though that isn't correct. There were no persons sampled from South America or Africa. This is likely due to a lack of representation in these areas overall. their pronouns as listed on their websites / invited talks,Pronouns corresponds somewhat to their preferred gender. Unfortunately, I could not directly ask all persons that I annotated information for due to time. Note that I did not find any non-binary pronouns in websites and/or talk introductions. their affiliation, whether they were in Academia/Industry/Government, and their seniority categorized in academic standards: student, < 7 years post-PhD is junior, > 7 years is senior.

⚠️ Caveats

⚠️⚠️⚠️

Data Quality

I manually selected only the PDF pages that contained reviewer information from the proceedings.

I then OCR’d the data using papermage, chunked it using newlines and/or number of tokens (250 max) with an average of between 10-30 characters, and then asked GPT-3.5-Turbo to extract the names from the noisier OCR text. I removed the names at the end which were often cut off due to line and page breaks (e.g. only the first or last name).

For an example of how this worked see the example below. It removed affiliations and the OCR’d page number (xi) and merged names that were split from the end of the page to the beginning of the newline:

An example of what the automated extraction did to the reviewer list

This process is potentially noisy, so I examined 50 different GPT-3.5-turbo based extractions. In 49 cases it was correct – however, in one case the last name was hyphenated when it should have not been. However, none of the manually annotated names that I annotated had hypens so this would only impact the overall statistics section.

The other source of noise is the ACL Anthology name matching. I checked 10 names and they all seemed correct, but it is possible there are other errors I missed – if your name doesn’t match up on the ACL Anthology please contact them!

Overall, I found the data extraction process to be fairly high quality. Although there may be errors that I didn’t find, in aggregate the data seems sound.

Analysis

Rather than release the full data and call people out by name, I decided to present the aggregate results. In each section I will break down specific groups of reviewers.

Overview Results

For the period of 2018 - June 2023 (including ACL 2023).

Overview:

Some surprising statistics:

⚠️ Reminder: when I say “review” that includes AC/SAC’ing!

I was quite shocked to see such a large number of people who haven’t reviewed: the community has qualified reviewers – they just aren’t reviewing over the course of five years!

I was even more shocked to see the number of senior people who aren’t doing any community service for *CL venues. This was probably my biggest reason for not naming-and-shaming – my career would be seriously hurt if these people decided not to hire me!

Most Egregious Non-Reviewers

Drilling in on those who are publishing > 5 papers but not reviewing/AC/SAC’ing, who are they?

I manually annotated demographics for the top 25 offenders (ranging from 30 to 100+ published *CL papers in the last 5 years):

Category Type Count
Region Europe
Asia
North America
4
12
9
Pronouns He/Him
Her/She
22
3
Seniority Student
Junior
Senior
1
2
22
Affiliation Government
Industry
Academia
2
7
16
H-Index Average
Min
Max
54.3
8
160+

We can see that these non-reviewer publishers are typically from Asia or North America, use male pronouns, are almost always senior researchers, and are generally from academia. They have an average h-index of 54, indicating that they are well-cited. They are from well-respected affiliations including the top academic and industrial groups.

Who is Reviewing the Most?

We can also examine the opposite – who are the kind souls who are reviewing despite not publishing in the past five years? I annotated the demographics for those who review the most with no papers, ranging from 10 to 15 reviews over the past 5 years:

Category Type Count
Region Europe
Asia
North America
8
1
16
Pronouns He/Him
Her/She
19
6
Seniority Student
Junior
Senior
0
6
19
Affiliation Government
Industry
Academia
1
15
9
H-Index Average
Min
Max
13.4
2
27

We see that those who are reviewing without publishing papers are typically from industry, from North America, and in the junior to senior researcher stage. This makes intuitive sense as there is less incentive to publish in industry, but many have ties to conferences and continue to serve. Their affiliations are generally from smaller startups as well as several from Google and Apple. We are grateful for their service to the community!

People who have published and reviewed

We’ve now covered the edge cases, where people either hadn’t published or hadn’t reviewed in the past five years. But what about those that have reviewed and published?

How much should researchers (and senior ones) review?

There is no magic number of times that one should review per publication – however, I think we could probably agree that it would be unkind to publish large numbers of papers and only review once. The percentiles of Papers Per Review is as follows:

Percentile 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.95 0.99
Papers Per Review 0.5 1.0 1.0 1.5 2.0 2.5 3.0 4.0 6.3 9.0 17.7

A relatively service-friendly ratio might be something like 5 publications to one time reviewing/AC/SAC’ing (given that reviewers are typically asked to review five papers per conference and that it is the 85th percentile).

On one hand, reviewing/AC/SAC’ing once per every 5 papers becomes difficult for very senior researchers, who publish lots of papers! On the other hand, our community needs senior reviewers/AC/SAC’ing along with the less experienced reviewers (especially for the AC’s and SAC’s), so researchers who publish many papers without reviewing as much leaves the community bereft of their wisdom and forces more and more junior people to take on these roles.

Statistics of Published and Reviewing Authors

What are the demographics of those who have the worst Papers Per Reviewing Time ratio?

Category Type Count
Region Europe
Asia
North America
8
6
11
Pronouns He/Him
Her/She
25
0
Seniority Student
Junior
Senior
0
1
24
Affiliation Government
Industry
Academia
1
8
16
H-Index Average
Min
Max
66.8
12
200+
Papers Per Reviewing Time Average
Min
Max
42.4
26
84

We can see that those who have reviewed but don’t do it much (e.g. a high Papers Per Reviewing Time) are typically from North America, use male pronouns (notably no female pronouns), are almost always senior researchers, and generally from academia. They have an astounding average h-index of 66, indicating that they are extremely well cited.

I can tell you that these people include some of the “hottest” names in the field, ranging from prominent universities to major industry groups.

What about people from normal ranges? I sampled people who have reviewed between 3 to 10 times and annotated until I got tired (38 total):

Category Type Count
Region Europe
Asia
North America
16
6
16
Pronouns He/Him
Her/She
31
7
Seniority Student
Junior
Senior
9
23
6
Affiliation Government
Industry
Academia
0
11
27
H-Index Average
Min
Max
12.7
5
32
Papers Per Reviewing Time Average
Min
Max
1.7
0.25
5.75

We note that this shows a much more balanced set of reviewers: mostly junior researchers (post-PhD, < 7 years), a few students, a few faculty, and mostly from academia. Those with female pronouns are still underrepresented but are more balanced than in the previous sections. Overall, reviewers come from a broad mix of university and industry affiliations, with no main pattern.

Total Statistics

I took all the annotated results and computed the following amount of reviews that each demographic does, on average:

Pronouns Average Reviews
He/Him 4.8
She/Her 6.7

A pretty large and striking difference between those with male vs female pronouns!

Affiliation Region Average Reviews
Asia 3.0
Europe 5.3
North America 5.7

North American affiliates review the most on average, followed by Europe affiliates, and then Asian affiliates last. ⚠️ Reminder, these are based on regional affiliation and not ethnicity.

Seniority Average Reviews
Student 5.8
Junior (<7 years post-PhD) 6.7
Senior (> 7 years post-PhD) 4.2

We see that those post-PhD but less than 7 years after review/AC/SAC the most. Senior researchers review/AC/SAC the least.

Academia vs Industry Average Reviews
Government 3.3
Industry 6.5
Academia 4.2

This debunks a lot of the common criticism I hear about reviewing, such as “people in industry aren’t reviewing” (actually no, they review more than academics on average). Even when separating these by Seniority we see that the trend holds:

Academia vs Industry by Seniority Average Reviews
Senior in Government 3.3
Junior in Industry 6.5
Senior in Industry 6.5
Student in Academia 5.8
Junior in Academia 6.7
Senior in Academia 2.6

I think a good argument can be made that not enough students are reviewing – but clearly we can see that students are reviewing a proportionally equivalent amount when they are invited to review.

H-Index vs Reviewing?

The above prompted me to plot the h-index against the number of times a person reviews. We can see a pretty strong negative correlation. This is to be somewhat expected, but the scale of it is quite intense - if no senior people are reviewing/AC’ing/SAC’ing, we lose all their experience! Of all the people I annotated, no one with an h-index above 55 reviewed/AC’d/SAC’d more than 2 times over five years.

Comparing the number of times someone has reviewed in the last five years to their H-Index

Conclusion

Thanks for reading this far and hopefully you enjoyed the data! If people want the data to conduct reasonable experiments with them feel free to get in touch, but as of right now I think it will remain private.

I’m not going to end with any suggestions, other than to say that I hope the community takes some steps to ensure reviewing is successful; as of right now, I don’t feel like I get any benefit from peer review. And this sentiment building through the community only puts peer review in a downward spiral. So let’s come together to counter that :)

FAQ

Will you release the data publicly?

Currently not - I don’t want prominent members of the community to dislike me and deny me opportunities for jobs! Perhaps post-PhD (feel free to send me job offers ;)).

Can you tell me what my (or my friends) Papers Per Reviewing Time number is?

Sure, send me an email to get yours (not your friends). But since you’ve done your own reviewing, feel free to compute your Papers Per Reviewing Time using those numbers!

Acknowledgments

I’d like to thank Luca Soldaini and Kyle Lo who helped inspire this project and provided useful feedback. I’d also like to thank Marc Marone, Nathaniel Weir, Aleem Khan, and Michael Saxon for their advice on data processing and proofreading.

BibText

If you found this blogpost useful and would like to cite it, you can cite it as:

@misc{weller2024reviewing,
      title={Who's Reviewing for Natural Language Processing Conferences?}, 
      author={Orion Weller},
      year={2024},
      url={https://orionweller.github.io/blog/2024/reviewing/}
}