Skip to content

Heavy Workload on Reviewing Papers? ReviewAdvisor Helps out

Notifications You must be signed in to change notification settings

GitLiyx/ReviewAdvisor

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

46 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Outline

Motivation

  • Can We Automate Scientific Reviewing?

  • How will we know if we succeed?

  • Heavy workload on reviewing papers? ReviewAdvisor helps out!

Demo

Have a TRY

Dataset

The details of our dataset can be found here

Evaluation

We first define two high-level objectives of scientific peer review following Jefferson et al. (2002a). and Smith (2006).

  1. Selecting high-quality submissions for publication.
  2. Improving different aspects of a paper by providing detailed comments.

We then make a step towards review evaluation for scientific papers by proposing seven quantifiable metrics that can characterize a review from multiple perspectives. The evaluation metrics are shown in Fig.1, red represents absolute metric and blue represents relative metric.

Model

We decompose review generation into two steps, and consider using a extract-then-generate paradigm.

We consider three extraction strategies, which can be visualized in Fig.2.

  1. Section-based Extraction
  2. Cross-entropy Method Extraction
  3. Hybrid Extraction

We consider two generation frameworks, which can be visualized in Fig.3.

  1. Vanilla Sequence to Sequence
  2. Jointly Sequence to Sequence and Sequence Labeling

Bias

To characterize potential bias existing in reviews, we define aspect score as the percentage of positive occurrences with respect to each review's aspects.

Absolute bias

Absolute bias can be visualized directly by plotting the aspect scores of different groups w.r.t each aspect. An example is shown in Fig.4.

Relative bias

Relative bias quantifies the relative favor for certain groups between human reviewers and system reviewers. And we define a metric called disparity difference to measure relative bias. The calculation of disparity difference can be visualized in Fig.5.

Challenges

We summairze eight eight challenges in the pursuit of a good review generation system.

  1. Model
    • Long Document Modeling
    • Pre-trained Model for Scientific Domain
    • Structure Information
    • External Knowledge
  2. Dataset
    • More Open, Fine-grained Review Data
    • More Accurate and powerful Scientific Paper Parser
  3. Evaluation
    • Fairness and Bias in Generated Text
    • Reliability

Acknowledgement

This work cannot be accomplished without the help of many researchers.

We would like to thank people for their generous support, especially, Gábor Berend, Zhouhan Lin, William W. Cohen, Pengcheng Yin, Tiange Luo, Yuki M. Asano, Junjie Yan, Tuomas Haarnoja, Dandan Guo, Jie Fu, Lei Chen, Jinlan Fu, Jiapeng Wu, Wenshan Wang, Ziyi Dou, Yixin Liu, Junxian He, Bahetiyaer Bare, Saizheng Zhang, Jiateng Xie, Spyros Gidaris.

Bib

About

Heavy Workload on Reviewing Papers? ReviewAdvisor Helps out

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published