Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Action items for benchmarking UCCA #1

Open
omriabnd opened this issue May 15, 2020 · 2 comments
Open

Action items for benchmarking UCCA #1

omriabnd opened this issue May 15, 2020 · 2 comments

Comments

@omriabnd
Copy link
Member

omriabnd commented May 15, 2020

  • Build a webpage similar to https://nlpprogress.com/english/semantic_parsing.html#ucca-parsing where we: (1) detailed description of the official evaluation protocol (for different corpora?) including eval scripts + versions, normalization, dataset versions etc.; (2) a leader-board with parser outputs, sorted by UCCA official score and another column where they are evaluated on the MRP metric; (3) a bottom part of the page with links to other (unofficial or legacy) exp setups and corresponding leader boards.
  • Post the description in (1) as a file in the ucca code repo. Evaluation documentation huji-nlp/ucca#92
  • Improve UCCA score to more sensible handle unary expansions / multiple categories over the same edge. This will become the new official score. Ask participants of the semeval shared task and conll shared tasks whether they'd like to re-evaluate their systems and post their scores. Evaluation treats multiple categories too leniently huji-nlp/ucca#91
  • Run the new script on the MRP 2019 and 2020 submitted UCCA parses, after converting them from JSON to XML.
@omriabnd omriabnd assigned omriabnd and unassigned omriabnd May 18, 2020
@danielhers
Copy link
Member

A leaderboard will require running the experiments with leading parsers on the latest data with native UCCA evaluation (not MRP). Maybe @OfirArviv could help.

@danielhers
Copy link
Member

Also add to https://datasets.quantumstat.com/

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants