Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What is the difference between dataset V1 and V2? #5

Closed
HongyuLi2018 opened this issue May 29, 2019 · 1 comment
Closed

What is the difference between dataset V1 and V2? #5

HongyuLi2018 opened this issue May 29, 2019 · 1 comment

Comments

@HongyuLi2018
Copy link

I noticed that there is an update in MRQA datasets. Is it possible to provide some details about the changes? Thanks!

@ajfisch
Copy link
Collaborator

ajfisch commented May 29, 2019

Yep, absolutely. The change is minor:

The issue: We identified that in one of the dataset creation scripts not all of the detected_answers were kept. This means that if an answer had more than one possible answer (e.g. Super Bowl vs Super Bowl 50) only occurrences (start/end token) of one were recorded. Note that answers, the original set of answers from which detected_answers are derived, is unchanged.

Effects: This only affected datasets that had more than one annotated gold span type per question for training. The full set of answers which the evaluation script uses was unaffected, so baseline results don't change. Visualization was affected; not all the true answer options were shown.

Update: The updated v2 datasets are exactly the same, just with their corrected detected_answers lists.

Thanks for the patience!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants