Skip to content
This repository has been archived by the owner on Feb 17, 2024. It is now read-only.

clarification on how MLQA eval is done in mt5 #70

Open
dorost1234 opened this issue Apr 4, 2021 · 1 comment
Open

clarification on how MLQA eval is done in mt5 #70

dorost1234 opened this issue Apr 4, 2021 · 1 comment

Comments

@dorost1234
Copy link

Hi

MLQA only have eval/test sets, could you please assist me and clarify on which dataset you have trained the model for the zero-shot experience? The result reported in the paper, are they computed on the eval set or the test set?

thank you.

@hanss0n
Copy link

hanss0n commented May 11, 2021

It says so in the paper. The English zero-shot evaluation on MLQA was done using SQuAD and the multilingual one was conducted using SQuAD + translations of SQuAD in the target languages, i.e. Arabic, Spanish, English, German, Hindi, Vietnamese and simplified Chinese. The MLQA set used is most likely the test one since its intended use is benchmarking.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants