-
Notifications
You must be signed in to change notification settings - Fork 21
XLM Evaluation results #9
Comments
Hi Farad, Which implementation of XLM are you using? The HPs for XLM were: We used the pytext implementation of XLM. The correct tokenization and preprocessing is very important for good performance, I'm not sure whether the HF version has this correct, as a number of people have struggled to get the good results with XLM on HF. |
Hello, any updates on the pytext code release ? I know the COVID situation may have changed the plans. I struggle to replicate of your paper on one shot learning (that is, trained on your MLQA-train chinese) with HF XLM-R (inference with a zero-shot model on chinese works fine). Thank you ! |
Hi Rachel! XLM-R wasn't included in our paper, so we can't directly help there. Patrick |
Hi @RachelKer To achieve a similar performance on zh test sets, you just need to add |
@nooralahzadeh Thank you, I saw your issue on HF repo a few days ago and with this change I manage to get the correct results for Bert and XLM trained on chinese, but not for XLM-R. Did you manage to train XLM-Roberta on chinese ? @patrick-s-h-lewis Oh indeed I confused XLM-R and XLM on your paper, I am sorry. I think the training problem that I have occurs with XLM-R only. Thanks for checking the code release anyway, and your quick answer ! |
Hey @RachelKer and @nooralahzadeh, I asked internally about XLMR (since there is some overlap between the teams), the pytext model is released, but there aren't instructions for how to run it on MLQA, so someone is going to write these instructions up :) Patrick |
Hi,
I performed some experiments using the XLM implementation of huggingfaces training on sQuAD v1.1 training and test in the MLQA test set. The results are as follows:( f1 / EM)
en 68.51/56.13
es 57.59/41.21
ar 47.88/31.41
de 51.99/38.16
zh 38.34/21.39
hi 46.13/31.72
vi 44.09/27.07
I am wondering what it has a large difference with yours. Did you do special thing except the early stopping on MLQA-en?
The text was updated successfully, but these errors were encountered: