-
Notifications
You must be signed in to change notification settings - Fork 463
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] Which is better for Cross Lingual classfication Task: LASER or XLM? #106
Comments
XLM is better than LASER |
You can consider further improving the results for regular document classification by following this approach http://nlp.fast.ai/classification/2019/09/10/multifit.html . We used LASER as XLM wasn't available when we were testing multifit. I would be super interested to see how it works with XLM |
XLM does not cover all 100 languages to the best of my knowledge. Which model/implementation did you use @Ayush-iitkgp ? |
Indeed, XLM with MLM+TLM only covers 15 languages currently... |
@Ayush-iitkgp From my reading of their paper a few weeks ago, my understanding is that the version MLM+TLM is the one that gets to the best results in terms of multilingual embeddings and that can outperform LASER. Indeed, multi-BERT already implemented MLM for a large number of languages and the quality of the multilingual embeddings is not optimal. |
Hello, |
@Ayush-iitkgp How is your approach doing? |
@Bachstelze my approach included fine-tuning the XLM model on English data and using zero-shot classification to predict on german and Spanish languages. However, the performance of the model on german and Spanish is less than even 20% (accuracy) so I am still figuring out what can be done. The problem in my case is that I only have labeled English data and a very small amount of non-English data for performance measurement. Do you have any recommendations? |
How is the performance of the model on English?
Hopefully you give the power of your knowledge back to the people. |
Now XML-R has 100 language, so it makes a lot of sense to replaceLASER with XML-R:
|
Hi @loretoparisi, Cheers, |
@MastafaF pretty interesting test! We did not test XLM-R yet, I wonder why XLM model does not outperform LASER bi-LSTM networks, because according to the results they have presented it should be the opposite, but we did not replace LASER anyways for several other reasons. |
Hi @loretoparisi , XLM-R gives poor results at the moment. Stay tuned for further experiments, will post the link soon for replication 😄 |
Hi @loretoparisi , you can check some tests here on WMT2012 reproducing experiments from LASER and doing a comparative study between other multilingual architectures. |
@MastafaF thank you very much! It's a very comprehensive and rigorous analysis 💯 |
Hello,
I have the training data with labels in English. Now, I want to use this data to predict for other languages. I saw XLM and LASER both support the cross-lingual classification. However, they don't have the benchmark on the same dataset, therefore, it's difficult to know which model is better. Does someone help me in determining which(XLM or LASER) is better for cross-lingual classification?
The text was updated successfully, but these errors were encountered: