Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

test model on other languages #56

Open
wonderfultina opened this issue Nov 7, 2019 · 5 comments
Open

test model on other languages #56

wonderfultina opened this issue Nov 7, 2019 · 5 comments

Comments

@wonderfultina
Copy link

Hi,
I want to ask a question ,if i want to use model that trained in English,and use it to test other languages.How do I run the code?

@PiotrCzapla
Copy link
Member

Do you mean in form of zeroshoot transfer learning?

If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

@wonderfultina
Copy link
Author

I understand now,thank you.

@vhargitai
Copy link

Hi @PiotrCzapla , have you or your colleagues already pretrained this model on English Wikipedia?

If not, would using prepare_wiki-en.sh to grab wikitext-103, then running postprocess_wikitext.py on it be identical to the dataset preparation you did for other languages in the MultiFiT paper?

I'd like to reproduce the monolingual supervised training procedure in the MultiFiT paper for English language classification. Thanks in advance!

@mhajiaghayi
Copy link

Do you mean in form of zeroshoot transfer learning?
If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

Q) in this case, you don't have a single model with the fixed tokenization that does zero-shot embedding for other language. am I right?

@iNeil77
Copy link

iNeil77 commented Jul 27, 2020

Do you mean in form of zeroshoot transfer learning?

If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

In the CLS-DE notebook I only see the classifier fine-tuning happening with DE Music Data, Label pairs. But if I understand what you said correctly, shouldn't the LASER classifier be fine-tuned with EN Music Data first before it can act as a teacher to fine-tune the DE Classifier? I don't see that in the notebook. Am I misunderstanding the training regime?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants