Coder Social home page Coder Social logo

test model on other languages about multifit HOT 5 OPEN

n-waves avatar n-waves commented on July 16, 2024
test model on other languages

from multifit.

Comments (5)

PiotrCzapla avatar PiotrCzapla commented on July 16, 2024

Do you mean in form of zeroshoot transfer learning?

If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

from multifit.

wonderfultina avatar wonderfultina commented on July 16, 2024

I understand now,thank you.

from multifit.

vhargitai avatar vhargitai commented on July 16, 2024

Hi @PiotrCzapla , have you or your colleagues already pretrained this model on English Wikipedia?

If not, would using prepare_wiki-en.sh to grab wikitext-103, then running postprocess_wikitext.py on it be identical to the dataset preparation you did for other languages in the MultiFiT paper?

I'd like to reproduce the monolingual supervised training procedure in the MultiFiT paper for English language classification. Thanks in advance!

from multifit.

mhajiaghayi avatar mhajiaghayi commented on July 16, 2024

Do you mean in form of zeroshoot transfer learning?
If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

Q) in this case, you don't have a single model with the fixed tokenization that does zero-shot embedding for other language. am I right?

from multifit.

iNeil77 avatar iNeil77 commented on July 16, 2024

Do you mean in form of zeroshoot transfer learning?

If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.

In the CLS-DE notebook I only see the classifier fine-tuning happening with DE Music Data, Label pairs. But if I understand what you said correctly, shouldn't the LASER classifier be fine-tuned with EN Music Data first before it can act as a teacher to fine-tune the DE Classifier? I don't see that in the notebook. Am I misunderstanding the training regime?

from multifit.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.