Comments (5)
Do you mean in form of zeroshoot transfer learning?
If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.
from multifit.
I understand now,thank you.
from multifit.
Hi @PiotrCzapla , have you or your colleagues already pretrained this model on English Wikipedia?
If not, would using prepare_wiki-en.sh
to grab wikitext-103, then running postprocess_wikitext.py
on it be identical to the dataset preparation you did for other languages in the MultiFiT paper?
I'd like to reproduce the monolingual supervised training procedure in the MultiFiT paper for English language classification. Thanks in advance!
from multifit.
Do you mean in form of zeroshoot transfer learning?
If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.
Q) in this case, you don't have a single model with the fixed tokenization that does zero-shot embedding for other language. am I right?
from multifit.
Do you mean in form of zeroshoot transfer learning?
If so we use Laser for that. First we train laser , to obtain zeroshoot predictions for other languages.
Then we use that zershoot predictions to train regular multifit (pretrained in the language that we are testing on). The unsupervised pretraining removes noise from the laser zeroshoot predictions and improves the results.
In the CLS-DE notebook I only see the classifier fine-tuning happening with DE Music Data, Label pairs. But if I understand what you said correctly, shouldn't the LASER classifier be fine-tuned with EN Music Data first before it can act as a teacher to fine-tune the DE Classifier? I don't see that in the notebook. Am I misunderstanding the training regime?
from multifit.
Related Issues (20)
- Different size of CLS unsupervised data between .csv and original .xml files HOT 1
- Multifit inference problem HOT 2
- Training custom classifier HOT 1
- Specifying a validation set HOT 1
- Problems with reproducing zero-shot learning results HOT 2
- Where can I find the dataset de.train.csv? HOT 3
- Kernel restarted HOT 1
- Saliency maps
- Tokenizer HOT 1
- Get activations of a specific layer of the multifit model
- Missing File in CLS-DE.ipynb HOT 1
- Always labels are tokenizing instead of text column, Kindly fix the issue facing HOT 1
- Download music/books data in german version
- OOM during finetuning
- File exists but it doesn't found it!!!!! HOT 1
- fp16
- Label_for_lm() takes too much time!
- multifit does'nt work on Google Colab HOT 4
- Port to Fastai 2 HOT 1
- Create classifier with fastai v1.0
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from multifit.