Comments (6)
I'm surprised by the perplexity of your language model:
valid_mn_mlm_ppl -> 12.698742
valid_zh_mlm_ppl -> 482.045657
482 is really high, I'm afraid the model does not encode Chinese sentences properly here. Can you pretrain for a longer time? Did you stop at epoch 11 because the model had stopped converging?
Another thing to check is whether the word embeddings in the pretrained LM lookup table are somehow aligned. You can for instance print the nearest neighbors of Chinese words and see if they are close to their Mongolian translations.
from xlm.
How can I get the nearest neighbors of Chinese words?@glample
from xlm.
This notebook will show you how to reload a model with the associated dictionary: https://github.com/facebookresearch/XLM/blob/master/generate-embeddings.ipynb
Once you have it, you can simply extract the embeddings with model.embeddings
. Then, for a given Chinese word X, just look for model.embeddings[dico.index(X)]
it will give you the embedding of the word. You can then simply do some nearest neighbors search of the closest vectors in the embeddings, and map word ids to their original words with dico[word_id]
.
from xlm.
If the word embeddings in the pretrained LM lookup table are aligned not good,How can I improve it.@glample
from xlm.
First you would need to fix the language model quality on the Chinese sentences, the perplexity should be much lower than this, so right now it is not surprising that it does not work. Maybe train with more data / check that the Chinese segmentation is correct and that there is a not a bug in your data preprocessing? Or simply train longer / with more GPUs, as 11 epochs is not much.
from xlm.
I got it. Thanks!
from xlm.
Related Issues (20)
- Add memory to transformer
- XLM LICENSE
- Error when using the uploaded en-fr model for NMT (translate from English to French) HOT 1
- Error in Training HOT 3
- Generate multiple optimal results(beam search)
- Training data details for XLM-15 model HOT 1
- Question about parameters for further training of a preexisiting model?
- default params for PKM
- supervised machine translation HOT 1
- How is sentence piece model trained in XLM-R?
- [Question] Does XLM-R follows RoBERTa or XLM for MLM?
- ./get-data-para.sh HOT 3
- Checkpoint for TLM objective
- confusion about `lm_head`'s size? HOT 2
- How can I expand it to a new language which is Romanised? For example, Marathi Romanized?
- e
- get-data-glue.sh 400 Bad Request
- bt_steps meaning
- how to save the entire model instead of just the model parameters
- Predict a masked word
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from xlm.