nlp_intro

10. Cross-lingual transfer and multilingual NLP

Explanations and visualisations

 

Reasons for studying multilingual NLP

 

Language vectors

 

Multilingual data sets

 

Multilingual pre-trained models

BERT-type

GPT-type

Full Transformers

Other pre-trained models are typically trained for a single language or a group of languages (e.g. Indic BERT, AraBERT, BERTić)

 

Transfer across languages

 

finetune

continuepretrain

test

 

Language similarity and sampling

 

mentions