Oppgaven er ikke lenger tilgjengelig

Experiments with fine-tuned language models

Fine-tuning language models is the process of fitting pre-trained language models to different tasks. What can we learn from this process?

"Fine-tuning" refers to the process of continuing to train language models (such as BERT or XLM), but this time with task-specific data, instead of just language modelling. While some studies have shown that this can be useful [1], there are many unanswered questions about what precisely changes within the language model when we fine-tune.

Does the language's notion of syntax or semantics change? What happens when we introduce other languages into the mix? Do different components in the language model contribute differently to fine-tuning? Merchant et al. run a variety of probes to try and quantify the differences between language models before and after fine-tuning [2]; this is a very general paper, and more specificity in one particular direction could be interesting.

Recommended reading:

 
Publisert 12. okt. 2020 13:57 - Sist endret 4. nov. 2020 18:35

Veileder(e)

Omfang (studiepoeng)

60