With the advances in Deep Learning, the term Transfer Learning (TL) has become ubiquitous in the field of Machine Learning. One of the most widely adopted strategies when working with pre-trained models is to fine-tune them on downstream tasks by leveraging a relatively smaller labeled dataset compared to the amount of training data used for the pre-training phase. Fine-tuning is in fact a common technique of transfer learning. In general TL, refers to a set of techniques and approaches which leverage training data sampled from a source distribution to improve performance on a test set, the target, containing elements sampled from a different, but related, distribution. This paradigm brings about two major advantages. First, it increases performance on the target domain by making the algorithm more robust and resilient, allowing us to leverage powerful pre-trained models that are trained on hardware not widely available. Second, it allows the application of data-intensive techniques to many scarce-resource domains where training an ad-hoc solution would be impossible. In this thesis, we explore applications of Heterogeneous Transfer Learning (HTL) to the field of Natural Language Processing. We identify two main exploratory spaces: (i) the heterogeneous space defined by different languages and (ii) the heterogeneous space defined by the intersection of languages and perceptual information. Lastly, we explore the benefits of HTL when dealing simultaneously with both multimodality and multilinguality.
Heterogeneous Transfer Learning in Natural Language Processing
PEDROTTI, ANDREA
2024
Abstract
With the advances in Deep Learning, the term Transfer Learning (TL) has become ubiquitous in the field of Machine Learning. One of the most widely adopted strategies when working with pre-trained models is to fine-tune them on downstream tasks by leveraging a relatively smaller labeled dataset compared to the amount of training data used for the pre-training phase. Fine-tuning is in fact a common technique of transfer learning. In general TL, refers to a set of techniques and approaches which leverage training data sampled from a source distribution to improve performance on a test set, the target, containing elements sampled from a different, but related, distribution. This paradigm brings about two major advantages. First, it increases performance on the target domain by making the algorithm more robust and resilient, allowing us to leverage powerful pre-trained models that are trained on hardware not widely available. Second, it allows the application of data-intensive techniques to many scarce-resource domains where training an ad-hoc solution would be impossible. In this thesis, we explore applications of Heterogeneous Transfer Learning (HTL) to the field of Natural Language Processing. We identify two main exploratory spaces: (i) the heterogeneous space defined by different languages and (ii) the heterogeneous space defined by the intersection of languages and perceptual information. Lastly, we explore the benefits of HTL when dealing simultaneously with both multimodality and multilinguality.| File | Dimensione | Formato | |
|---|---|---|---|
|
phd_report.pdf
non disponibili
Licenza:
Tutti i diritti riservati
Dimensione
139.05 kB
Formato
Adobe PDF
|
139.05 kB | Adobe PDF | |
|
phd_thesis_andreapedrotti.afterreview.pdf
embargo fino al 01/10/2027
Licenza:
Tutti i diritti riservati
Dimensione
30.07 MB
Formato
Adobe PDF
|
30.07 MB | Adobe PDF |
I documenti in UNITESI sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/20.500.14242/216375
URN:NBN:IT:UNIPI-216375