With the advances in Deep Learning, the term Transfer Learning (TL) has become ubiquitous in the field of Machine Learning. One of the most widely adopted strategies when working with pre-trained models is to fine-tune them on downstream tasks by leveraging a relatively smaller labeled dataset compared to the amount of training data used for the pre-training phase. Fine-tuning is in fact a common technique of transfer learning. In general TL, refers to a set of techniques and approaches which leverage training data sampled from a source distribution to improve performance on a test set, the target, containing elements sampled from a different, but related, distribution. This paradigm brings about two major advantages. First, it increases performance on the target domain by making the algorithm more robust and resilient, allowing us to leverage powerful pre-trained models that are trained on hardware not widely available. Second, it allows the application of data-intensive techniques to many scarce-resource domains where training an ad-hoc solution would be impossible. In this thesis, we explore applications of Heterogeneous Transfer Learning (HTL) to the field of Natural Language Processing. We identify two main exploratory spaces: (i) the heterogeneous space defined by different languages and (ii) the heterogeneous space defined by the intersection of languages and perceptual information. Lastly, we explore the benefits of HTL when dealing simultaneously with both multimodality and multilinguality.

Heterogeneous Transfer Learning in Natural Language Processing

PEDROTTI, ANDREA
2024

Abstract

With the advances in Deep Learning, the term Transfer Learning (TL) has become ubiquitous in the field of Machine Learning. One of the most widely adopted strategies when working with pre-trained models is to fine-tune them on downstream tasks by leveraging a relatively smaller labeled dataset compared to the amount of training data used for the pre-training phase. Fine-tuning is in fact a common technique of transfer learning. In general TL, refers to a set of techniques and approaches which leverage training data sampled from a source distribution to improve performance on a test set, the target, containing elements sampled from a different, but related, distribution. This paradigm brings about two major advantages. First, it increases performance on the target domain by making the algorithm more robust and resilient, allowing us to leverage powerful pre-trained models that are trained on hardware not widely available. Second, it allows the application of data-intensive techniques to many scarce-resource domains where training an ad-hoc solution would be impossible. In this thesis, we explore applications of Heterogeneous Transfer Learning (HTL) to the field of Natural Language Processing. We identify two main exploratory spaces: (i) the heterogeneous space defined by different languages and (ii) the heterogeneous space defined by the intersection of languages and perceptual information. Lastly, we explore the benefits of HTL when dealing simultaneously with both multimodality and multilinguality.
4-ott-2024
Italiano
cross-lingual
multi-modal
transfer learning
Moreo Fernández, Alejandro
Sebastiani, Fabrizio
File in questo prodotto:
File Dimensione Formato  
phd_report.pdf

non disponibili

Licenza: Tutti i diritti riservati
Dimensione 139.05 kB
Formato Adobe PDF
139.05 kB Adobe PDF
phd_thesis_andreapedrotti.afterreview.pdf

embargo fino al 01/10/2027

Licenza: Tutti i diritti riservati
Dimensione 30.07 MB
Formato Adobe PDF
30.07 MB Adobe PDF

I documenti in UNITESI sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14242/216375
Il codice NBN di questa tesi è URN:NBN:IT:UNIPI-216375