Finding effective ways to reach an integration between learning and reasoning within deep neural networks is a long-standing problem of AI. A major breakthrough in this area has the potential of bringing the capabilities of current machine learning systems to the next level, making deep neural networks able to tackle a whole new range of problems, in a way that would have been unthinkable just a few years ago. This thesis explores possible ways to enrich the internal representations of deep learning models, under the long-term perspective of finding good inductive biases for supporting a smooth integration between learning and reasoning. The contributions presented in this document approach the problem from three different directions. First, a novel way to structure the latent representation of deep neural networks is introduced, allowing such representations to disentangle the different generative factors underlying the data. Then, a technique to enrich the same latent representations with external prior information is described, demonstrating its application on the challenging task of automatic music generation. Finally, a new benchmark for accurately measuring the systematic generalization capabilities of reasoning models is presented, based on the prediction of stoichiometrically-balanced chemical reactions. We hope that this thesis could give an in-depth overview of the current research in the fields of representation learning and learning/reasoning integration, as well as making some noteworthy contributions to the research community.

Learning Representations for Deep Neural Reasoners

VALENTI, ANDREA
2023

Abstract

Finding effective ways to reach an integration between learning and reasoning within deep neural networks is a long-standing problem of AI. A major breakthrough in this area has the potential of bringing the capabilities of current machine learning systems to the next level, making deep neural networks able to tackle a whole new range of problems, in a way that would have been unthinkable just a few years ago. This thesis explores possible ways to enrich the internal representations of deep learning models, under the long-term perspective of finding good inductive biases for supporting a smooth integration between learning and reasoning. The contributions presented in this document approach the problem from three different directions. First, a novel way to structure the latent representation of deep neural networks is introduced, allowing such representations to disentangle the different generative factors underlying the data. Then, a technique to enrich the same latent representations with external prior information is described, demonstrating its application on the challenging task of automatic music generation. Finally, a new benchmark for accurately measuring the systematic generalization capabilities of reasoning models is presented, based on the prediction of stoichiometrically-balanced chemical reactions. We hope that this thesis could give an in-depth overview of the current research in the fields of representation learning and learning/reasoning integration, as well as making some noteworthy contributions to the research community.
17-feb-2023
Italiano
deep learning
machine learning
machine reasoning
representation learning
Bacciu, Davide
File in questo prodotto:
File Dimensione Formato  
phd_thesis_main_revised_final.pdf

accesso aperto

Dimensione 6.33 MB
Formato Adobe PDF
6.33 MB Adobe PDF Visualizza/Apri
report_attivita_phd.pdf

non disponibili

Dimensione 112.37 kB
Formato Adobe PDF
112.37 kB Adobe PDF

I documenti in UNITESI sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14242/216661
Il codice NBN di questa tesi è URN:NBN:IT:UNIPI-216661