dc.contributor.advisor | Kaelbling, Leslie P. | |
dc.contributor.advisor | Lozano-Pérez, Tomás | |
dc.contributor.advisor | Tenenbaum, Joshua B. | |
dc.contributor.author | Alet i Puig, Ferran | |
dc.date.accessioned | 2023-01-19T19:58:40Z | |
dc.date.available | 2023-01-19T19:58:40Z | |
dc.date.issued | 2022-09 | |
dc.date.submitted | 2022-10-19T19:06:36.727Z | |
dc.identifier.uri | https://hdl.handle.net/1721.1/147561 | |
dc.description.abstract | Many recent achievements in machine learning have followed different variations on a single recipe: we pick a supervised training dataset and assume there exists a function mapping inputs to outputs. We then leverage the expressivity of deep learning (together with few but carefully chosen inductive biases for each domain) and train a neural network to approximate this unknown function. In this thesis, we show that this single-function, single-neural-network approach can be too constraining and instead suggest spawning per-point models. This allows us to encode inductive biases in flexible ways and model expressive, structured generative models of the data distribution.
First, we present Tailoring: a novel way of encoding inductive biases by optimizing unsupervised objectives inside the prediction function. This ensures the structure is imposed both at training and test time. Furthermore, its generality allows applications in domains as diverse as physics time-series prediction, adversarial defenses, and contrastive representation learning. We also propose Noether Networks, which automatically discover these inductive biases, in the form of conservation laws.
Finally, we propose Functional risk minimization(FRM), an alternative framework to the standard Empirical risk minimization(ERM) setting where loss functions act in function space rather than output space. We show how we can make learning in this new framework efficient and can lead to improved performance compared to the standard ML setting. | |
dc.publisher | Massachusetts Institute of Technology | |
dc.rights | In Copyright - Educational Use Permitted | |
dc.rights | Copyright MIT | |
dc.rights.uri | http://rightsstatements.org/page/InC-EDU/1.0/ | |
dc.title | Learning inside the prediction function | |
dc.type | Thesis | |
dc.description.degree | Ph.D. | |
dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | |
mit.thesis.degree | Doctoral | |
thesis.degree.name | Doctor of Philosophy | |