Regularization Through Feature Knock Out
Author(s)
Wolf, Lior; Martin, Ian
DownloadMIT-CSAIL-TR-2004-072.ps (15843Kb)
Additional downloads
Metadata
Show full item recordAbstract
In this paper, we present and analyze a novel regularization technique based on enhancing our dataset with corrupted copies of the original data. The motivation is that since the learning algorithm lacks information about which parts of thedata are reliable, it has to produce more robust classification functions. We then demonstrate how this regularization leads to redundancy in the resulting classifiers, which is somewhat in contrast to the common interpretations of the OccamÂs razor principle. Using this framework, we propose a simple addition to the gentle boosting algorithm which enables it to work with only a few examples. We test this new algorithm on a variety of datasets and show convincing results.
Date issued
2004-11-12Other identifiers
MIT-CSAIL-TR-2004-072
AIM-2004-025
CBCL-242
Series/Report no.
Massachusetts Institute of Technology Computer Science and Artificial Intelligence Laboratory
Keywords
AI