Show simple item record

dc.contributor.advisorTomaso Poggio.en_US
dc.contributor.authorRifkin, Ryan Michael, 1972-en_US
dc.contributor.otherSloan School of Management.en_US
dc.date.accessioned2005-06-02T16:08:29Z
dc.date.available2005-06-02T16:08:29Z
dc.date.copyright2002en_US
dc.date.issued2002en_US
dc.identifier.urihttp://hdl.handle.net/1721.1/17549
dc.descriptionThesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, 2002.en_US
dc.descriptionIncludes bibliographical references (leaves 213-225).en_US
dc.description.abstractThis thesis shows that several old, somewhat discredited machine learning techniques are still valuable in the solution of modern, large-scale machine learning problems. We begin by considering Tikhonov regularization, a broad framework of schemes for binary classification. Tikhonov regularization attempts to find a function which simultaneously has small empirical loss on a training set and small norm in a Reproducing Kernel Hilbert Space. The choice of loss function determines the learning scheme. Using the hinge loss gives rise to the now well-known Support Vector Machine algorithm. We present SvmFu, a state-of-the-art SVM solver developed as part of thesis. We discuss the design and implementation issues involved in SvmFu, present empirical results on its performance, and offer general guidance on the use of SVMs to solve machine learning problems. We also consider, and advocate in many cases, the use of the more classical square loss, giving rise to the Regularized Least Squares Classifiation algorithm. RLSC is "trained" by solving a single system of linear equations. While it is widely believed that the SVM will perform substantially better than RLSC, we note that the same generalization bounds that apply to SVMs apply to RLSC, and we demonstrate empirically on both toy and real-world examples that RLSC's performance is essentially equivalent to SVMs across a wide range of problems, implying that the choice between SVM and RLSC should be based on computational tractability considerations. We demonstrate the empirical advantages and properties of RLSC, discussing the tradeoffs between RLSC and SVMs.en_US
dc.description.abstract(cont.) We also prove leave-one-out bounds for RLSC classification. Next, we turn to the problem of multiclass classification. Although a large body of recent literature exists suggesting the use of sophisticated schemes involving joint optimization of multiple discriminant functions or the use of error-correcting codes, we instead advocate a simple "one-vs-all" approach in which one classifier is trained to discriminate each class from all the others, and a new point is classified according to which classifier fires most strongly. We present empirical evidence of the strength of this scheme on real-world problems, and, in the context of RLSC as the base classifier, compelling arguments as to why the simple one-vs-all scheme is hard to beat. We also present leave-one-out bounds for multiclass classification where the base learners belong to a broad class of Tikhonov regularizers. Finally, we consider algorithmic stability, a relatively new theory that results in very elegant generalization bounds for algorithms which are "stable". We compare and contrast Tikhonov regularization, to which algorithmic stability applies, with Ivanov regularization, the form of regularization that is the basis for structural risk minimization and its related generalization bounds. We present some interesting examples which highlight the differences between the Tikhonov and Ivanov approaches, showing that the Tikhonov form has a much stronger stability than the Ivanov form ...en_US
dc.description.statementofresponsibilityby Ryan Michael Rifkin.en_US
dc.format.extent225 leavesen_US
dc.format.extent8585765 bytes
dc.format.extent8585573 bytes
dc.format.mimetypeapplication/pdf
dc.format.mimetypeapplication/pdf
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsM.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582
dc.subjectSloan School of Management.en_US
dc.titleEverything old is new again : a fresh look at historical approaches in machine learningen_US
dc.typeThesisen_US
dc.description.degreePh.D.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Operations Research Center
dc.contributor.departmentSloan School of Management
dc.identifier.oclc51896466en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record