A Projected Subgradient Method for Scalable Multi-Task Learning
Author(s)
Quattoni, Ariadna; Carreras, Xavier; Collins, Michael; Darrell, Trevor
DownloadMIT-CSAIL-TR-2008-045.pdf (355.5Kb)
Additional downloads
Other Contributors
Vision
Advisor
Trevor Darrell
Metadata
Show full item recordAbstract
Recent approaches to multi-task learning have investigated the use of a variety of matrix norm regularization schemes for promoting feature sharing across tasks.In essence, these approaches aim at extending the l1 framework for sparse single task approximation to the multi-task setting. In this paper we focus on the computational complexity of training a jointly regularized model and propose an optimization algorithm whose complexity is linear with the number of training examples and O(n log n) with n being the number of parameters of the joint model. Our algorithm is based on setting jointly regularized loss minimization as a convex constrained optimization problem for which we develop an efficient projected gradient algorithm. The main contribution of this paper is the derivation of a gradient projection method with l1ââ constraints that can be performed efficiently and which has convergence rates.
Date issued
2008-07-23Other identifiers
MIT-CSAIL-TR-2008-045