Accelerating greedy coordinate descent methods
Author(s)
Lu, Haihao; Freund, Robert Michael; Mirrokni Banadaki, Vahab Seyed
Download1806.02476.pdf (408.6Kb)
OPEN_ACCESS_POLICY
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
We introduce and study two algorithms to accelerate greedy coordinate descent in theory and in practice: Accelerated Semi-Greedy Coordinate Descent (ASCD) and Accelerated Greedy Co-ordinate Descent (AGCD). On the theory side, our main results are for ASCD: We show that ASCD achieves 0(l/k[superscript 2]) convergence, and it also achieves accelerated linear convergence for strongly convex functions. On the empirical side, while both AGCD and ASCD outperform Accelerated Randomized Coordinate Descent on most instances in our numerical experiments, we note that AGCD significantly outperforms the other two methods in our experiments, in spite of a lack of theoretical guarantees for this method. To complement this empirical finding for AGCD, we present an explanation why standard proof techniques for acceleration cannot work for AGCD, and we introduce a technical condition under which AGCD is guaranteed to have accelerated convergence. Finally, we confirm that this technical condition holds in our numerical experiments.
Date issued
2018-07Department
Massachusetts Institute of Technology. Department of Mathematics; Sloan School of ManagementJournal
Proceedings of Machine Learning Research
Publisher
Proceedings of Machine Learning Research
Citation
Lu, Haihao, Robert Freund, and Vahab Mirrokni. "Accelerating Greedy Coordinate Descent Methods." Proceedings of Machine Learning Research, 80 (2018): 3257-3266.
Version: Author's final manuscript