Building efficient algorithms by learning to compress
Author(s)Blalock, Davis W.(Davis Whitaker)
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science.
John V. Guttag.
MetadataShow full item record
The amount of data in the world is doubling every two years. Such abundant data offers immense opportunities, but also imposes immense computation, storage, and energy costs. This thesis introduces efficient algorithms for reducing these costs for bottlenecks in real world data analysis and machine learning pipelines. Concretely, we introduce algorithms for: -- Lossless compression of time series. This algorithm compresses better than any existing method, despite requiring only the resources available on a low-power edge device. -- Approximate matrix-vector multiplies. This algorithm accelerates approximate similarity scans by an order of magnitude relative to existing methods. -- Approximate matrix-matrix multiplies. This algorithm often outperforms existing approximation methods by more than 10x and non-approximate computation by more than 100x. We provide extensive empirical analyses of all three algorithms using real-world datasets and realistic workloads. We also prove bounds on the errors introduced by the two approximation algorithms. The theme unifying all of these contributions is learned compression. While compression is typically thought of only as a means to reduce data size, we show that specially designed compression schemes can also dramatically increase computation speed and reduce memory requirements.
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, September, 2020Cataloged from student-submitted PDF of thesis.Includes bibliographical references (pages 137-152).
DepartmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
Massachusetts Institute of Technology
Electrical Engineering and Computer Science.