Show simple item record

dc.contributor.authorWu, Ziniu
dc.contributor.authorMarcus, Ryan
dc.contributor.authorLiu, Zhengchun
dc.contributor.authorNegi, Parimarjan
dc.contributor.authorNathan, Vikram
dc.contributor.authorPfeil, Pascal
dc.contributor.authorSaxena, Gaurav
dc.contributor.authorRahman, Mohammad
dc.contributor.authorNarayanaswamy, Balakrishnan
dc.contributor.authorKraska, Tim
dc.date.accessioned2024-07-23T20:14:08Z
dc.date.available2024-07-23T20:14:08Z
dc.date.issued2024-06-09
dc.identifier.isbn979-8-4007-0422-2
dc.identifier.urihttps://hdl.handle.net/1721.1/155774
dc.description.abstractQuery performance (e.g., execution time) prediction is a critical component of modern DBMSes. As a pioneering cloud data warehouse, Amazon Redshift relies on an accurate execution time prediction for many downstream tasks, ranging from high-level optimizations, such as automatically creating materialized views, to low-level tasks on the critical path of query execution, such as admission, scheduling, and execution resource control. Unfortunately, many existing execution time prediction techniques, including those used in Redshift, suffer from cold start issues, inaccurate estimation, and are not robust against workload/data changes. In this paper, we propose a novel hierarchical execution time predictor: the Stage predictor. The Stage predictor is designed to leverage the unique characteristics and challenges faced by Redshift. The Stage predictor consists of three model states: an execution time cache, a lightweight local model optimized for a specific DB instance with uncertainty measurement, and a complex global model that is transferable across all instances in Redshift. We design a systematic approach to use these models that best leverages optimality (cache), instance-optimization (local model), and transferable knowledge about Redshift (global model). Experimentally, we show that the Stage predictor makes more accurate and robust predictions while maintaining a practical inference latency and memory overhead. Overall, the Stage predictor can improve the average query execution latency by 20% on these instances compared to the prior query performance predictor in Redshift.en_US
dc.publisherACM|Companion of the 2024 International Conference on Management of Dataen_US
dc.relation.isversionof10.1145/3626246.3653391en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleStage: Query Execution Time Prediction in Amazon Redshiften_US
dc.typeArticleen_US
dc.identifier.citationWu, Ziniu, Marcus, Ryan, Liu, Zhengchun, Negi, Parimarjan, Nathan, Vikram et al. 2024. "Stage: Query Execution Time Prediction in Amazon Redshift."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-07-01T07:54:29Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-07-01T07:54:30Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record