MIT Libraries logoDSpace@MIT

MIT
View Item 
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Fast convolutional neural networks on FPGAs with hls4ml

Author(s)
Aarrestad, Thea; Loncar, Vladimir; Ghielmetti, Nicolò; Pierini, Maurizio; Summers, Sioni; Ngadiuba, Jennifer; Petersson, Christoffer; Linander, Hampus; Iiyama, Yutaro; Di Guglielmo, Giuseppe; Duarte, Javier; Harris, Philip; Rankin, Dylan; Jindariani, Sergo; Pedro, Kevin; Tran, Nhan; Liu, Mia; Kreinar, Edward; Wu, Zhenbin; Hoang, Duc; ... Show more Show less
Thumbnail
DownloadPublished version (2.183Mb)
Publisher with Creative Commons License

Publisher with Creative Commons License

Creative Commons Attribution

Terms of use
Creative Commons Attribution 4.0 International license https://creativecommons.org/licenses/by/4.0/
Metadata
Show full item record
Abstract
<jats:title>Abstract</jats:title> <jats:p>We introduce an automated tool for deploying ultra low-latency, low-power deep neural networks with convolutional layers on field-programmable gate arrays (FPGAs). By extending the <jats:monospace>hls4ml</jats:monospace> library, we demonstrate an inference latency of 5 <jats:italic>µ</jats:italic>s using convolutional architectures, targeting microsecond latency applications like those at the CERN Large Hadron Collider. Considering benchmark models trained on the Street View House Numbers Dataset, we demonstrate various methods for model compression in order to fit the computational constraints of a typical FPGA device used in trigger and data acquisition systems of particle detectors. In particular, we discuss pruning and quantization-aware training, and demonstrate how resource utilization can be significantly reduced with little to no loss in model accuracy. We show that the FPGA critical resource consumption can be reduced by 97% with zero loss in model accuracy, and by 99% when tolerating a 6% accuracy degradation.</jats:p>
Date issued
2021
URI
https://hdl.handle.net/1721.1/142113
Department
Massachusetts Institute of Technology. Department of Physics
Journal
Machine Learning: Science and Technology
Publisher
IOP Publishing
Citation
Aarrestad, Thea, Loncar, Vladimir, Ghielmetti, Nicolò, Pierini, Maurizio, Summers, Sioni et al. 2021. "Fast convolutional neural networks on FPGAs with hls4ml." Machine Learning: Science and Technology, 2 (4).
Version: Final published version

Collections
  • MIT Open Access Articles

Browse

All of DSpaceCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

My Account

Login

Statistics

OA StatisticsStatistics by CountryStatistics by Department
MIT Libraries
PrivacyPermissionsAccessibilityContact us
MIT
Content created by the MIT Libraries, CC BY-NC unless otherwise noted. Notify us about copyright concerns.