Show simple item record

dc.contributor.authorSpangher, Lucas
dc.contributor.authorBonotto, Matteo
dc.contributor.authorArnold, William
dc.contributor.authorChayapathy, Dhruva
dc.contributor.authorGallingani, Tommaso
dc.contributor.authorSpangher, Alexander
dc.contributor.authorCannarile, Francesco
dc.contributor.authorBigoni, Daniele
dc.contributor.authorde Marchi, Eliana
dc.contributor.authorRea, Cristina
dc.date.accessioned2025-11-19T15:44:47Z
dc.date.available2025-11-19T15:44:47Z
dc.date.issued2025-05-24
dc.identifier.urihttps://hdl.handle.net/1721.1/163763
dc.description.abstractPlasma disruptions remain a major obstacle to sustained commercial operation of tokamak-based fusion devices. Although machine learning (ML) methods have shown promise for predicting disruptions, their performance and generalizability suffer from a lack of common benchmarks and comprehensive multi-device evaluations. To address this, we present DisruptionBench, a new benchmarking platform designed to standardize how ML-driven disruption prediction systems are trained and evaluated on multi-machine data. DisruptionBench spans three devices - Alcator C-Mod, DIII-D, and EAST - and includes tasks of varying difficulty: zero-shot, few-shot, and many-shot training regimes to assess each model’s ability to transfer learned representations to new or data-limited machines. We evaluate four state-of-the-art ML architectures. Two are re-implementations of notable prior work: a random forest (Cristina Rea in PPCF 60:084008, 2018) and the Hybrid Deep Learner (HDL) (Zhu in NC 61: 026607, 2020). We also propose two new approaches tailored for disruption prediction: a transformer-based model inspired by GPT-2, capable of learning long-range temporal dependencies through self-attention, and a Continuous Convolutional Neural Network (CCNN) that leverages continuous kernels to capture subtle variations in plasma signals. Across the nine benchmarking tasks, the CCNN demonstrates consistently strong performance and achieves the highest overall Area Under the ROC Curve (AUC) in intra-machine tests (up to 0.97 on C-Mod). Nevertheless, the GPT-2-based approach and HDL can outperform CCNN in specific transfer scenarios, particularly when the test machine is underrepresented in training data. We further analyze the significance of memory length in capturing precursor phenomena, providing evidence that longer context windows can boost predictive accuracy.en_US
dc.publisherSpringer USen_US
dc.relation.isversionofhttps://doi.org/10.1007/s10894-025-00495-2en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceSpringer USen_US
dc.titleDisruptionBench and Complimentary New Models: Two Advancements in Machine Learning Driven Disruption Predictionen_US
dc.typeArticleen_US
dc.identifier.citationSpangher, L., Bonotto, M., Arnold, W. et al. DisruptionBench and Complimentary New Models: Two Advancements in Machine Learning Driven Disruption Prediction. J Fusion Energ 44, 26 (2025).en_US
dc.contributor.departmentMassachusetts Institute of Technology. Plasma Science and Fusion Centeren_US
dc.relation.journalJournal of Fusion Energyen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-07-18T15:31:49Z
dc.language.rfc3066en
dc.rights.holderThe Author(s)
dspace.embargo.termsN
dspace.date.submission2025-07-18T15:31:49Z
mit.journal.volume44en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record