Show simple item record

dc.contributor.authorXu, Lei
dc.contributor.authorAlnegheimish, Sarah
dc.contributor.authorBerti‐Equille, Laure
dc.contributor.authorCuesta‐Infante, Alfredo
dc.contributor.authorVeeramachaneni, Kalyan
dc.date.accessioned2025-10-22T16:38:19Z
dc.date.available2025-10-22T16:38:19Z
dc.date.issued2025-07-07
dc.identifier.urihttps://hdl.handle.net/1721.1/163365
dc.description.abstractIn text classification, creating an adversarial example means subtly perturbing a few words in a sentence without changing itsmeaning, causing it to be misclassified by a classifier. A concerning observation is that a significant portion of adversarial exam-ples generated by existing methods change only one word. This single-word perturbation vulnerability represents a significantweakness in classifiers, which malicious users can exploit to efficiently create a multitude of adversarial examples. This paperstudies this problem and makes the following key contributions: (1) We introduce a novel metric 𝜌 to quantitatively assess a clas-sifier's robustness against single-word perturbation. (2) We present the SP-Attack, designed to exploit the single-word perturbationvulnerability, achieving a higher attack success rate, better preserving sentence meaning, while reducing computation costscompared to state-of-the-art adversarial methods. (3) We propose SP-Defence, which aims to improve 𝜌 by applying data augmen-tation in learning. Experimental results on 4 datasets and 2 masked language models show that SP-Defence improves 𝜌 by 14.6%and 13.9% and decreases the attack success rate of SP-Attack by 30.4% and 21.2% on two classifiers respectively, and decreasesthe attack success rate of existing attack methods that involve multiple-word perturbation.en_US
dc.language.isoen
dc.publisherWileyen_US
dc.relation.isversionofhttps://doi.org/10.1111/exsy.70079en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceWileyen_US
dc.titleSingle Word Change Is All You Need: Using LLMs to Create Synthetic Training Examples for Text Classifiersen_US
dc.typeArticleen_US
dc.identifier.citationXu, L., S. Alnegheimish, L. Berti-Equille, A. Cuesta-Infante, and K. Veeramachaneni. 2025. “ Single Word Change Is All You Need: Using LLMs to Create Synthetic Training Examples for Text Classifiers.” Expert Systems 42, no. 8: e70079.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Laboratory for Information and Decision Systemsen_US
dc.relation.journalExpert Systemsen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-10-22T16:30:03Z
dspace.orderedauthorsXu, L; Alnegheimish, S; Berti‐Equille, L; Cuesta‐Infante, A; Veeramachaneni, Ken_US
dspace.date.submission2025-10-22T16:30:04Z
mit.journal.volume42en_US
mit.journal.issue8en_US
mit.licensePUBLISHER_CC


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record