| dc.contributor.author | Xu, Lei | |
| dc.contributor.author | Alnegheimish, Sarah | |
| dc.contributor.author | Berti‐Equille, Laure | |
| dc.contributor.author | Cuesta‐Infante, Alfredo | |
| dc.contributor.author | Veeramachaneni, Kalyan | |
| dc.date.accessioned | 2025-10-22T16:38:19Z | |
| dc.date.available | 2025-10-22T16:38:19Z | |
| dc.date.issued | 2025-07-07 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/163365 | |
| dc.description.abstract | In text classification, creating an adversarial example means subtly perturbing a few words in a sentence without changing itsmeaning, causing it to be misclassified by a classifier. A concerning observation is that a significant portion of adversarial exam-ples generated by existing methods change only one word. This single-word perturbation vulnerability represents a significantweakness in classifiers, which malicious users can exploit to efficiently create a multitude of adversarial examples. This paperstudies this problem and makes the following key contributions: (1) We introduce a novel metric 𝜌 to quantitatively assess a clas-sifier's robustness against single-word perturbation. (2) We present the SP-Attack, designed to exploit the single-word perturbationvulnerability, achieving a higher attack success rate, better preserving sentence meaning, while reducing computation costscompared to state-of-the-art adversarial methods. (3) We propose SP-Defence, which aims to improve 𝜌 by applying data augmen-tation in learning. Experimental results on 4 datasets and 2 masked language models show that SP-Defence improves 𝜌 by 14.6%and 13.9% and decreases the attack success rate of SP-Attack by 30.4% and 21.2% on two classifiers respectively, and decreasesthe attack success rate of existing attack methods that involve multiple-word perturbation. | en_US |
| dc.language.iso | en | |
| dc.publisher | Wiley | en_US |
| dc.relation.isversionof | https://doi.org/10.1111/exsy.70079 | en_US |
| dc.rights | Creative Commons Attribution | en_US |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | en_US |
| dc.source | Wiley | en_US |
| dc.title | Single Word Change Is All You Need: Using LLMs to Create Synthetic Training Examples for Text Classifiers | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Xu, L., S. Alnegheimish, L. Berti-Equille, A. Cuesta-Infante, and K. Veeramachaneni. 2025. “ Single Word Change Is All You Need: Using LLMs to Create Synthetic Training Examples for Text Classifiers.” Expert Systems 42, no. 8: e70079. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Laboratory for Information and Decision Systems | en_US |
| dc.relation.journal | Expert Systems | en_US |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/JournalArticle | en_US |
| eprint.status | http://purl.org/eprint/status/PeerReviewed | en_US |
| dc.date.updated | 2025-10-22T16:30:03Z | |
| dspace.orderedauthors | Xu, L; Alnegheimish, S; Berti‐Equille, L; Cuesta‐Infante, A; Veeramachaneni, K | en_US |
| dspace.date.submission | 2025-10-22T16:30:04Z | |
| mit.journal.volume | 42 | en_US |
| mit.journal.issue | 8 | en_US |
| mit.license | PUBLISHER_CC | |