Richard Plant
Evaluating Language Model Vulnerability to Poisoning Attacks in Low-Resource Settings
Plant, Richard; Giuffrida, Mario Valerio; Pitropakis, Nikolaos; Gkatzia, Dimitra
Authors
Dr VALERIO GIUFFRIDA VALERIO.GIUFFRIDA@NOTTINGHAM.AC.UK
Assistant Professor in Computer Vision
Nikolaos Pitropakis
Dimitra Gkatzia
Abstract
Pre-trained language models are a highly effective source of knowledge transfer for natural language processing tasks, as their development represents an investment of resources beyond the reach of most researchers and end users. The widespread availability of such easily adaptable resources has enabled high levels of performance, which is especially valuable for low-resource language users who have typically been overlooked when it comes to NLP applications. However, these models introduce vulnerabilities in NLP toolchains, since they may prove vulnerable to attacks from malicious actors with access to the data used for downstream training. By perturbing instances from the training set, such attacks seek to undermine model capabilities and produce radically different outcomes during inference. We show that adversarial data manipulation has a severe effect on model performance, with BERT's performance dropping by more than 30% on average across all tasks at a poisoning ratio greater than 50%. Additionally, we conduct the first evaluation of this kind in the Basque language domain, establishing the vulnerability of low-resource models to the same form of attack.
Citation
Plant, R., Giuffrida, M. V., Pitropakis, N., & Gkatzia, D. (2025). Evaluating Language Model Vulnerability to Poisoning Attacks in Low-Resource Settings. IEEE Transactions on Audio, Speech and Language Processing, 33, 54-67. https://doi.org/10.1109/taslp.2024.3507565
Journal Article Type | Article |
---|---|
Acceptance Date | Nov 21, 2024 |
Online Publication Date | Nov 28, 2024 |
Publication Date | 2025 |
Deposit Date | Jun 2, 2025 |
Publicly Available Date | Jun 4, 2025 |
Journal | IEEE Transactions on Audio, Speech and Language Processing |
Print ISSN | 1558-7916 |
Publisher | Institute of Electrical and Electronics Engineers |
Peer Reviewed | Peer Reviewed |
Volume | 33 |
Pages | 54-67 |
DOI | https://doi.org/10.1109/taslp.2024.3507565 |
Keywords | language modelling, machine learning methods for hlt, language understanding and computational semantics |
Public URL | https://nottingham-repository.worktribe.com/output/44826414 |
Publisher URL | https://ieeexplore.ieee.org/document/10771712 |
Files
LLM_poisoning_taslp AAM
(1.3 Mb)
PDF
You might also like
Cloud Detection Challenge - Methods and Results
(2025)
Journal Article
Synchronization Is All You Need: Exocentric-to-Egocentric Transfer for Temporal Action Segmentation with Unlabeled Synchronized Video Pairs
(2024)
Presentation / Conference Contribution
TADM: Temporally-Aware Diffusion Model for Neurodegenerative Progression on Brain MRI
(2024)
Presentation / Conference Contribution
Adapting Vision Foundation Models for Plant Phenotyping
(2023)
Presentation / Conference Contribution
Downloadable Citations
About Repository@Nottingham
Administrator e-mail: discovery-access-systems@nottingham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search