Steganographic Backdoor Attacks in NLP: Ultra-Low Poisoning and Defense Evasion
arXiv:2511.14301v2 Announce Type: replace-cross Abstract: Transformer models are foundational to natural language processing (NLP) applications, yet remain vulnerable to backdoor attacks introduced through poisoned data, which implant hidden behaviors during training. To strengthen the ability to prevent such compromises, recent…
