On the Ability of LLMs to Handle Character-Level Perturbations: How Well and How?
Anyuan Zhuo 1, Xuefei Ning 2, Ningyuan Li 2, Jingyi Zhu 1, Yu Wang 2, Pinyan Lu 1
Published on arXiv
2510.14365
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Despite heavy Unicode character injection that fragments tokenization and drastically reduces signal-to-noise ratio, many contemporary LLMs maintain notable performance, undermining UCC-Inj as an effective anti-cheating defense.
UCC-Inj
Novel technique introduced
This work investigates the resilience of contemporary LLMs against frequent and structured character-level perturbations, specifically through the insertion of noisy characters after each input character. We introduce UCC-Inj, a practical method that inserts invisible Unicode control characters into text to discourage LLM misuse in scenarios such as online exam systems. Surprisingly, despite strong obfuscation that fragments tokenization and reduces the signal-to-noise ratio significantly, many LLMs still maintain notable performance. Through comprehensive evaluation across model-, problem-, and noise-related configurations, we examine the extent and mechanisms of this robustness, exploring both the handling of character-level tokenization and implicit versus explicit denoising mechanism hypotheses of character-level noises. We hope our findings on the low-level robustness of LLMs will shed light on the risks of their misuse and on the reliability of deploying LLMs across diverse applications.
Key Contributions
- Introduces UCC-Inj, a practical Unicode control character injection method that inserts invisible characters after each input character to discourage LLM misuse in exam systems
- Provides comprehensive evaluation of LLM robustness to character-level perturbations across model, problem, and noise configurations, revealing that many LLMs maintain notable performance despite severe tokenization fragmentation
- Analyzes the mechanisms behind LLM robustness to character-level noise, exploring implicit vs. explicit denoising hypotheses and tokenization handling