benchmark 2025

On the Ability of LLMs to Handle Character-Level Perturbations: How Well and How?

Anyuan Zhuo 1, Xuefei Ning 2, Ningyuan Li 2, Jingyi Zhu 1, Yu Wang 2, Pinyan Lu 1

0 citations · 17 references · arXiv

α

Published on arXiv

2510.14365

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

Despite heavy Unicode character injection that fragments tokenization and drastically reduces signal-to-noise ratio, many contemporary LLMs maintain notable performance, undermining UCC-Inj as an effective anti-cheating defense.

UCC-Inj

Novel technique introduced


This work investigates the resilience of contemporary LLMs against frequent and structured character-level perturbations, specifically through the insertion of noisy characters after each input character. We introduce UCC-Inj, a practical method that inserts invisible Unicode control characters into text to discourage LLM misuse in scenarios such as online exam systems. Surprisingly, despite strong obfuscation that fragments tokenization and reduces the signal-to-noise ratio significantly, many LLMs still maintain notable performance. Through comprehensive evaluation across model-, problem-, and noise-related configurations, we examine the extent and mechanisms of this robustness, exploring both the handling of character-level tokenization and implicit versus explicit denoising mechanism hypotheses of character-level noises. We hope our findings on the low-level robustness of LLMs will shed light on the risks of their misuse and on the reliability of deploying LLMs across diverse applications.


Key Contributions

  • Introduces UCC-Inj, a practical Unicode control character injection method that inserts invisible characters after each input character to discourage LLM misuse in exam systems
  • Provides comprehensive evaluation of LLM robustness to character-level perturbations across model, problem, and noise configurations, revealing that many LLMs maintain notable performance despite severe tokenization fragmentation
  • Analyzes the mechanisms behind LLM robustness to character-level noise, exploring implicit vs. explicit denoising hypotheses and tokenization handling

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
inference_timeblack_box
Applications
online exam systemsllm misuse preventiontext obfuscation