Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI researchers,Security professionals,Developers of LLM-based applications,Educators 3 weeks ago

On the Ability of LLMs to Handle Character-Level Perturbations: How Well and How?

large-language-models › robustness
📄 Abstract

Abstract: This work investigates the resilience of contemporary LLMs against frequent and structured character-level perturbations, specifically through the insertion of noisy characters after each input character. We introduce UCC-Inj, a practical method that inserts invisible Unicode control characters into text to discourage LLM misuse in scenarios such as online exam systems. Surprisingly, despite strong obfuscation that fragments tokenization and reduces the signal-to-noise ratio significantly, many LLMs still maintain notable performance. Through comprehensive evaluation across model-, problem-, and noise-related configurations, we examine the extent and mechanisms of this robustness, exploring both the handling of character-level tokenization and implicit versus explicit denoising mechanism hypotheses of character-level noises. We hope our findings on the low-level robustness of LLMs will shed light on the risks of their misuse and on the reliability of deploying LLMs across diverse applications.
Authors (5)
Anyuan Zhuo
Xuefei Ning
Ningyuan Li
Yu Wang
Pinyan Lu
Submitted
October 16, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This work introduces UCC-Inj, a practical method for obfuscating text using invisible Unicode control characters to deter LLM misuse. It demonstrates that many LLMs maintain notable performance despite significant tokenization fragmentation and reduced signal-to-noise ratio, shedding light on LLM misuse risks and deployment reliability.

Business Value

Enhances the security of online platforms and systems that rely on LLMs by providing a method to prevent malicious use, such as cheating in online exams or generating harmful content.