Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: This work investigates the resilience of contemporary LLMs against frequent
and structured character-level perturbations, specifically through the
insertion of noisy characters after each input character. We introduce UCC-Inj,
a practical method that inserts invisible Unicode control characters into text
to discourage LLM misuse in scenarios such as online exam systems.
Surprisingly, despite strong obfuscation that fragments tokenization and
reduces the signal-to-noise ratio significantly, many LLMs still maintain
notable performance. Through comprehensive evaluation across model-, problem-,
and noise-related configurations, we examine the extent and mechanisms of this
robustness, exploring both the handling of character-level tokenization and
implicit versus explicit denoising mechanism hypotheses of character-level
noises. We hope our findings on the low-level robustness of LLMs will shed
light on the risks of their misuse and on the reliability of deploying LLMs
across diverse applications.
Authors (5)
Anyuan Zhuo
Xuefei Ning
Ningyuan Li
Yu Wang
Pinyan Lu
Submitted
October 16, 2025
Key Contributions
This work introduces UCC-Inj, a practical method for obfuscating text using invisible Unicode control characters to deter LLM misuse. It demonstrates that many LLMs maintain notable performance despite significant tokenization fragmentation and reduced signal-to-noise ratio, shedding light on LLM misuse risks and deployment reliability.
Business Value
Enhances the security of online platforms and systems that rely on LLMs by providing a method to prevent malicious use, such as cheating in online exams or generating harmful content.