PatrickVuscan
View original ↗Build a CLI or library that detects hidden malicious instructions in LLM prompts using unicode normalization and character analysis. This helps developers audit prompts and model inputs against adversarial obfuscation attempts.
Suggested repo: stealthcheck
"Your LLM is being lied to by invisible text; expose the hidden characters."
Estimated effort: 20h