Hacker News new | past | comments | ask | show | jobs | submit login

Seems pretty easy to mitigate. Just strip out invisible characters from input?





Easy enough for a human to understand but tricky for a computer

It’s not tricky for a computer to sanitize all io to/from an LLM. You can even build it into an inference engine itself to avoid mistakes. The article just shifts (not necessarily intentionally) a well-known problem with crappycode into AI FUD territory.

Did you read the full article?

> As researcher Thacker explained: The issue is they’re not fixing it at the model level, so every application that gets developed has to think about this or it's going to be vulnerable. And that makes it very similar to things like cross-site scripting and SQL injection, which we still see daily because it can’t be fixed at central location. Every new developer has to think about this and block the characters.


Why dont we fix it at the "API" level, then? I.e OpenAPI or Claude's API could do this for everyone. I know some people host their own models, but this would lower the attack surface.

Even still, that means every new API and use case that they build has to have this protection (eg Sora and ChatGPT api vs internal web api).

Seems perfectly reasonable to me.

We already have to protect against SQL and script injection, now we need to protect against Unicode injection.

Honestly surprised invisible Unicode characters haven't already been used for other types of attacks and that this is only an issue now.


Hosted models have APIs that will be patched, local models tend to use a handful of libraries that will also be patched. Yet Another Nothing Burger.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: