Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think OP is claiming that if enough people are using these obfuscators, the training data will be poisoned. The LLM being able to translate it right now is not a proof that this won't work, since it has enough "clean" data to compare against.




If enough people are doing that then venacular English has changed to be like that.

And it still isn't a problem for LLMs. There is sufficient history for it to learn on, and in any case low resource language learning shows them better than humans at learning language patterns.

If it follows an approximate grammar then an LLM will learn from it.


I don't mean people actually conversing like this on the internet, but using programs like what is in the article to feed it to the bots only.

This is exactly like those search engine traps people implemented in the late 90s and is roughly as effective.

But sure.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: