Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> LLMs can now detect garbage much more cheaply than humans can.

Off the top of my head, I don't think this is true for training data. I could be wrong, but it seems very fallible to let GPT-5 be the source of ground truth for GPT-6.





I dotn think an LLM even can detect garbage during a training run. While training the system is only tasked with predicting the next token in the training set, it isn't trying to reason about the validity of the training set itself.

You can triage with an LLM, at least. Throw away the obvious junk, have a human look at anything doubtful.

Llm-as-a-judge has been working well for years now.

RL from LLMs works.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: