Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I thought that was about LLMs being trained on compressed data. But I might be thinking about a different paper.


Gzip + kNN for text classification:

https://aclanthology.org/2023.findings-acl.426.pdf


Not LLM, just BERT, also did not actually outperform it.

source: https://kenschutte.com/gzip-knn-paper/




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: