Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

GGUF is just a file format. The ability to offload some layers to CPU is not specific to it nor to llama.cpp in general - indeed, it was available before llama.cpp was even a thing.


I'm pretty sure I didn't assert that it was more that a file format or that llama.cpp was a pioneer in that regard?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: