Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Llama and Candle are a lot more modern for these things than PyTorch/libtorch, though libtorch is still the de-facto standard.


That's wrong. Llama.cpp / Candle doesn't offer anything on the table that PyTorch cannot do (design wise). What they offer is smaller deployment footprint.

What's modern about LLM is the training infrastructure and single coordinator pattern, which PyTorch just started and inferior to many internal implementations: https://pytorch.org/blog/integration-idea-monarch/


Pytorch is still pretty dominant in cloud hosting. I’m not aware of anyone not using it (usually by way of vLLM or similar). It’s also completely dominant for training. I’m not aware of anyone using anything else.

It’s not dominant in terms of self-hosted where llama.cpp wins but there’s also not really that much self-hosting going on (at least compared with the amount of requests that hosted models are serving)




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: