Hacker Newsnew | past | comments | ask | show | jobs | submit | fromlogin
Writing an LLM from scratch, part 25 – instruction fine-tuning (gilesthomas.com)
2 points by gpjt 16 hours ago | past | discuss
Writing an LLM from scratch, part 24 – the transcript hack (gilesthomas.com)
1 point by gpjt 1 day ago | past | discuss
Retro Language Models: Rebuilding Karpathy's RNN in PyTorch (gilesthomas.com)
1 point by ibobev 2 days ago | past | discuss
Writing an LLM from scratch, part 23 – fine-tuning for classification (gilesthomas.com)
1 point by ibobev 3 days ago | past | discuss
Retro Language Models: Rebuilding Karpathy's RNN in PyTorch (gilesthomas.com)
3 points by gpjt 5 days ago | past | discuss
Writing an LLM from scratch, part 23 – fine-tuning for classification (gilesthomas.com)
1 point by gpjt 7 days ago | past | discuss
Writing an LLM from scratch, part 22 – training our LLM (gilesthomas.com)
254 points by gpjt 14 days ago | past | 10 comments
Revisiting Karpathy's 'The Unreasonable Effectiveness of RNNs' (gilesthomas.com)
1 point by ibobev 17 days ago | past
Revisiting Karpathy's 'Unreasonable Effectiveness of Recurrent Neural Networks' (gilesthomas.com)
2 points by gpjt 19 days ago | past
Writing an LLM from scratch, part 21 – perplexed by perplexity (gilesthomas.com)
1 point by ibobev 21 days ago | past
Writing an LLM from scratch, part 21 – perplexed by perplexity (gilesthomas.com)
1 point by gpjt 22 days ago | past
Writing an LLM from scratch, part 20 – starting training, and cross entropy loss (gilesthomas.com)
41 points by gpjt 27 days ago | past | 3 comments
How Do LLMs Work? (gilesthomas.com)
2 points by gpjt 42 days ago | past | 1 comment
How Do LLMs Work? (gilesthomas.com)
1 point by ibobev 44 days ago | past
The maths you need to start understanding LLMs (gilesthomas.com)
616 points by gpjt 57 days ago | past | 120 comments
What AI chatbots are doing under the hood (gilesthomas.com)
2 points by gpjt 61 days ago | past
LLM from scratch, part 18 – residuals, shortcut connections, and the Talmud (gilesthomas.com)
2 points by gpjt 72 days ago | past
The fixed length bottleneck and the feed forward network (gilesthomas.com)
1 point by gpjt 76 days ago | past
Writing an LLM from scratch, part 17 – the feed-forward network (gilesthomas.com)
8 points by gpjt 78 days ago | past
Writing an LLM from scratch, part 16 – layer normalisation (gilesthomas.com)
1 point by gpjt 3 months ago | past
Leaving PythonAnywhere (gilesthomas.com)
3 points by gpjt 4 months ago | past
Writing an LLM from scratch, part 15 – from context vectors to logits (gilesthomas.com)
7 points by gpjt 5 months ago | past
Writing an LLM from scratch, part 14 – the complexity of self-attention at scale (gilesthomas.com)
1 point by gpjt 5 months ago | past
Writing an LLM from scratch, part 13 – attention heads are dumb (gilesthomas.com)
351 points by gpjt 5 months ago | past | 67 comments
Writing an LLM from scratch, part 12 – multi-head attention (gilesthomas.com)
3 points by gpjt 6 months ago | past
Writing an LLM from scratch, part 11 – batches (gilesthomas.com)
2 points by gpjt 6 months ago | past
Writing an LLM from scratch, part 10 – dropout (gilesthomas.com)
90 points by gpjt 7 months ago | past | 8 comments
Adding /Llms.txt (gilesthomas.com)
1 point by gpjt 7 months ago | past
Writing an LLM from scratch, part 9 – causal attention (gilesthomas.com)
4 points by gpjt 7 months ago | past
Writing an LLM from scratch, part 8 – trainable self-attention (gilesthomas.com)
380 points by gpjt 7 months ago | past | 31 comments

Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: