GuppyLM: A 9-Million-Parameter LLM Built in 130 Lines of PyTorch That Trains in 5 Minutes on a Free GPU
A developer has built GuppyLM — a tiny but functional language model with 9 million parameters, trained on 60,000 synthetic conversations using a vanilla transformer architecture written in roughly 130 lines of PyTorch. It trains to conversational competence in about 5 minutes on a free Google Colab T4 GPU. The project has 892 upvotes on Hacker News from developers who say it is the clearest educational LLM implementation they have seen.