Andrej Karpathy describes in a recent video LLMs as lossy compressors, which helps explain why having more parameters is so important and why Small Language Models (SLMs) can't quite match their capabilities, even with better pre-training data.
Keep reading with a 7-day free trial
Subscribe to Text Generation to keep reading this post and get 7 days of free access to the full post archives.