“Nobody knows what makes humans so much more efficient”: Tiny language models based on Homo Sapiens could help explain how we learn and improve AI efficiency, for better or worse

Technology companies are shifting their focus from building larger language models (LLMs) to developing smaller models (SLMs) that can match or even surpass them.

Meta's Llama 3 (400 billion parameters), OpenAI's GPT-3.5 (175 billion parameters), and GPT-4 (about 1.8 trillion parameters) are famously larger models, while Microsoft's Phi-3 family ranges from 3.8 billion to 14 billion parameters, and Apple Intelligence “alone” has about 3 billion parameters.

scroll to top