670: LLaMA: GPT-3 performance, 10x smaller

670: LLaMA: GPT-3 performance, 10x smaller

Super Data Science: ML & AI Podcast with Jon Krohn

How does Meta AI's natural language model, LLaMa compare to the rest? Based on the Chinchilla scaling laws, LLaMa is designed to be smaller but more performant. But how exactly does it achieve this feat? It's all done by…

Related tracks

See all