40% off TNW Conference!

But the team found that other areas of LLM architecture didnt benefit as much from brute force.

Dont get me wrong, Gopher has significantly more parameters than GPT-3.

DeepMind’s new 280 billion-parameter language model kicks GPT-3’s butt in accuracy

Also tagged with

A figure from DeepMind’s press release