Tweeted By @DeepMind
Chinchilla: A 70 billion parameter language model that outperforms much larger models, including Gopher. By revisiting how to trade-off compute between model & dataset size, users can train a better and smaller model. Read more: https://t.co/RaZGUclBYQ 1/3 pic.twitter.com/TNWI1RLloA
— DeepMind (@DeepMind) April 12, 2022