Extrapolating GPT-N performance

In 2020, I wrote a post extrapolating how fast language models would improve on certain benchmarks, if they were scaled up. I've since followed-up with a couple of posts on how well those extrapolations seem to be holding up.

Read the last post for graphs with the most data-points on them. Read the first post for details and thoughts about how this exercise might be useful.