AI ALIGNMENT FORUMTags
AF

Scaling Laws

EditHistorySubscribe
Discussion (0)
Help improve this page
EditHistorySubscribe
Discussion (0)
Help improve this page
Scaling Laws
Random Tag
Contributors
1plex

Scaling Laws refer to the observed trend of some machine learning architectures (notably transformers) to scale their performance on predictable power law when given more compute, data, or parameters (model size), assuming they are not bottlenecked on one of the other resources. This has been observed as highly consistent over more than six orders of magnitude.

Scaling laws graph from Scaling Laws for Neural Language Models
Posts tagged Scaling Laws
Most Relevant
0
29Thoughts on the Alignment Implications of Scaling Language Models
leogao
1y
2
1
19NVIDIA and Microsoft releases 530B parameter transformer model, Megatron-Turing NLG
Ozyrus
7mo
1
1
15[Link] Training Compute-Optimal Large Language Models
nostalgebraist
2mo
7
0
48How much chess engine progress is about adapting to bigger computers?
Paul Christiano
10mo
8
1
32Compute Trends Across Three eras of Machine Learning
Jaime Sevilla, Pablo Villalobos, Lennart Heim, Marius Hobbhahn, Tamay Besiroglu, Anson Ho
3mo
5
1
15Parameter counts in Machine Learning
Jaime Sevilla, Pablo Villalobos
1y
8
0
15Estimating training compute of Deep Learning models
Lennart Heim, Jaime Sevilla, Marius Hobbhahn, Tamay Besiroglu, Anson Ho
4mo
4
1
8How to measure FLOP/s for Neural Networks empirically?
Marius Hobbhahn
6mo
0
Add Posts