My model of what is going on with LLMs
Epistemic status: You probably already know if you want to read this kind of post, but in case you have not decided: my impression is that people are acting very confused about what we can conclude about scaling LLMs from the evidence, and I believe my mental model cuts through a lot of this confusion - I have tried to rebut what I believe to be misconceptions in a scattershot way, but will attempt to collect the whole picture here. I am a theoretical computer scientist and this is a theory. Soon I want to do some more serious empirical research around it - but be aware that most of my ideas about LLMs have not had the kind of careful, detailed contact with reality that I would like at the time of writing this post. If you're a good engineer (or just think I am dropping the ball somewhere) and are interested in helping dig into this please reach out. This post is not about timelines, though I think it has obvious implications for timelines. We have seen LLMs scale to impressively general performance. This does not mean they will soon reach human level because intelligence is not just a knob that needs to get turned further, it comprises qualitatively distinct functions. At this point it is not plausible that we can precisely predict how far we are from unlocking all remaining functions since it will probably require more insights. The natural guess is that the answer is on the scale of decades. It's important to take a step back and understand the history of how progress in A.I. takes place, following the main line of connectionist algorithms that (in hindsight, back-chaining from the frontier) are load-bearing. This story is relatively old and well-known, but I still need to retell it because I want to make a couple of points clear. First, deep learning has made impressive steps several times over the course of decades. Second, "blind scaling" has contributed substantially but has not been the whole story, conceptual insights piled on top of (and occasionally
Hmm, psychosecurity is an interesting reframing of epistemic rationality.