Seems like a good New Year activity for rationalists, so I thought I'd post it early instead of late.
Here are some steps I recommend:
- Go looking through old writings and messages.
- If you keep a journal, go look at a few entries from throughout the last year or two.
- Skim over your LessWrong activity from the last year.
- If you're active on a slack or discord, do a search of
from: [your username]
and skim your old messages from the last year (or the last 3 months on the free tier of slack). - Sample a few of your reddit comments and tweets from the last year.
- Same for text messages.
- Think back through major life events (if you had any this year) and see if they changed your mind about anything. Maybe you changed jobs or turned down an offer; maybe you tried a new therapeutic intervention or recreational drug; maybe you finally told your family something important; maybe you finally excommunicated someone from your life; maybe you tried mediating a conflict between your friends.
- Obvious, but look over your records of Manifold trades, quantitative predictions, and explicit bets. See if there's anything interesting.
Here are some emotional loadings that I anticipate seeing:
- "Well, that aged poorly."
- "Wow, bullet dodged!"
- "But I mean...how could I have known?"
- "Ah. Model uncertainty strikes again."
- "Yeah ok, but this year it'll happen for sure!"
- "[Sigh] I did have an inkling, but I guess I just didn't want to admit it at the time."
- "I tested that hypothesis and got a result."
- "Okay, they were right about this one thing. Doesn't mean I have to like them."
- "Now I see what people mean when they say X"
- "This is huge, why does no one talk about this?!"
At the beginning of the year I thought a decent model of how LLMs work was 10 years or so out. I’m now thinking it may be five years or less. What do I mean?
In the days of classical symbolic AI, researchers would use a programming language, often some variety of LISP, but not always, to implement a model of some set of linguistic structures and processes, such as those involved in story understanding and generation, or question answering. I see a similar division of conceptual labor in figuring out what’s going on inside LLMs. In this analogy I see mechanistic understanding as producing the equivalent of the programming languages of classical AI. These are the structures and mechanisms of the virtual machine that operates the domain model, where the domain is language in the broadest sense. I’ve been working on figuring out a domain model and I’ve had unexpected progress in the last month. I’m beginning to see how such models can be constructed. Call these domain models meta-models for LLMs.
It’s those meta models that I’m thinking are five years out. What would the scope of such a meta model be? I don’t know. But I’m not thinking in terms of one meta model that accounts for everything a given LLM can do. I’m thinking of more limited meta models. I figure that various communities will begin creating models in areas that interest them.
I figure we start with some hand-crafting to work out some standards. Then we’ll go to work on automating the process of creating the model. How will that work? I don’t know. Noone’s ever done it.
My confidence in this project has just gone up. It seems that I now have a collaborator. That is, he's familiar with my work in general and my investigations of ChatGPT in particular, we've had some email correspondence, and a couple of Zoom conversations. During today's conversation we decided to collaborate on a paper on the theme of 'demystifying LLMs.'
A word of caution. We haven't written the paper yet, so who knows? But all the signs are good. He's an expert on computer vision systems on the faculty of Goethe University in Frankfurt: Visvanathan... (read more)