I also got a Fatebook account thanks to this post.
This post lays out a bunch of tools that address what I've previously found lacking in personal forecasts, so thanks for the post! I've definitely gone observables-first, forecasted primarily the external world (rather than e.g. "if I do X, will I afterwards think it was a good thing to do?"), and have had the issue of feeling vaguely neutral about everything you touched on in Frame 3.
I'll now be trying these techniques out and see whether that helps.
...and as I wrote that sentence, I came to think about how Humans are not automatically strategic - particularly that we do not "ask ourselves what we’re trying to achieve" and "ask ourselves how we could tell if we achieved it" - and that this is precisely the type of thing you were using Fatebook for in this post. So, I actually sat down, thought about it and made a few forecasts:
And noticing that making these forecasts was cognitively heavy and not fluent at all, I made one more forecast:
So far I've made a couple of forecasts of the form "if I go to event X, will I think it was clearly worth it" that already resolved, and felt like I got useful data points to calibrate my expectations on.
Woo, great. :)
Whether this works out or not for you, I quite appreciate you laying out the details. Hope it's useful for you!
I think a basic possible concern (but likely minor in the long term, given sufficient rationality training and experience) is that making a lot of explicit predictions about things you have causal control over can have a self-fulfilling prophecy effect. So, for example, if you expect you won't be able to accomplish something, that belief could propagate through your mind and subconsciously make you not try as hard to achieve it (because you would expect you ultimately wouldn't be able to anyway), which later on makes your subjective estimation of the likelihood of your success go lower, so you try even less hard, etc. A self-reinforcing loop of anti-productivity.
Following this post, I made 4 forecasts on the output and impact of my MATS project, which led me to realize some outcomes I expected were less likely than I felt, absent active effort on my part to make them happen.
I got a Fatebook account at the beginning of 2024. After concluding the "Frictionless" section of this post, I felt encouraged to install the Firefox extension and make a prediction to share here.
Given that I haven't had time to read the full post in detail and that I feel that I want to get better at actively using predictions as a planning and calibration tool - and knowing that I may have time to dedicate to this activity within the next two weeks - I made the following, simple prediction.
⚖ Will I finish reading the post "Fluent, Cruxy Predictions" by Raemon within the next two weeks? (Nonmonotonica X: 65%)
Since I would like to take something concrete away from this post, I would want to find an area where setting up an action plan is relatively straightforward, so that I can focus on the prediction part. I already have few ideas, so I am quite confident that I will have found one by the time I will come back to this article:
⚖ Will I find a suitable area or goal where I can pair planning with prediction making using Fatebook and following Raemon's article? (Nonmonotonica X: 75%)
Thank you for taking the time to writing down this post.
If you do end up installing the Fatebook browser extension and making some predictions, let me know!
I did.
I believe that one of the effects of the current schooling system is training people to shy away from cruxyness, which makes you somewhat less of a liability for employment.
Update:
At the time I posted this, it was at a stage where I had just unlocked the skill, felt it's promise, but it hadn't really paid off yet. I feel like it has paid off in some important ways since then.
Some of my first round of "Important, Cruxy Predictions" (from 6 months ago, 3 months before I made this post), resolved recently. Many of them were of the form "6 months from now will I have recently used [various rationality techniques I was inventing]", of which "Fluent, Cruxy Predictions" was one.
I had given relatively low chances of most of them panning out. But, it turns out I've used basically my entire toolkit in the past ~week.
I'd also made an overall prediction of "a year from now (6 months ago) will the Lightcone team be using various tools." This had essentially been the motivating example behind the "Murphijitsu, and refusing to be satisfied with 'maybe'" section. I noticed I didn't feel optimistic about Lightcone adopting any of the stuff I was generating. Then I made a set of predictions where "in worlds where it was overwhelmingly obvious that these tools were going to end up helping the rest of Lightcone, what other observations would I make along the way?". This was helpful in clarifying further actions I needed to take for it to work (such as "find ways to mention it more often that felt helpful instead of overbearing").
I've more recently started workshopping a "group decisionmaking" UI that helps turn vague strategy disagreements into concrete disagreements. You can see a prototype spreadsheet here.
The latest in the Feedback Loop Rationality series.
Periodically, people (including me) try to operationalize predictions, or bets, and... it doesn't seem to help much.
I think I recently "got good" at making "actually useful predictions." I currently feel on-the-cusp of unlocking a host of related skills further down the rationality tech tree. This post will attempt to spell out some of the nuances of how I currently go about it, and paint a picture of why I think it's worth investing in.
The takeaway that feels most important to me is: it's way better to be "fluent" at operationalizing predictions, compared to "capable at all."
Previously, "making predictions" was something I did separately from my planning process. It was a slow, clunky process.
Nowadays, reasonably often I can integrate predictions into the planning process itself, because it feels lightweight. I'm better at quickly feeling-around for "what sort of predictions would actually change my plans if they turned out a certain way?", and then quickly check in on my intuitive sense of "what do I expect to happen?"
Fluency means you can actually use it day-to-day to help with whatever work is most important to you. Day-to-day usage means you can actually get calibrated for predictions in whatever domains you care about. Calibration means that your intuitions will be good, and you'll know they're good.
If I were to summarize the change-in-how-I-predict, it's a shift from:
"Observables-first". i.e. looking for things I could easily observe/operationalize, that were somehow related to what I cared about.
to:
"Cruxy-first". i.e. Look for things that would change my decisionmaking, even if vague, and learn to either better operationalize those vague things, or, find a way to get better data. (and then, there's a cluster of skills and shortcuts to make that easier)
Disclaimer:
This post is on the awkward edge of "feels full of promise", but "I haven't yet executed on the stuff that'd make it clearly demonstrably valuable." (And, I've tracked the results of enough "feels promise" stuff to know they have a <50% hit rate)
I feel like I can see the work needed to make this technique actually functional. It's a lot of work. I'm not sure if it's worth it. (Alas, I'm inventing this technique because I don't know how to tell if my projects are worth it and I really want a principled way of forming beliefs about that. Since I haven't finished vetting it yet it's hard to tell!)
There's a failure mode where rationality schools proliferate without evidence, where people get excited about their new technique that sounds good on paper, and publish exciting blogposts about it.
There's an unfortunate catch-22 where:
But, I nonetheless feel like the rationality community was overall healthier when it was full of excited people who still believed a bit in their heart that rationality would give them superpowers, and posted excitedly about it. It also seems intellectually valuable for notes to be posted along the way, so others can track it and experiment.
Rationality Training doesn't give you superpowers – it's a lot of work, and then "it's a pretty useful skill, among other skills." But I think it's an important skill if you are trying to problems you don't understand.
I expect, a year from now, I'll have a better conception of the ideas in this post. This post is still my best guess about how this skill can work, and I'd like it if other people excitedly tried to make it work, and report how it went.
Predicting Outcomes vs Comparing Plans
I'm using forecasting as one tool in a larger toolset. I'll have another post delving deeper into this, but basically: whenever I'm spending a bunch of resources on something, I try to:
The "fluent, cruxy predictions" skill is a tool for step 3 and 4. It's related to Murphijitsu (generally asking "what do I expect to go wrong here?", and then improving your plan, and then asking again until you feel like you plan is good enough).
Often, if I have an important uncertainty, instead of running an expensive experiment to resolve it, I can get a surprising amount of mileage out of just asking "well, what do I actually expect to happen here, if I force myself to get concrete and relevant?". The act of asking the question prompts my unconscious mind (i.e System 1 or Inner Simulator) to reveal what it actually believes when I'm not bullshitting yourself.
The trick here is using your deliberate, System 2 reasoning, to notice important questions that your System 1 would be good at answering.
But, your unconscious mind isn't magic. It doesn't know everything. So I think it's important to follow this up with making explicit predictions, that you check later, so your inner-simulator can get more accurate over time.
Subskills that go into this are:
Of those: training the skill of "operationalizing cruxy questions" is most important.
But, reducing friction is the easiest part to get traction on, so let's start there.
Frictionless
It should be as easy as possible to make a prediction. I particularly like the app Fatebook.io. It's very lightweight – I can open up the site, and my cursor will automatically be highlighted in the "new prediction" text box. I type in a prediction, tab over to the "forecast" input to enter a probability, and then hit return and boom, I'm done.
But I like it even more because:
For example, right now I'm writing a LessWrong doc. Are there any cruxy predictions I can make about this? What actually am I trying to accomplish with this post? After thinking a little, I hit Cmd-Shift-F, entered these predictions and then hit Cmd-V to paste them into the doc:
⚖ A week after posting this, between 1-3 people will have commented saying they got a Fatebook account because of this post (65%)
⚖ A week after posting this, 1-2 people will have actually made a chrome/firefox extension prediction that they paste into a comment. (50%)
Awhile later, I added:
⚖ A post (not by me) that gets 100+ karma will link approvingly to this post within 2 years (10%)
This is actually a decent example of how to use predictions, so I'll walk through what just went through my head. (This will be a little messy, because it's a real thought process. I think it's good to showcase what the real messiness might look like)
I started by asking myself "what actually am I even trying to do with this post?"
The most obvious goal with this post is to get people into the habit of making fluent, cruxy predictions. Ideally, I care that people are still doing it years later, skillfully. But I anticipate getting a very weak signal about that – I won't necessarily know who read the post. If I made Fatebook prediction like "a year from now, someone who read this post will have gone to develop a solid Fatebook habit", I wouldn't even know who to followup with to ask if it had worked.
As I thought about that, I started to feel despair – is this post really going to accomplish anything? I'm actually pretty pessimistic that many people will read this post and take action. And the longterm second-order effects feel pretty murky. This post is a fair amount of effort to write. Why am I even doing this?
After thinking a bit, two things occurred to me:
And that hopefully illustrates a major point of this whole "fluent cruxy prediction" concept: Asking "what effects do I expect to see from my actions?" often throws into relief that I don't really expect to see results from my actions. I'm just kinda running on autopilot. But, having noticed that, I can followup and ask "okay, what would actually need to be true, to be in the higher-likelihood of success world?"
With that in mind:
Hey there, reader.
If this feels at least somewhat compelling, what if you just got yourself to Fatebook right now, and make a couple predictions that'll resolve within couple days, or a week? Fatebook will send you emails reminding you about it, which can help bootstrap a habit.
Feeling around for "cruxiness"
The previous section started getting into this already. But now let's focus more explicitly: What sort of predictions, if true, would change your decisionmaking?
Cruxy operationalization is a murky artform. But here are some principles, and tricks I currently use. Most of my technique involves asking myself various questions, and then seeing what bubbles up.
Over time, I keep track of which question-prompts feel most useful.
Ideal Prerequisite:
Have at least two (plans/approaches/options)
I'll hopefully write up a whole blogpost about this someday. But one failure mode is "Well, I only really had one general idea. I made predictions about whether that idea would go well. Maybe they returned 'it'll go amazingly well!' and maybe they returned 'it'll go at least pretty well'. 'Pretty well' is at least 'pretty good.' I can't think of anything else to do, so... I guess I'll go with my original idea?"
Having at least two options, that are somewhat mutually exclusive, forces me to start thinking "okay, but how does my favorite plan compare to my other realistic options?".
It's important to find at least two options you really believe in. (Ideally: have at least one alternative that fits reasonably well into your current lifestyle, or strategic framework, and one that is more radically different).
It's a form of leaving a Line of Retreat, that helps you be more cognitively agile as you decide whether to pivot. If you're having trouble coming up with alternative plans, try Metastrategic Brainstorming.
Frames 1: Costs and benefits
So, one set of questions are "high level cost/benefit comparison?", i.e:
And then:
For me, these activate a different mode, that cuts more to what I actually care about.
Frame 2: Will this really help my deeper, longterm goals?
Often, my plans are really "step 1 of a much larger plan", or, a cluster of strategies that other people in the world are working towards.
Frame 3: Murphijitsu, and being dissatisfied with "maybe"
Often while thinking about the first two sets of questions, I find my intuitive sense of "will my project be successful?" is "... kinda?". Either the outcome I'm predicting feels uncertain ("~55%?"), or it's hard to get that clear a visualizable outcome at all.
I then ask "okay, so, if this was definitely, clearly a resounding success, a year from now, what are all the things I would see?"
The way this works best at first sometimes feels like unrealistic dreaming, but then often highlights that the dream is maybe achievable (but perhaps a lot more work than I'd initially envisioned). I generate specific new intermediate actions I might want to plan for. Then, I go back to the original prediction and see if I expect greater success.
This is pretty similar to the process of Murphijitsu (asking "am I surprised if this goes wrong?", and if you wouldn't be too surprised, iterating on your plan until you'd feel actively surprised if it didn't work). The two novel components I'm adding here are:
The feeling of "oh... that's cruxy"
It's possible to do "fake work" – where you come up with some defensible sounding predictions, but they still don't actually cleave at your cruxes or help you think.
One thing I particularly look for is a feeling of "oh man, that's cruxy." For me it often comes with a feeling of destabilization/vertigo.
I remember once a few years ago when I was arguing with someone about whether empathy was useful. I was like "Empathy isn't just nice. It makes people more effective so they'll be better at their job." My conversation partner (habryka) said "do you really think Elon Musk would do better at his job if he was more empathetic", and I felt a sinking feeling of "oh no, my beliefs just became falsifiable, and, I'm not sure how had I'd actually bet on them."
It's hard to actually resolve a bet on whether empathy would help Elon Musk, but, operationalizing the bet, even if it's a hypothetical fantasy bet, helps boot up my realistic intuitions, rather than abstract hopes about "surely this 'make people more empathetic plan will be good' because empathy is generically good."
Example
The Fractal Strategy Workshop
While I was leading up to my Fractal Strategy Workshop, 10 days beforehand, my colleague and I sat down to make some predictions about the workshop.
A naive thing we might have written down is "after the workshop, someone will use a technique we taught them." But this wouldn't be very concrete. How would you know when to check? What counts as using a technique from the workshop?
The things my colleague wrote down tended to be questions like:
Meanwhile I was making predictions more like:
In some sense, my colleague's questions are quite reasonable. If we aren't seeing participants use the techniques from the workshop, or generally improve their planmaking, the workshop can't be that useful.
But, that wasn't that cruxy for my decisionmaking. I know that the first workshop I run will probably have lots of problems. I also don't really expect a workshop to work in isolation (what I hoped would work was workshop + 6 months of followup coaching to help cement the skills and collaboratively iterate on applying them to the participant's lives). I was deliberately not committing to doing the followup coaching for this workshop, because it was an early beta test. The purpose was to get a sense of whether the curriculum was roughly pointed in the right direction, and decide whether to do a full workshop+coaching later on.
It is cruxy to me whether, if I worked on it for a year, the metastrategy framework would produce clear results. But for the immediate future, the more relevant question was "will the major stakeholders of this project think it is worth paying for?". If people aren't willing to pay significant money for the workshops, that's evidence that they don't expect them to be seriously valuable. And if my boss didn't think it made sense, I'd need to raise money from somewhere and strike out on my own.
This shaped my decisionmaking in a few ways. I realized:
(It's worth noting: my colleague had the opposite impression. They thought it was relatively easy to get funding, they were much more worried about the metastrategy practice failing to impact people. Which is reasonable! But, part of the point here is that what matters is what is cruxy for you. They're your decisions and plans!)
Tips and Tricks
Practice via Videogames
Real life takes a long time to give you feedback on your predictions. I've found videogames a good testbed to practice making predictions, in a way that feels fun, rapid-feedback, while connected to a "real-ish decisionmaking process."
I recommend using Turn-Based games that you haven't played before, since those give you lots of opportunities to pause and think.
At first, simply focus on making as many predictions as you can about the game's mechanics, to drill the basic move of "notice that there are multiple plausible worlds that you're living in, and that you have some useful intuitions about it even if you're missing lots of information."
Once you get a basic handle on it, try to specifically ask "what are some unknowns in this game that would affect my strategy?". When you're about to make a choice, consider "what could turn out to be true, that'd lead me to wish I'd made a different choice?".
NOTES for Prediction-Followup
Often, I'll notice "hmm, I could make a prediction about that" while I'm in the middle of a longer thought process. Something I find helpful is to jot down a note about it somewhere I can easily come back to after the train-of-thought completes. (If I'm in a brainstorming document, literally write something like PREDICTION, in capital letters).
The most common note I use is the word "PROMISING," which specifically refers to me having an idea that feels really compelling. (I developed this while practicing solving Baba is You levels, and making predictions about whether my current problem-solving process was on the right track).
I encourage you to develop your own shorthand, based on the mental qualia that are useful for you.
Stuck on making a perfect prediction? Make an imperfect one
(Relatedly: it's okay to make fuzzy predictions that only make sense to you, but when doing so, try to make "extremized" ones)
Sometimes I want to decision – quitting a job, or deciding to do a complex plan. If I can't pin down a concrete observable thing, I can always default to "Subjectively, a year from now, I'll think it was pretty clearly correct to have [done X]."
But, the words "pretty obviously" there are important.
Oftentimes when I ask "was it a good decision [to quite that job / to start that new relationship / etc]?" the answer is "...maybe? Yeah, kinda?". And that's too vague to be useful as a resolution criteria.
But if you set the standard for "it has to be pretty clearly correct", not just "probably correct", I find that leaves less room for hemming/hawing about it.
Make multiple predictions
Relatedly: If you're not sure if one prediction captures The True Spirit of What You're Uncertain about, try making a few different predictions on the same topic.
If I'm considering leaving my job, I might separately try asking things like:
Takeaways
We just covered a lot of stuff. A quick recap of some of the key concepts:
To reach fluency, you need some way of getting in lots of practice. The advice I'd give depends somewhat on where you're starting from. If you're totally new to predictions, I think it's necessary to first get a basic fluency with "making predictions, at all." Make tons of predictions about things you care about in your life, that will resolve reasonably soon. (i.e. days or weeks). Make tons of predictions about videogames or other fun domains with uncertainty and fast-feedback.
But, the thing to ultimately be aiming for is to hit a kind of "escape velocity", with a combination of skills make predictions useful for your day job, or major personal projects. For that, the skill of identifying and operationalizing "strategic cruxiness" is an important building block.
If you do end up installing the Fatebook browser extension and making some predictions, let me know! It is helpful to know when people actually take actions based on posts like this.