少点错误 2024年07月11日
Fluent, Cruxy Predictions
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了将预测融入日常决策的技巧,作者分享了其从“观察结果优先”到“关键因素优先”的转变,以及如何通过流畅的预测来提升决策效率。文章还介绍了“Fatebook”等工具,以及如何识别关键因素、制定可操作的预测,并通过预测结果的反馈来不断提升内在模拟器的准确性。

🤔 作者从“观察结果优先”转变为“关键因素优先”,不再仅仅关注可观察的结果,而是着重于寻找那些能够改变决策的关键因素。这种转变使得预测更加实用,并能有效地引导决策方向。

💡 作者认为,流畅的预测意味着能够将预测融入日常决策过程,并通过持续的实践来校准内在模拟器,使其更加准确地反映现实情况。

🚀 作者分享了“Fatebook”等工具,这些工具可以帮助用户轻松记录预测,并通过邮件提醒等功能来培养预测习惯。

🎯 作者强调了“关键因素”的重要性,并建议用户在制定预测时,要寻找那些能够显著影响决策结果的关键因素。

🧠 作者认为,预测的准确性并非一蹴而就,需要不断地练习和校准。通过不断地观察预测结果,并根据结果调整内在模拟器,可以逐步提升预测的准确性。

💪 作者还强调了预测过程中的“摩擦力”问题,并建议用户使用轻量级工具,并简化预测的记录流程,以降低预测的“摩擦力”,提高预测的效率。

Published on July 10, 2024 5:36 AM GMT

Periodically, people (including me) try to operationalize predictions, or bets, and... it doesn't seem to help much.

I think I recently "got good" at making "actually useful predictions." I currently feel on-the-cusp of unlocking a host of related skills further down the rationality tech tree. This post will attempt to spell out some of the nuances of how I currently go about it, and paint a picture of why I think it's worth investing in.

The takeaway that feels most important to me is: it's way better to be "fluent" at operationalizing predictions, compared to "capable at all."

Previously, "making predictions" was something I did separately from my planning process. It was a slow, clunky process.

Nowadays, reasonably often I can integrate predictions into the planning process itself, because it feels lightweight. I'm able to quickly feel-around for "what sort of predictions would actually change my plans if they turned out a certain way?", and then quickly check in on my intuitive sense of "what do I expect to happen?"

Fluency means you can actually use it day-to-day to help with whatever work is most important to you. Day-to-day usage means you can actually get calibrated for predictions in whatever domains you care about. Calibration means that your intuitions will be good, and you'll know they're good

If I were to summarize the change-in-how-I-predict, it's a shift from:

"Observables-first". i.e. looking for things I could easily observe/operationalize, that were somehow related to what I cared about.

to:

"Cruxy-first". i.e. Look for things that would change my decisionmaking, even if vague, and learn to either better operationalize those vague things, or, find a way to get better data. (and then, there's a cluster of skills and shortcuts to make that easier)

Disclaimer: 

This post is on the awkward edge of "feels full of promise", but "I haven't yet executed on the stuff that'd make it clearly demonstrably valuable." (And, I've tracked the results of enough "full of promise-feeling" stuff to know they have a <50% hit rate)

I feel like I can see the work needed to make this technique actually functional. It's a lot of work. I'm not sure if it's worth it. (Alas, I'm inventing this technique because I don't know how to tell if my projects are worth it and I really want a principled way of forming beliefs about that. Since I haven't finished vetting it yet it's hard to tell!)

There's a failure mode where rationality schools proliferate without evidence, where people get excited about their new technique that sounds good on paper, and publish exciting blogposts about it. 

There's an unfortunate catch-22 where naive rationality gurus post excitedly. immediately, as if they already know what they're doing. This is even somewhat helpful, because believing in the thing really does help make it more true.

The more humble and realistic rationality gurus wait to publish until they're more confident their thing really works. They feel so in-the-dark, fumbling around. As they should. Because, like, they (we) are. But, I nonetheless feel like the rationality community was overall healthier when it was full of excited people who still believed a bit in their heart that rationality would give them superpowers, and posted excitedly about it. And it feels better for notes to be shared along the way.

Rationality Training doesn't give you superpowers – it's a lot of work, and then "it's a pretty useful skill, among other skills."

I expect, a year from now, I'll have a better conception of the ideas in this post. This post is still my best guess about how this skill can work, and I'd like it if other people excitedly tried to make it work, and report how it went. 

Predicting Outcomes vs Comparing Plans

I'm using forecasting as one tool in a larger toolset. I'll have another post delving deeper into this, but basically: whenever I'm spending a bunch of resources on something, I try to:

    Explicitly ask "what are my goals?"Come up with at least two ideas for approaching that goal, that you really believe in.Identify possible cruxes for pursuing one plan, or the other.Reduce uncertainty on those cruxes.Pick a thing and do it.

The "fluent, cruxy predictions" skill is a tool for step 3 and 4. It's related to Murphijitsu (generally asking "what do I expect to go wrong here?", and then improving your plan, and then asking again until you feel like you plan is good enough)

Often, if I have an important uncertainty, instead of running an expensive experiment to resolve it, I can get a surprising amount of mileage out of just asking "well, what do I actually expect to happen here, if I force myself to get concrete and relevant?". The act of asking the question prompts my unconscious mind (i.e System 1 or Inner Simulator) to reveal what it actually believes when I'm not bullshitting yourself. 

The trick here is using your deliberate, System 2 reasoning, to notice important questions that your System 1 would be good at answering.

But, your unconscious mind isn't magic. It doesn't know everything. So I think it's important to follow this up with making explicit predictions, that you check later, so your inner-simulator can get more accurate over time.

Subskills that go into this are:

Of those: training the skill of "operationalizing cruxy questions" is most important. 

But, reducing friction is the easiest part to get traction on, so let's start there.

Frictionless

It should be as easy as possible to make a prediction. I particularly like the app Fatebook.io. It's very lightweight – I can open up the site, and my cursor will automatically be highlighted in the "new prediction" text box. I type in a prediction, tab over to the "forecast" input to enter a probability, and then hit return and boom, I'm done.

But I like it even more because: 

For example, right now I'm writing a LessWrong doc. Are there any cruxy predictions I can make about this? What actually am I trying to accomplish with this post? After thinking a little, I hit Cmd-Shift-F, entered these predictions and then hit Cmd-V to paste them into the doc:

⚖ A week after posting this, between 1-3 people will have commented saying they got a Fatebook account because of this post (65%)

⚖ A week after posting this, 1-2 people will have actually made a chrome/firefox extension prediction that they paste into a comment. (50%)

Awhile later, I added:

⚖ A post (not by me) that gets 100+ karma will link approvingly to this post within 2 years (10%)

This is actually a decent example of how to use predictions, so I'll walk through what just went through my head. (This will be a little messy, because it's a real thought process. I think it's good to showcase what the real messiness might look like)

I started by asking myself "what actually am I even trying to do with this post?"

The most obvious goal with this post is to get people into the habit of making fluent, cruxy predictions. Ideally, I care that people are still doing it years later, skillfully. But I anticipate getting a very weak signal about that – I won't necessarily know who read the post. If I made Fatebook prediction like "a year from now, someone who read this post will have gone to develop a solid Fatebook habit", I wouldn't even know who to followup with to ask if it had worked.

As I thought about that, I started to feel despair – is this post really going to accomplish anything? I'm actually pretty pessimistic that many people will read this post and take action. And the longterm second-order effects feel pretty murky. This post is a fair amount of effort to write. Why am I even doing this? 

After thinking a bit, two things occurred to me:

And that hopefully illustrates a major point of this whole "fluent cruxy prediction" concept: Asking "what effects do I expect to see from my actions?" often throws into relief that I don't really expect to see results from my actions. I'm just kinda running on autopilot. But, having noticed that, I can followup and ask "okay, what would actually need to be true, to be in the higher-likelihood of success world?"

With that in mind: 

Hey there, reader. 

If this feels at least somewhat compelling, what if you just got yourself to Fatebook right now, and make a couple predictions that'll resolve within couple days, or a week? Fatebook will send you emails reminding you about it, which can help bootstrap a habit.

Feeling around for "cruxiness"

The previous section started getting into this already. But now let's focus more explicitly: What sort of predictions, if true, would change your decisionmaking?

Cruxy operationalization is a murky artform. But here are some principles, and tricks I currently use. Most of my technique involves asking myself various questions, and then seeing what bubbles up.

Over time, I keep track of which question-prompts feel most useful.

Ideal Prerequisite: 
Have at least two (plans/approaches/options)

I'll hopefully write up a whole blogpost about this someday. But one failure mode is "Well, I only really had one general idea. I made predictions about whether that idea would go well. Maybe they returned 'it'll go amazingly well!' and maybe they returned 'it'll go at least pretty well'. 'Pretty well' is at least 'pretty good.' I can't think of anything else to do, so... I guess I'll go with my original idea?"

Having at least two options, that are somewhat mutually exclusive, forces me to start thinking "okay, but how does my favorite plan compare to my other realistic options?". 

It's important to find at least two options you really believe in. (Ideally: have at least one alternative that fits reasonably well into your current lifestyle, or strategic framework, and one that is more radically different).

It's a form of leaving a Line of Retreat, that helps you be more cognitively agile as you decide whether to pivot. If you're having trouble coming up with alternative plans, try Metastrategic Brainstorming.

Frames 1: Costs and benefits

So, one set of questions are "high level cost/benefit comparison?", i.e:

And then:

For me, these activate a different mode, that cuts more to what I actually care about.

Frame 2: Will this really help my deeper, longterm goals?

Often, my plans are really "step 1 of a much larger plan", or, a cluster of strategies that other people in the world are working towards. 

Frame 3: Murphijitsu, and being dissatisfied with "maybe" 

Often while thinking about the first two sets of questions, I find my intuitive sense of "will my project be successful?" is "... kinda?". Either the outcome I'm predicting feels uncertain ("~55%?"), or it's hard to get that clear a visualizable outcome at all.

I then ask "okay, so, if this was definitely, clearly a resounding success, a year from now, what are all the things I would see?"

The way this works best at first sometimes feels like unrealistic dreaming, but then often highlights that the dream is maybe achievable (but perhaps a lot more work than I'd initially envisioned). I generate specific new intermediate actions I might want to plan for. Then, I go back to the original prediction and see if I expect greater success.

This is pretty similar to the process of Murphijitsu (asking "am I surprised if this goes wrong?", and if you wouldn't be too surprised, iterating on your plan until you'd feel actively surprised if it didn't work). The two novel components I'm adding here are:

The feeling of "oh... that's cruxy" 

It's possible to do "fake work" – where you come up with some defensible sounding predictions, but they still don't actually cleave at your cruxes or help you think. 

One thing I particularly look for is a feeling of "oh man, that's cruxy." For me it often comes with a feeling of destabilization/vertigo.

I remember once a few years ago when I was arguing with someone about whether empathy was useful. I was like "Empathy isn't just nice. It makes people more effective so they'll be better at their job." My conversation partner (habryka) said "do you really think Elon Musk would do better at his job if he was more empathetic", and I felt a sinking feeling of "oh no, my beliefs just became falsifiable, and, I'm not sure how had I'd actually bet on them." 

It's hard to actually resolve a bet on whether empathy would help Elon Musk, but, operationalizing the bet, even if it's a hypothetical fantasy bet, helps boot up my realistic intuitions, rather than abstract hopes about "surely this 'make people more empathetic plan will be good' because empathy is generically good."

Example

The Fractal Strategy Workshop

While I was leading up to my Fractal Strategy Workshop, 10 days beforehand, my colleague and I sat down to make some predictions about the workshop.

A naive thing we might have written down is "after the workshop, someone will use a technique we taught them." But this wouldn't be very concrete. How would you know when to check? What counts as using a technique from the workshop?

The things my colleague wrote down tended to be questions like: 

Meanwhile I was making predictions more like:

In some sense, my colleague's questions are quite reasonable. If we aren't seeing participants use the techniques from the workshop, or generally improve their planmaking, the workshop can't be that useful.

But, that wasn't that cruxy for my decisionmaking. I know that the first workshop I run will probably have lots of problems. I also don't really expect a workshop to work in isolation (what I hoped would work was workshop + 6 months of followup coaching to help cement the skills and collaboratively iterate on applying them to the participant's lives). I was deliberately not committing to doing the followup coaching for this workshop, because it was an early beta test. The purpose was to get a sense of whether the curriculum was roughly pointed in the right direction, and decide whether to do a full workshop+coaching later on.

It is cruxy to me whether, if I worked on it for a year, the metastrategy framework would produce clear results. But for the immediate future, the more relevant question was "will the major stakeholders of this project think it is worth paying for?". If people aren't willing to pay significant money for the workshops, that's evidence that they don't expect them to be seriously valuable. And if my boss didn't think it made sense, I'd need to raise money from somewhere and strike out on my own.

This shaped my decisionmaking in a few ways. I realized:

(It's worth noting: my colleague had the opposite impression. They thought it was relatively easy to get funding, they were much more worried about the metastrategy practice failing to impact people. Which is reasonable! But, part of the point here is that what matters is what is cruxy for you. They're your decisions and plans!)

Tips and Tricks

Practice via Videogames

Real life takes a long time to give you feedback on your predictions. I've found videogames a good testbed to practice making predictions, in a way that feels fun, rapid-feedback, while connected to a "real-ish decisionmaking process." 

I recommend using Turn-Based games that you haven't played before, since those give you lots of opportunities to pause and think. 

At first, simply focus on making as many predictions as you can about the game's mechanics, to drill the basic move of "notice that there are multiple plausible worlds that you're living in, and that you have some useful intuitions about it even if you're missing lots of information."

Once you get a basic handle on it, try to specifically ask "what are some unknowns in this game that would affect my strategy?". When you're about to make a choice, consider "what could turn out to be true, that'd lead me to wish I'd made a different choice?".

NOTES for Prediction-Followup

Often, I'll notice "hmm, I could make a prediction about that" while I'm in the middle of a longer thought process. Something I find helpful is to jot down a note about it somewhere I can easily come back to after the train-of-thought completes. (If I'm in a brainstorming document, literally write something like PREDICTION, in capital letters).

The most common note I use is the word "PROMISING," which specifically refers to me having an idea that feels really compelling. (I developed this while practicing solving Baba is You levels, and making predictions about whether my current problem-solving process was on the right track).

I encourage you to develop your own shorthand, based on the mental qualia that are useful for you.

Stuck on making a perfect prediction? Make an imperfect one

(Relatedly: it's okay to make fuzzy predictions that only make sense to you, but when doing so, try to make "extremized" ones)

Sometimes I want to decision – quitting a job, or deciding to do a complex plan. If I can't pin down a concrete observable thing, I can always default to "Subjectively, a year from now, I'll think it was pretty clearly correct to have [done X]."

But, the words "pretty obviously" there are important. 

Oftentimes when I ask "was it a good decision [to quite that job / to start that new relationship / etc]?" the answer is "...maybe? Yeah, kinda?". And that's too vague to be useful as a resolution criteria. 

But if you set the standard for "it has to be pretty clearly correct", not just "probably correct", I find that leaves less room for hemming/hawing about it. 

Make multiple predictions

Relatedly: If you're not sure if one prediction captures The True Spirit of What You're Uncertain about, try making a few different predictions on the same topic.

If I'm considering leaving my job, I might separately try asking things like:

Takeaways

We just covered a lot of stuff. A quick recap of some of the key concepts:

The advice I'd give depends somewhat on where you're starting from. I think it's necessary to first get a basic fluency with "making predictions, at all." 

I recommend getting started via making a bunch of predictions while playing videogames, or other activities with uncertainty + fast feedback loops. I also recommend just making a ton of predictions about your life, to get yourself the basic fluency of completing a prediction-loop, and starting to get contact with "what sort of stuff is in my life that's predictable?"

But, the thing to ultimately be aiming for is to hit a kind of "escape velocity", with a combination of skills make predictions useful for your day job, or major personal projects. For that, the skill of identifying and operationalizing "strategic cruxiness" is an important building block.

If you do end up installing the Fatebook browser extension and making some predictions, let me know! It is helpful to know when people actually 



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

预测 决策 关键因素 内在模拟器 流畅性
相关文章