This is part of LessWrong for EA, a LessWrong repost & low-commitment discussion group (inspired by this comment). Each week I will revive a highly upvoted, EA-relevant post from the LessWrong Archives, more or less at random
Excerpt from the post:
How often do you make predictions (either about future events, or about information that you don’t yet have)? If you’re a regular Less Wrong reader you’re probably familiar with the idea that you should make your beliefs pay rent by saying, “Here’s what I expect to see if my belief is correct, and here’s how confident I am,” and that you should then update your beliefs accordingly, depending on how your predictions turn out.
And yet… my impression is that few of us actually make predictions on a regular basis. Certainly, for me, there has always been a gap between how useful I think predictions are, in theory, and how often I make them.
I don’t think this is just laziness. I think it’s simply not a trivial task to find predictions to make that will help you improve your models of a domain you care about.
At this point I should clarify that there are two main goals predictions can help with:
Improved Calibration (e.g., realizing that I’m only correct about Domain X 70% of the time, not 90% of the time as I had mistakenly thought).
Improved Accuracy (e.g., going from being correct in Domain X 70% of the time to being correct 90% of the time)
If your goal is just to become better calibrated in general, it doesn’t much matter what kinds of predictions you make. So calibration exercises typically grab questions with easily obtainable answers, like “How tall is Mount Everest?” or “Will Don Draper die before the end of Mad Men?” See, for example, the Credence Game, Prediction Book, and this recent post. And calibration training really does work.
But even though making predictions about trivia will improve my general calibration skill, it won’t help me improve my models of the world. That is, it won’t help me become more accurate, at least not in any domains I care about. If I answer a lot of questions about the heights of mountains, I might become more accurate about that topic, but that’s not very helpful to me.
So I think the difficulty in prediction-making is this: The set {questions whose answers you can easily look up, or otherwise obtain} is a small subset of all possible questions. And the set {questions whose answers I care about} is also a small subset of all possible questions. And the intersection between those two subsets is much smaller still, and not easily identifiable. As a result, prediction-making tends to seem too effortful, or not fruitful enough to justify the effort it requires. (Full post on LW)
LW4EA: 16 types of useful predictions
Link post
Written by LW user Julia_Galef.
This is part of LessWrong for EA, a LessWrong repost & low-commitment discussion group (inspired by this comment). Each week I will revive a highly upvoted, EA-relevant post from the LessWrong Archives, more or less at random
Excerpt from the post:
Please feel free to,
Discuss in the comments
Subscribe to the LessWrong for EA tag to be notified of future posts
Tag other LessWrong reposts with LessWrong for EA.
Recommend additional posts