Edit 11 Feb 2022: Jeremy made a post about starting a low-commitment LW Article club where heād be linkposting articles on a weekly basis from this list for people to engage with!
Context /ā Motivation:
I am interested in thinking a lot recently about how we could share ideas from the rationalist community to EA and related subcommunities (perhaps communicating the same ideas in different ways).
Iāve been diving into LessWrong recently and remembered why I hadnāt for a whileāitās really overwhelming. Even with the sequences and curation, itās a lot of content, and itās not always obvious to me which posts Iād find most valuable.
I think itās better to read fewer posts in more depth to properly understand them.
I think itās likely that some posts or ideas will be much more relevant to EAs than others, but Iām not sure which ones
My ask:
Iād be interested in recommendations for standlone posts (e.g. All debates are bravery debates), specific concepts (e.g. schelling fence or doublecrux) , or specific sequences (e.g.)
If you have time, Iād love to know why itās valuable to you
Iām a big fan of some of the early LessWrong content, e.g.
Confidence levels inside and outside an argument
Humans are not automatically strategic
Parapsychology: the control group for science
Politics as charity
Why the tails come apart
More generally, Iād recommend much of the content by Scott Alexander (āYvainā), Paul Christiano, Wei Dai, Gwern, Greg Lewis (āThrasymachusā), Anna Salamon and Carl Shulman (Iām probably forgetting other names).
I think a lot of old Scott Alexander posts are quite accessible. Some top ones are:
https://āāwww.lesswrong.com/āāposts/āāgFMH3Cqw4XxwL69iy/āāeight-short-studies-on-excuses
https://āāwww.lesswrong.com/āāposts/āā895quRDaK6gR2rM82/āādiseased-thinking-dissolving-questions-about-disease
Some other good classical posts:
https://āāwww.lesswrong.com/āāposts/āāaHaqgTNnFzD7NGLMx/āāreason-as-memetic-immune-disorder
https://āāwww.lesswrong.com/āāposts/āāPBRWb2Em5SNeWYwwB/āāhumans-are-not-automatically-strategic
https://āāwww.lesswrong.com/āāposts/āāyA4gF5KrboK2m2Xu7/āāhow-an-algorithm-feels-from-inside
Privileging the Question changed my life in college. I donāt know how useful it would be for the average person already involved in EA, but it played a huge role in my not getting distracted by random issues and controversies, and instead focusing on big-picture problems that werenāt as inherently interesting. Iād at least recommend it to new members of university EA groups, if not āmost community membersā.
This got me to leave my girlfriend and has remained a permanent way that I think:
https://āāwww.lesswrong.com/āāposts/āā627DZcvme7nLDrbZu/āāupdate-yourself-incrementally
I read it as part of all the sequences, no idea how helpful it will be to others or as a standalone post
My take is that LessWrong is best understood as a mix of individual voices, each with their own style and concerns. The approach Iād recommend is to select one writer whose voice you find compelling, and spend some time digesting their ideas. A common refrain is āread the sequences,ā but thatās not where I started. I like John Wentworthās writing.
Alternatively, you might find yourself interested in a particular topic. LessWrongās tags can help you both find an interesting topic and locate relevant posts, though itās not super fine grained or comprehensive.
One of the key sources of value on LessWrong is that it provides a common language for some complex ideas, presented in a relatively fun and accessible format. The combination of all those ideas can elevate thinking, although itās no panacaea. My intuition is that itās best to slowly follow your curiosity over a period of a few years, rather than trying to digest the whole thing all at once, or pick a couple highlights.
Any particular Wentworth posts that stand out to you? Iād like to include some in the LCLWBC (full credit to you for the name!), but I am not too familiar.
John had several posts highly ranked in the 2020 LessWrong review, and one in the 2019 LessWrong review, so thereās a community consensus that theyāre good. There was also a 2018 LessWrong review, though John didnāt place there.
In general, the review is a great resource for navigating more recent LW content. Although old posts are a community touchstone, the review includes posts that reflect the live interests of community members that have also been extensively vetted not only for being exciting, but for maintaining their value a year later.
Thank you!
Here are a few from Eliezer:
Making Beliefs Pay Rent (in Anticipated Experiences)
Politics is the Mind-Killer
Money: The Unit of Caring
Taboo Your Words
Purchase Fuzzies and Utilons Separately
Leave a Line of Retreat
Dissolving the Question
The Bottom Line
Is That Your True Rejection
Avoiding Your Beliefās Real Weak Points
And a few from others:
Reason as a Memetic Immune Disorder
Is the potential astronomical waste in our universe too small to care about?
Beyond astronomical waste
I might add more later.
I really like Ends Donāt Justify Means (Among Humans) and think itās a bit underrated. (In that I donāt hear people reference it much.)
I think I find the lesson generally useful: that in some cases it can be bad for me to āfollow consequentialism,ā (because in some cases Iām an idiot) without consequentialism being itself bad.
The noncentral fallacy nicely categorizes a very common source of ethical disagreement in my experience.
[Edit:] Somewhat more niche, but considering how important AI risk is to many EAs, Iād also recommend Against GDP as a metric for timelines and takeoff speeds, for rebutting what is in my estimation a bizarrely common error in forecasting AI takeoff.