I reference Buck Shlegeris several times in this post, because appreciating his dyed hair seemed like a fun example and analogy to start the post off with, but any impressive-seeming EA could act as a stand in.
I obtained Buck’s permission to use his name in this post.
Wanting to dye my hair like Buck
When I think of the word cool, my mind conjures refreshing breezes, stars in the sky, friends I whole-heartedly admire, and Buck Shlegeris’ dyed hair.
Why Buck’s hair, you ask? Well it is badass, and the reasoning process behind it is probably also great fun, but my guess is that I mostly think of Buck’s dyed hair as cool because I strongly associate Buck’s hair… with Buck (it is a distinctive feature of his).
People tell me that Buck does a lot of pretty good work as the CTO of Redwood Research, which is one of the major EA aligned organizations working on alignment, which may very well be the present most important bottleneck on humanity’s future going well.
So by extension, Buck’s hair is cool.
I don’t think this is logically sound. What I do know is that I generally don’t attach an emotional valence to dyed hair, but when I see Buck at retreats or conferences, I notice his hair and I’m slightly more likely to think, “huh, Buck has cool hair” even though I generally don’t care what color your hair is.
It’s a stupid update; my appreciation for Buck’s work shouldn’t extend to his hair, but I definitely do it a bit anyways. It’s like there’s a node in my mind for Buck that is positively valenced and a node for dyed hair that is neutrally valenced. When I see his dyed hair, the nodes become associated with each other and a previously neutral dyed hair node becomes tinted with positive feelings.
My guess is that I’m more inclined to dye my hair myself because Buck dyes his hair. This is true even though we’re optimizing for different aesthetic preferences and I don’t actually obtain almost any information on how I should optimize for my aesthetic preferences based on Buck’s decisions.
Conclusion: When I see people I admire commit to action X, I want to commit to X myself a little more, regardless of the informational value of their decision to me.
Buck does alignment research. Should I do alignment research?
Now let’s broaden this example. Lots of people I admire believe alignment research is very important. Even discounting the specific reasons they believe alignment research is important, I should probably be more inclined to do alignment research, because these people are super smart and thoughtful and largely value aligned with me in our altruistic efforts.
But how much more?
So I first take into account their epistemic status, adjusting my opinion on the importance of alignment research based on how much I trust their judgment and how confident they are and how altruistically aligned we are.
But remember: I want to dye my hair slightly more because Buck dyes his hair, even though his dyeing his hair doesn’t provide (basically) any information to me on how I should optimize for my own aesthetic preferences.
There’s a separate motivation of “wanting to do what this person I like does” that I have to separate out and subtract from how much I want to do alignment research if I want to have an accurate sense of the importance of alignment research.
But this is hard. Disentangling the informational value people I admire have provided through their beliefs that alignment research is important from the positive feelings I hold towards alignment research because it is now associated with them is hard.
And I’m probably not doing it very well.
Conclusion: I’m likely updating too much in favor of dyed hair and alignment research based on the beliefs of and my positive feelings for EA’s I admire. Maybe I really should do the thing where I sit down for a few dozen hours and formulate my empirical and moral uncertainties into precise questions whose answers are cruxy to me working on alignment and then try to answer those questions.
Optimizing for validation instead of impact
But it gets worse. Dyeing my hair might not better fulfill my aesthetic preferences, but it may very well make Buck like me marginally more. Doing alignment research might not be the right path to impact for me, but it may very well make lots of people I admire like me marginally more. My monke brain is way more attracted instinctively to “optimizing for self-validation within the EA community by doing the things impressive EAs say are good” rather than “actually figuring out how to save the world and doing it,” and the two look similar enough, and the former so much easier, that I’m probably falling into this trap.
Conclusion: It is likely that at least some of the time, I’m unintentionally optimizing for validation within the EA community, rather than fulfilling the EA project. Sometimes optimizing for either leads to the same action, but whenever they pull apart and I choose to follow the EA community incentives rather than the EA project ones, I pay in dead children. This makes me somewhat aggrieved. So what do I do?
I don’t know. I’m writing this post because I’m not sure; I have some thoughts, but they’re very scattered. Do you have suggestions?
Note:
This is my first forum post, so feedback is really appreciated! In particular, I realize that I’m singling out Buck in this post and I’m unsure if this was useful in communicating my point, or if this was just weird.
I think there’s definitely some part of me that is trying to feel more legible in the community through this post. I’ve deleted the karma elements from the EA forum for myself using my adblock and am writing this under a pseudonym to try to mitigate this.
An early reader noted that “it’s badass to dye your hair like [Buck’s], it’s contrarian, it’s free. and come on, deciding your hair color based on quantum events such that theoretically all bucks throughout the multiverse sport all the colors of the rainbow is just fucking cool”. So those are also reasons to dye your hair like Buck.
Edit: changed the title because jwpieters noted it was misleading. It was originally the first section title “wanting to dye my hair like buck”. Thanks jwpieters!
Wanting to dye my hair a little more because Buck dyes his hair
I reference Buck Shlegeris several times in this post, because appreciating his dyed hair seemed like a fun example and analogy to start the post off with, but any impressive-seeming EA could act as a stand in.
I obtained Buck’s permission to use his name in this post.
Wanting to dye my hair like Buck
When I think of the word cool, my mind conjures refreshing breezes, stars in the sky, friends I whole-heartedly admire, and Buck Shlegeris’ dyed hair.
Why Buck’s hair, you ask? Well it is badass, and the reasoning process behind it is probably also great fun, but my guess is that I mostly think of Buck’s dyed hair as cool because I strongly associate Buck’s hair… with Buck (it is a distinctive feature of his).
People tell me that Buck does a lot of pretty good work as the CTO of Redwood Research, which is one of the major EA aligned organizations working on alignment, which may very well be the present most important bottleneck on humanity’s future going well.
So by extension, Buck’s hair is cool.
I don’t think this is logically sound. What I do know is that I generally don’t attach an emotional valence to dyed hair, but when I see Buck at retreats or conferences, I notice his hair and I’m slightly more likely to think, “huh, Buck has cool hair” even though I generally don’t care what color your hair is.
It’s a stupid update; my appreciation for Buck’s work shouldn’t extend to his hair, but I definitely do it a bit anyways. It’s like there’s a node in my mind for Buck that is positively valenced and a node for dyed hair that is neutrally valenced. When I see his dyed hair, the nodes become associated with each other and a previously neutral dyed hair node becomes tinted with positive feelings.
My guess is that I’m more inclined to dye my hair myself because Buck dyes his hair. This is true even though we’re optimizing for different aesthetic preferences and I don’t actually obtain almost any information on how I should optimize for my aesthetic preferences based on Buck’s decisions.
Conclusion: When I see people I admire commit to action X, I want to commit to X myself a little more, regardless of the informational value of their decision to me.
Buck does alignment research. Should I do alignment research?
Now let’s broaden this example. Lots of people I admire believe alignment research is very important. Even discounting the specific reasons they believe alignment research is important, I should probably be more inclined to do alignment research, because these people are super smart and thoughtful and largely value aligned with me in our altruistic efforts.
But how much more?
So I first take into account their epistemic status, adjusting my opinion on the importance of alignment research based on how much I trust their judgment and how confident they are and how altruistically aligned we are.
But remember: I want to dye my hair slightly more because Buck dyes his hair, even though his dyeing his hair doesn’t provide (basically) any information to me on how I should optimize for my own aesthetic preferences.
There’s a separate motivation of “wanting to do what this person I like does” that I have to separate out and subtract from how much I want to do alignment research if I want to have an accurate sense of the importance of alignment research.
But this is hard. Disentangling the informational value people I admire have provided through their beliefs that alignment research is important from the positive feelings I hold towards alignment research because it is now associated with them is hard.
And I’m probably not doing it very well.
Conclusion: I’m likely updating too much in favor of dyed hair and alignment research based on the beliefs of and my positive feelings for EA’s I admire. Maybe I really should do the thing where I sit down for a few dozen hours and formulate my empirical and moral uncertainties into precise questions whose answers are cruxy to me working on alignment and then try to answer those questions.
Optimizing for validation instead of impact
But it gets worse. Dyeing my hair might not better fulfill my aesthetic preferences, but it may very well make Buck like me marginally more. Doing alignment research might not be the right path to impact for me, but it may very well make lots of people I admire like me marginally more. My monke brain is way more attracted instinctively to “optimizing for self-validation within the EA community by doing the things impressive EAs say are good” rather than “actually figuring out how to save the world and doing it,” and the two look similar enough, and the former so much easier, that I’m probably falling into this trap.
Conclusion: It is likely that at least some of the time, I’m unintentionally optimizing for validation within the EA community, rather than fulfilling the EA project. Sometimes optimizing for either leads to the same action, but whenever they pull apart and I choose to follow the EA community incentives rather than the EA project ones, I pay in dead children. This makes me somewhat aggrieved. So what do I do?
I don’t know. I’m writing this post because I’m not sure; I have some thoughts, but they’re very scattered. Do you have suggestions?
Note:
This is my first forum post, so feedback is really appreciated! In particular, I realize that I’m singling out Buck in this post and I’m unsure if this was useful in communicating my point, or if this was just weird.
I think there’s definitely some part of me that is trying to feel more legible in the community through this post. I’ve deleted the karma elements from the EA forum for myself using my adblock and am writing this under a pseudonym to try to mitigate this.
An early reader noted that “it’s badass to dye your hair like [Buck’s], it’s contrarian, it’s free. and come on, deciding your hair color based on quantum events such that theoretically all bucks throughout the multiverse sport all the colors of the rainbow is just fucking cool”. So those are also reasons to dye your hair like Buck.
Edit: changed the title because jwpieters noted it was misleading. It was originally the first section title “wanting to dye my hair like buck”. Thanks jwpieters!