I agree that we shouldn’t use e2g as a shorthand for skillmaxing.
I am less optimistic about the ‘fit’ vs raw competence point. It’s not clear to me that a good fit for the work position can easily be gleaned by work tests—a very competent person may be able to acquire that ‘fit’ within a few weeks on the job, for example, once they have more context for the kind of work the organization wants. So even if the candidates at the point of hiring looked very different, their comparison may differ unless we imagine both in an applied job context, having learned things they did not know at the time of hiring.
I am more broadly worried about ‘fit’ in EA hiring contexts, because as opposed to markers of raw competence, ‘fit’ provides a lot of flexibility for selecting traits that are relatively tangential to work performance and often unreliable. For example, value-fit might select for hiring likeminded folks who have read the same stuff the hiring manager has, and reduce epistemic diversity. A fit for similar research interests reduces epistemic diversity and locks in certain research agendas for a long time. A vibe-fit may select simply for friends and those who have internalized norms. A worktest that is on an explicitly EA project may select for those already more familiar with EA, even if it would be easy for an outsider candidate to pick up on basic EA knowledge quickly if they got the job.
My impression is that overall, EA does have a noticeable suboptimal tendency to hire likeminded folks and folks in overlapping social circles (i.e. friends; friends of friends). Insofar as ‘fit’ makes it easier to justify this tendency internally and externally, I worry that it will lead to suboptimal hiring. I acknowledge we may have very different kinds of ‘fit’ in mind here. I do think the examples I provide above do exist in EA hiring decisions.
I haven’t done hiring rounds for EA, so I may be completely wrong—maybe your experience has been that after a few worktests it becomes abundantly clear who the right candidate is.
I agree that we shouldn’t use e2g as a shorthand for skillmaxing.
I am less optimistic about the ‘fit’ vs raw competence point. It’s not clear to me that a good fit for the work position can easily be gleaned by work tests—a very competent person may be able to acquire that ‘fit’ within a few weeks on the job, for example, once they have more context for the kind of work the organization wants. So even if the candidates at the point of hiring looked very different, their comparison may differ unless we imagine both in an applied job context, having learned things they did not know at the time of hiring.
I am more broadly worried about ‘fit’ in EA hiring contexts, because as opposed to markers of raw competence, ‘fit’ provides a lot of flexibility for selecting traits that are relatively tangential to work performance and often unreliable. For example, value-fit might select for hiring likeminded folks who have read the same stuff the hiring manager has, and reduce epistemic diversity. A fit for similar research interests reduces epistemic diversity and locks in certain research agendas for a long time. A vibe-fit may select simply for friends and those who have internalized norms. A worktest that is on an explicitly EA project may select for those already more familiar with EA, even if it would be easy for an outsider candidate to pick up on basic EA knowledge quickly if they got the job.
My impression is that overall, EA does have a noticeable suboptimal tendency to hire likeminded folks and folks in overlapping social circles (i.e. friends; friends of friends). Insofar as ‘fit’ makes it easier to justify this tendency internally and externally, I worry that it will lead to suboptimal hiring. I acknowledge we may have very different kinds of ‘fit’ in mind here. I do think the examples I provide above do exist in EA hiring decisions.
I haven’t done hiring rounds for EA, so I may be completely wrong—maybe your experience has been that after a few worktests it becomes abundantly clear who the right candidate is.