Some research questions that you may want to tackle

Formerly titled “Write up my research ideas for someone else to tackle? Fine—you asked for it!”

Unrelatedly, thanks to Jessica McCurdy for telling me to write down some of my research ideas and questions in case someone else wants to tackle one (or a few).

The list

  1. Cause prio but for earning to give

    1. As far as I know, SBF relied on his personal knowledge and intuition when deciding to try building FTX.

    2. It doesn’t have to be this way! I can imagine a more systematic effort to identify and describe which earning to give opportunities are most promising. Is there a $100B idea with a 1% chance of working? A $1T idea with a 0.1% chance? I think we can and should find out.

  2. Are there cheap and easy ways to kill fish quickly?

    1. Right now, I estimate 250 million fish years are spent in agony each year as wild fish are killed by asphyxiation or being gutted alive, which takes a surprisingly long time to cause death. There must be a better way.

      1. Related: can we just raise (farm) a ton of fish ourselves, but using humane practices, with donations subsidizing the cost difference relative to standard aquaculture

  3. From my red teaming project on extinction risk reduction:

    1. Unpacking which particular biorisk prevention activities seem robust to a set of plausible empirical and ethical assumptions and which do not; and

    2. Seeking to identify any AI alignment research programs that would reduce s-risks by a greater magnitude than “mainstream” x-risk-oriented alignment research.

  4. From my “half baked ideas comment” on the Forum:

    1. Figure out how to put to good use some greater proportion of the approximately 1 Billion recent college grads who want to work at an “EA org”

      1. This might look like a collective of independent-ish researchers?

    2. There should be way more all-things-considered, direct comparisons between cause areas.

      1. So I guess the research question is: what is the most important cause area to work on and/​or donate to, all things considered?

        1. No more “agreeing to disagree”—I want an (intellectual) fight to the death. Liberal-spending longtermists should make an affirmative case that this ethos is the best way to spend money on the margin, and objectors should argue that it isn’t.

        2. In particular, I don’t think a complete case has been made (even from a total utilitarian, longtermist perspective) that at the current funding margin, it makes sense to spend marginal dollars on longtermism-motivated projects instead of animal welfare projects. I’d be very interested to see this comparison in particular

  5. [Related to above] Is anyone actually arguing that neartermist, human-centric interventions are the most ethical way to spend time or money?

    1. That’s not a rhetorical question! The hundreds of millions of dollars being directed to AMF et al. instead of some other charity or cause area should be more seriously justified or defended, IMO.

    2. For anyone who does think that improving human welfare in the developing world is the best thing to do: do AMF-type charities actually increase the number of human life-years lived?

  6. (As I asked on Twitter) What jobs/​tasks/​roles are high impact (by normal EA standards) but relatively low status within EA?

    1. I think one of the big ways EA could screw up is by having intra-EA status incongruent (at least ordinally) with expected impact.

  7. What would an animal welfare movement with the ambition, epistemic quality, and enthusiasm (and maybe funding) of the longtermist movement look like?

  8. [I might tackle this] What can AI safety learn from human brains’ bilateral asymmetry

    1. The whole “brain hemisphere difference” thing is surrounded by plenty pop science myths, surrounding it, but there really are some quite profound differences as described in Ian McGilchrist’s The Master and His Emissary

  9. What positions of power and/​or influence in the world are most neglected or easiest to access, perhaps because they’re low prestige and/​or low pay?

  10. S-risk people: what can we actually do, in the real world and the foreseeable future, to decrease s-risks?

    1. It seems to me most of this research is quite abstract and theoretical—which may not make sense if transformative AI is only a few years away!

  11. It seems like the default view is that some time in the future, the world and/​or EA is going to decide that AI systems are sentient. This seems totally implausible.

    1. What should we do under radical uncertainty as to whether any given “thing” or process is sentient?

    2. What empirical observations, if any, should change our actions, plans, or ethics?