AI alignment
TagLast edit: 22 Jul 2022 20:58 UTC by Leo AI alignment is research on how to align AI systems with human or moral goals.
Evaluation
80,000 Hours rates AI alignment a “highest priority area”: a problem at the top of their ranking of global issues assessed by importance, tractability and neglectedness.
Further reading
Christiano, Paul (2020) Current work in AI alignment, Effective Altruism Forum, April 3.
Shah, Rohin (2020) What’s been happening in AI alignment?, Effective Altruism Forum, July 29.
External links
AI Alignment Forum.
Related entries
AI governance | AI forecasting | alignment tax | Center for Human-Compatible Artificial Intelligence | Machine Intelligence Research Institute | rationality community
Ajeya21 Sep 2021 15:35 UTC153 points
14 min readEA link(www.cold-takes.com)
Jacy20 Feb 2018 18:29 UTC107 points
35 min readEA link(www.sentienceinstitute.org)
EJT20 Feb 2023 21:52 UTC107 points
19 min readEA link evhub12 Jan 2024 19:51 UTC65 points
1 min readEA link(arxiv.org)
KR7 Jul 2020 17:53 UTC18 points
8 min readEA link Ben23 Jul 2020 19:02 UTC10 points
3 min readEA link Chi15 Aug 2020 19:59 UTC38 points
39 min readEA link KR22 Jul 2020 19:07 UTC21 points
3 min readEA link joe9 Feb 2023 11:55 UTC27 points
18 min readEA link DavidW15 Feb 2023 20:12 UTC20 points
1 min readEA link(www.lesswrong.com)
evhub1 Jul 2020 20:59 UTC13 points
1 min readEA link(futureoflife.org)
Ajeya27 Mar 2023 19:46 UTC198 points
1 min readEA link(www.planned-obsolescence.org)
mic30 Jun 2022 18:37 UTC53 points
11 min readEA link leopold29 Mar 2023 14:26 UTC327 points
9 min readEA link(www.forourposterity.com)
Akash11 Sep 2022 23:45 UTC31 points
6 min readEA link(www.lesswrong.com)
Pablo21 Nov 2020 12:02 UTC36 points
1 min readEA link(www.lesswrong.com)
ts20 Jan 2022 21:10 UTC11 points
1 min readEA link Chi3 Mar 2024 18:07 UTC113 points
21 min readEA link Gavin23 Mar 2021 13:18 UTC31 points
8 min readEA link(www.gleech.org)
Gavin8 Apr 2022 22:08 UTC22 points
1 min readEA link(marginalrevolution.com)
Ajeya15 Dec 2020 12:10 UTC35 points
1 min readEA link(alignmentforum.org)
Hzn16 Dec 2024 14:01 UTC−2 points
3 min readEA link plex18 May 2024 15:06 UTC13 points
1 min readEA link(aisafety.info)
Garrison23 Oct 2024 23:42 UTC57 points
7 min readEA link(garrisonlovely.substack.com)
yixiong14 Nov 2024 13:34 UTC18 points
5 min readEA link(yixiong.substack.com)
Jeremy21 Nov 2022 17:47 UTC15 points
1 min readEA link(scottaaronson.blog)
Mark Xu24 Dec 2020 23:08 UTC23 points
2 min readEA link(www.alignmentforum.org)
Raemon19 Nov 2018 2:21 UTC26 points
1 min readEA link(www.lesswrong.com)
Habryka28 Jan 2020 21:14 UTC28 points
6 min readEA link(www.lesswrong.com)
Jeremy17 May 2022 3:05 UTC11 points
1 min readEA link(www.lesswrong.com)
ac10 Feb 2020 10:10 UTC26 points
16 min readEA link Fai25 Feb 2022 20:43 UTC46 points
8 min readEA link JakubK13 Dec 2022 19:04 UTC21 points
2 min readEA link(www.lesswrong.com)
rgb25 Apr 2022 13:42 UTC91 points
1 min readEA link adamShimi20 Jul 2022 10:44 UTC43 points
9 min readEA link(www.alignmentforum.org)
mic2 Feb 2024 18:22 UTC27 points
22 min readEA link(papers.ssrn.com)
Miguel3 Feb 2023 19:32 UTC18 points
1 min readEA link(www.whitehatstoic.com)
Miguel4 Feb 2023 16:49 UTC4 points
12 min readEA link(www.whitehatstoic.com)
Matt Keene10 Feb 2023 18:15 UTC−9 points
5 min readEA link(www.creatingafuturewewant.com)
Roko12 Dec 2024 10:46 UTC−7 points
1 min readEA link(www.transhumanaxiology.com)
Otto24 Jul 2023 10:18 UTC36 points
7 min readEA link(time.com)
leopold29 Mar 2023 15:19 UTC56 points
5 min readEA link(www.forourposterity.com)
Dan H30 Mar 2023 13:09 UTC41 points
2 min readEA link(arxiv.org)