RSS

AI alignment

Core TagLast edit: 16 May 2021 12:51 UTC by Pablo

The AI alignment tag is used for posts that discuss aligning AI systems with human interests, and for meta-discussion about whether this goal is worthwhile, achievable, etc.

Related entries

AI forecasting | AI governance | Center for Human-Compatible Artificial Intelligence | Machine Intelligence Research Institute | rationality community

2019 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks19 Dec 2019 2:58 UTC
146 points
28 comments62 min readEA link

2018 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks18 Dec 2018 4:48 UTC
115 points
28 comments63 min readEA link

Ben Garfinkel: How sure are we about this AI stuff?

Ben Garfinkel9 Feb 2019 19:17 UTC
93 points
16 commentsEA link

Disen­tan­gling ar­gu­ments for the im­por­tance of AI safety

richard_ngo23 Jan 2019 14:58 UTC
59 points
14 comments8 min readEA link

Del­e­gated agents in prac­tice: How com­pa­nies might end up sel­l­ing AI ser­vices that act on be­half of con­sumers and coal­i­tions, and what this im­plies for safety research

remmelt26 Nov 2020 16:39 UTC
11 points
0 comments4 min readEA link

My cur­rent thoughts on MIRI’s “highly re­li­able agent de­sign” work

Daniel_Dewey7 Jul 2017 1:17 UTC
50 points
64 commentsEA link

2016 AI Risk Liter­a­ture Re­view and Char­ity Comparison

Larks13 Dec 2016 4:36 UTC
53 points
22 commentsEA link

My per­sonal cruxes for work­ing on AI safety

Buck13 Feb 2020 7:11 UTC
124 points
33 comments44 min readEA link

2017 AI Safety Liter­a­ture Re­view and Char­ity Comparison

Larks20 Dec 2017 21:54 UTC
43 points
17 commentsEA link

How do take­off speeds af­fect the prob­a­bil­ity of bad out­comes from AGI?

KR7 Jul 2020 17:53 UTC
18 points
0 comments8 min readEA link

Scru­ti­niz­ing AI Risk (80K, #81) - v. quick summary

Louis_Dixon23 Jul 2020 19:02 UTC
10 points
0 comments3 min readEA link

Cog­ni­tive Science/​Psy­chol­ogy As a Ne­glected Ap­proach to AI Safety

Kaj_Sotala5 Jun 2017 13:46 UTC
34 points
37 commentsEA link

The aca­demic con­tri­bu­tion to AI safety seems large

technicalities30 Jul 2020 10:30 UTC
105 points
24 comments9 min readEA link

Hiring en­g­ineers and re­searchers to help al­ign GPT-3

Paul_Christiano1 Oct 2020 18:52 UTC
106 points
19 comments3 min readEA link

AGI safety from first principles

richard_ngo21 Oct 2020 17:42 UTC
68 points
0 comments3 min readEA link
(www.alignmentforum.org)

An­nounc­ing AI Safety Support

Linda Linsefors19 Nov 2020 20:19 UTC
53 points
0 comments4 min readEA link

TAI Safety Biblio­graphic Database

Jess_Riedel22 Dec 2020 16:03 UTC
57 points
9 comments17 min readEA link

AMA: Ajeya Co­tra, re­searcher at Open Phil

Ajeya28 Jan 2021 17:38 UTC
83 points
107 comments1 min readEA link

[Link post] Co­or­di­na­tion challenges for pre­vent­ing AI conflict

stefan.torges9 Mar 2021 9:39 UTC
48 points
0 comments1 min readEA link
(longtermrisk.org)

Draft re­port on ex­is­ten­tial risk from power-seek­ing AI

Joe_Carlsmith28 Apr 2021 21:41 UTC
76 points
33 comments1 min readEA link

[Question] How much EA anal­y­sis of AI safety as a cause area ex­ists?

richard_ngo6 Sep 2019 11:15 UTC
83 points
19 comments2 min readEA link

Rele­vant pre-AGI possibilities

kokotajlod20 Jun 2020 13:15 UTC
22 points
0 comments1 min readEA link
(aiimpacts.org)

Par­allels Between AI Safety by De­bate and Ev­i­dence Law

Cullen_OKeefe20 Jul 2020 22:52 UTC
30 points
2 comments2 min readEA link
(cullenokeefe.com)

[Question] How strong is the ev­i­dence of un­al­igned AI sys­tems caus­ing harm?

evelynciara21 Jul 2020 4:08 UTC
31 points
1 comment1 min readEA link

In­tel­lec­tual Diver­sity in AI Safety

KR22 Jul 2020 19:07 UTC
19 points
8 comments3 min readEA link

Why the Orthog­o­nal­ity Th­e­sis’s ve­rac­ity is not the point:

Etoile de Scauchy23 Jul 2020 15:40 UTC
1 point
0 comments3 min readEA link

AI Risk: In­creas­ing Per­sua­sion Power

kewlcats3 Aug 2020 20:25 UTC
4 points
0 comments1 min readEA link

My Un­der­stand­ing of Paul Chris­ti­ano’s Iter­ated Am­plifi­ca­tion AI Safety Re­search Agenda

Chi15 Aug 2020 19:59 UTC
34 points
3 comments39 min readEA link

[Link] How un­der­stand­ing valence could help make fu­ture AIs safer

Milan_Griffes8 Oct 2020 18:53 UTC
22 points
2 comments3 min readEA link

An­nounc­ing AXRP, the AI X-risk Re­search Podcast

DanielFilan23 Dec 2020 20:10 UTC
30 points
1 comment1 min readEA link

Buck Sh­legeris: How I think stu­dents should ori­ent to AI safety

EA Global25 Oct 2020 5:48 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

In­tro­duc­ing The Non­lin­ear Fund: AI Safety re­search, in­cu­ba­tion, and funding

Kat Woods18 Mar 2021 14:07 UTC
62 points
32 comments5 min readEA link

Long-Term Fu­ture Fund: April 2019 grant recommendations

Habryka23 Apr 2019 7:00 UTC
144 points
242 comments46 min readEA link

I’m Buck Sh­legeris, I do re­search and out­reach at MIRI, AMA

Buck15 Nov 2019 22:44 UTC
122 points
231 comments2 min readEA link

Tech­ni­cal AGI safety re­search out­side AI

richard_ngo18 Oct 2019 15:02 UTC
81 points
5 comments3 min readEA link

I’m Cul­len O’Keefe, a Policy Re­searcher at OpenAI, AMA

Cullen_OKeefe11 Jan 2020 4:13 UTC
44 points
68 comments1 min readEA link

Align­ment Newslet­ter One Year Retrospective

rohinmshah10 Apr 2019 7:00 UTC
62 points
22 comments21 min readEA link

[Link] EAF Re­search agenda: “Co­op­er­a­tion, Con­flict, and Trans­for­ma­tive Ar­tifi­cial In­tel­li­gence”

stefan.torges17 Jan 2020 13:28 UTC
63 points
0 comments1 min readEA link

[AN #80]: Why AI risk might be solved with­out ad­di­tional in­ter­ven­tion from longtermists

rohinmshah3 Jan 2020 7:52 UTC
58 points
12 comments10 min readEA link
(www.alignmentforum.org)

AI Im­pacts: His­toric trends in tech­nolog­i­cal progress

Aaron Gertler12 Feb 2020 0:08 UTC
55 points
5 comments3 min readEA link

Ought: why it mat­ters and ways to help

Paul_Christiano26 Jul 2019 1:56 UTC
53 points
5 comments5 min readEA link

Crit­i­cal Re­view of ‘The Precipice’: A Re­assess­ment of the Risks of AI and Pandemics

Fods1211 May 2020 11:11 UTC
80 points
32 comments26 min readEA link

[Question] What are the challenges and prob­lems with pro­gram­ming law-break­ing con­straints into AGI?

MichaelStJules2 Feb 2020 20:53 UTC
6 points
34 comments1 min readEA link

FLI AI Align­ment pod­cast: Evan Hub­inger on In­ner Align­ment, Outer Align­ment, and Pro­pos­als for Build­ing Safe Ad­vanced AI

evhub1 Jul 2020 20:59 UTC
13 points
2 comments1 min readEA link
(futureoflife.org)

AMA or dis­cuss my 80K pod­cast epi­sode: Ben Garfinkel, FHI researcher

Ben Garfinkel13 Jul 2020 16:17 UTC
87 points
140 comments1 min readEA link

A list of good heuris­tics that the case for AI X-risk fails

Aaron Gertler16 Jul 2020 9:56 UTC
23 points
9 comments2 min readEA link
(www.alignmentforum.org)

Per­sonal thoughts on ca­reers in AI policy and strategy

carrickflynn27 Sep 2017 16:52 UTC
49 points
29 commentsEA link

Ro­hin Shah: What’s been hap­pen­ing in AI al­ign­ment?

EA Global29 Jul 2020 20:15 UTC
17 points
0 comments14 min readEA link
(www.youtube.com)

Is GPT-3 the death of the pa­per­clip max­i­mizer?

matthias_samwald3 Aug 2020 11:34 UTC
4 points
1 comment1 min readEA link

Con­ver­sa­tion on AI risk with Adam Gleave

AI Impacts27 Dec 2019 21:43 UTC
18 points
3 comments4 min readEA link
(aiimpacts.org)

Dis­con­tin­u­ous progress in his­tory: an update

AI Impacts17 Apr 2020 16:28 UTC
60 points
3 comments24 min readEA link

Thoughts on short timelines

Tobias_Baumann23 Oct 2018 15:59 UTC
22 points
15 commentsEA link

New re­port on how much com­pu­ta­tional power it takes to match the hu­man brain (Open Philan­thropy)

Aaron Gertler15 Sep 2020 1:06 UTC
33 points
1 comment17 min readEA link
(www.openphilanthropy.org)

Align­ing Recom­mender Sys­tems as Cause Area

IvanVendrov8 May 2019 8:56 UTC
152 points
32 comments13 min readEA link

Con­sider pay­ing me to do AI safety re­search work

Rupert5 Nov 2020 8:09 UTC
6 points
3 comments2 min readEA link

AGI Predictions

Pablo21 Nov 2020 12:02 UTC
36 points
0 comments1 min readEA link
(www.lesswrong.com)

[Question] How can I bet on short timelines?

kokotajlod7 Nov 2020 12:45 UTC
33 points
10 comments2 min readEA link

[Question] Is this a good way to bet on short timelines?

kokotajlod28 Nov 2020 14:31 UTC
17 points
16 comments1 min readEA link

Draft re­port on AI timelines

Ajeya15 Dec 2020 12:10 UTC
28 points
0 comments1 min readEA link
(alignmentforum.org)

Some AI re­search ar­eas and their rele­vance to ex­is­ten­tial safety

critch15 Dec 2020 12:15 UTC
10 points
0 comments56 min readEA link
(alignmentforum.org)

Open Philan­thropy’s AI gov­er­nance grant­mak­ing (so far)

Aaron Gertler17 Dec 2020 12:00 UTC
63 points
0 comments6 min readEA link
(www.openphilanthropy.org)

2020 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks21 Dec 2020 15:25 UTC
134 points
14 comments68 min readEA link

Take­aways from safety by de­fault interviews

AI Impacts7 Apr 2020 2:01 UTC
24 points
2 comments13 min readEA link
(aiimpacts.org)

What does it mean to be­come an ex­pert in AI Hard­ware?

Christopher_Phenicie9 Jan 2021 4:15 UTC
65 points
10 comments11 min readEA link

[Question] How should we in­vest in “long-term short-ter­mism” given the like­li­hood of trans­for­ma­tive AI?

James_Banks12 Jan 2021 23:54 UTC
7 points
0 comments1 min readEA link

Shar­ing the World with Digi­tal Minds

Aaron Gertler1 Dec 2020 8:00 UTC
12 points
0 comments1 min readEA link
(www.nickbostrom.com)

What Should the Aver­age EA Do About AI Align­ment?

Raemon25 Feb 2017 20:07 UTC
30 points
41 commentsEA link

Op­por­tu­ni­ties for in­di­vi­d­ual donors in AI safety

alexflint12 Mar 2018 2:10 UTC
13 points
13 commentsEA link

Quan­tify­ing the Far Fu­ture Effects of Interventions

MichaelDickens18 May 2016 2:15 UTC
8 points
1 commentEA link

Po­ten­tial Risks from Ad­vanced AI

EA Global13 Aug 2017 7:00 UTC
8 points
0 comments17 min readEA link

What does (and doesn’t) AI mean for effec­tive al­tru­ism?

EA Global12 Aug 2017 7:00 UTC
8 points
0 comments12 min readEA link

Some global catas­trophic risk estimates

Tamay10 Feb 2021 19:32 UTC
99 points
12 comments1 min readEA link

Three Im­pacts of Ma­chine Intelligence

Paul_Christiano23 Aug 2013 10:10 UTC
33 points
4 comments8 min readEA link
(rationalaltruist.com)

Some promis­ing ca­reer ideas be­yond 80,000 Hours’ pri­or­ity paths

Ardenlk26 Jun 2020 10:34 UTC
129 points
25 comments15 min readEA link

Jade Le­ung: Who do we want to con­trol hu­man-level AI?

EA Global8 Jun 2018 7:15 UTC
8 points
0 comments22 min readEA link
(www.youtube.com)

Jesse Clif­ton: Open-source learn­ing — a bar­gain­ing approach

EA Global18 Oct 2019 18:05 UTC
9 points
0 comments1 min readEA link
(www.youtube.com)

How to build a safe ad­vanced AI (Evan Hub­inger) | What’s up in AI safety? (Asya Ber­gal)

EA Global25 Oct 2020 5:48 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Ar­tifi­cial in­tel­li­gence ca­reer stories

EA Global25 Oct 2020 6:56 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Tan Zhi Xuan: AI al­ign­ment, philo­soph­i­cal plu­ral­ism, and the rele­vance of non-Western philosophy

EA Global21 Nov 2020 8:12 UTC
7 points
1 comment1 min readEA link
(www.youtube.com)

AGI risk: analo­gies & arguments

technicalities23 Mar 2021 13:18 UTC
23 points
3 comments7 min readEA link

He­len Toner: The Open Philan­thropy Pro­ject’s work on AI risk

EA Global3 Nov 2017 7:43 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Owain Evans and Vic­to­ria Krakovna: Ca­reers in tech­ni­cal AI safety

EA Global3 Nov 2017 7:43 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Owen Cot­ton-Bar­ratt: What does (and doesn’t) AI mean for effec­tive al­tru­ism?

EA Global11 Aug 2017 8:19 UTC
8 points
0 comments12 min readEA link
(www.youtube.com)

Katja Grace: AI safety

EA Global11 Aug 2017 8:19 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Daniel Dewey: The Open Philan­thropy Pro­ject’s work on po­ten­tial risks from ad­vanced AI

EA Global11 Aug 2017 8:19 UTC
6 points
0 comments17 min readEA link
(www.youtube.com)

Jan Leike, He­len Toner, Malo Bour­gon, and Miles Brundage: Work­ing in AI

EA Global11 Aug 2017 8:19 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Michael Page, Dario Amodei, He­len Toner, Tasha McCauley, Jan Leike, & Owen Cot­ton-Bar­ratt: Mus­ings on AI

EA Global11 Aug 2017 8:19 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Max Teg­mark: Risks and benefits of ad­vanced ar­tifi­cial intelligence

EA Global5 Aug 2016 9:19 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

An­drew Critch: Log­i­cal in­duc­tion — progress in AI alignment

EA Global6 Aug 2016 0:40 UTC
6 points
0 comments1 min readEA link
(www.youtube.com)

Co­her­ence ar­gu­ments im­ply a force for goal-di­rected behavior

Katja_Grace6 Apr 2021 21:44 UTC
13 points
1 comment14 min readEA link
(worldspiritsockpuppet.com)

On AI and Compute

johncrox3 Apr 2019 21:26 UTC
39 points
12 comments8 min readEA link

AI al­ign­ment prize win­ners and next round [link]

RyanCarey20 Jan 2018 12:07 UTC
7 points
1 commentEA link

Atari early

AI Impacts2 Apr 2020 23:28 UTC
34 points
2 comments5 min readEA link
(aiimpacts.org)

[Question] Is there ev­i­dence that recom­mender sys­tems are chang­ing users’ prefer­ences?

zdgroff12 Apr 2021 19:11 UTC
60 points
15 comments1 min readEA link

Why I ex­pect suc­cess­ful (nar­row) alignment

Tobias_Baumann29 Dec 2018 15:46 UTC
18 points
10 commentsEA link
(s-risks.org)

Ma­hen­dra Prasad: Ra­tional group de­ci­sion-making

EA Global8 Jul 2020 15:06 UTC
14 points
0 comments16 min readEA link
(www.youtube.com)

[Question] Brief sum­mary of key dis­agree­ments in AI Risk

iarwain26 Dec 2019 19:40 UTC
31 points
3 comments1 min readEA link

[Question] What con­sid­er­a­tions in­fluence whether I have more in­fluence over short or long timelines?

kokotajlod5 Nov 2020 19:57 UTC
18 points
0 comments1 min readEA link

AGI in a vuln­er­a­ble world

AI Impacts2 Apr 2020 3:43 UTC
17 points
0 comments1 min readEA link
(aiimpacts.org)

Three kinds of competitiveness

AI Impacts2 Apr 2020 3:46 UTC
10 points
0 comments5 min readEA link
(aiimpacts.org)

In­for­ma­tion se­cu­rity ca­reers for GCR reduction

ClaireZabel20 Jun 2019 23:56 UTC
158 points
34 comments8 min readEA link

Nat­u­ral­ism and AI alignment

Michele Campolo24 Apr 2021 16:20 UTC
10 points
3 comments7 min readEA link

Why AI is Harder Than We Think—Me­lanie Mitchell

evelynciara28 Apr 2021 8:19 UTC
42 points
6 comments2 min readEA link
(arxiv.org)

In­for­mat­ica: Spe­cial Is­sue on Superintelligence

RyanCarey3 May 2017 5:05 UTC
7 points
0 commentsEA link

Eric Drexler: Pare­to­topian goal alignment

EA Global15 Mar 2019 14:51 UTC
3 points
0 comments10 min readEA link
(www.youtube.com)

[Question] What harm could AI safety do?

SeanEngelhart15 May 2021 1:11 UTC
12 points
9 comments1 min readEA link

[Question] Why should we *not* put effort into AI safety re­search?

vvh16 May 2021 5:11 UTC
15 points
5 comments1 min readEA link

Pre­dict re­sponses to the “ex­is­ten­tial risk from AI” survey

RobBensinger28 May 2021 1:38 UTC
36 points
8 comments2 min readEA link

Long-Term Fu­ture Fund: May 2021 grant recommendations

abergal27 May 2021 6:44 UTC
110 points
15 comments57 min readEA link

Fi­nal Re­port of the Na­tional Se­cu­rity Com­mis­sion on Ar­tifi­cial In­tel­li­gence (NSCAI, 2021)

MichaelA1 Jun 2021 8:19 UTC
48 points
3 comments4 min readEA link
(www.nscai.gov)

“Ex­is­ten­tial risk from AI” sur­vey results

RobBensinger1 Jun 2021 20:19 UTC
75 points
35 comments11 min readEA link

Sur­vey on AI ex­is­ten­tial risk scenarios

SamClarke8 Jun 2021 17:12 UTC
135 points
5 comments7 min readEA link

Some AI Gover­nance Re­search Ideas

MarkusAnderljung3 Jun 2021 10:51 UTC
72 points
2 comments2 min readEA link

[Question] What is an ex­am­ple of re­cent, tan­gible progress in AI safety re­search?

Aaron Gertler14 Jun 2021 5:29 UTC
34 points
4 comments1 min readEA link

AI Gover­nance Ca­reer Paths for Europeans

careersthrowaway16 May 2020 6:40 UTC
66 points
1 comment12 min readEA link

[Link and com­men­tary] Beyond Near- and Long-Term: Towards a Clearer Ac­count of Re­search Pri­ori­ties in AI Ethics and Society

MichaelA14 Mar 2020 9:04 UTC
17 points
0 comments6 min readEA link

[Question] How do you talk about AI safety?

evelynciara19 Apr 2020 16:15 UTC
10 points
5 comments1 min readEA link

Database of ex­is­ten­tial risk estimates

MichaelA15 Apr 2020 12:43 UTC
87 points
35 comments5 min readEA link

AI Benefits Post 1: In­tro­duc­ing “AI Benefits”

Cullen_OKeefe22 Jun 2020 16:58 UTC
10 points
2 comments3 min readEA link

AI Benefits Post 2: How AI Benefits Differs from AI Align­ment & AI for Good

Cullen_OKeefe29 Jun 2020 16:59 UTC
9 points
0 comments2 min readEA link

Book re­view: Ar­chi­tects of In­tel­li­gence by Martin Ford (2018)

ofer11 Aug 2020 17:24 UTC
11 points
1 comment2 min readEA link

An­i­mal Rights, The Sin­gu­lar­ity, and Astro­nom­i­cal Suffering

deluks91720 Aug 2020 20:23 UTC
42 points
0 comments3 min readEA link

Sin­ga­pore’s Tech­ni­cal AI Align­ment Re­search Ca­reer Guide

yiyang26 Aug 2020 8:09 UTC
32 points
7 comments8 min readEA link

Asya Ber­gal: Rea­sons you might think hu­man-level AI is un­likely to hap­pen soon

EA Global26 Aug 2020 16:01 UTC
23 points
2 comments18 min readEA link
(www.youtube.com)

A course for the gen­eral pub­lic on AI

LeandroD31 Aug 2020 1:29 UTC
1 point
0 comments1 min readEA link

Does gen­er­al­ity pay? GPT-3 can provide pre­limi­nary ev­i­dence.

evelynciara12 Jul 2020 18:53 UTC
20 points
4 comments2 min readEA link

Short-Term AI Align­ment as a Pri­or­ity Cause

len.hoang.lnh11 Feb 2020 16:22 UTC
19 points
11 comments7 min readEA link

[Question] Are so­cial me­dia al­gorithms an ex­is­ten­tial risk?

BarryGrimes15 Sep 2020 8:52 UTC
24 points
13 comments1 min readEA link

Sum­mary of Stu­art Rus­sell’s new book, “Hu­man Com­pat­i­ble”

rohinmshah19 Oct 2019 19:56 UTC
30 points
1 comment15 min readEA link
(www.alignmentforum.org)

Feed­back Re­quest on EA Philip­pines’ Ca­reer Ad­vice Re­search for Tech­ni­cal AI Safety

BrianTan3 Oct 2020 10:39 UTC
18 points
5 comments4 min readEA link

AI risk hub in Sin­ga­pore?

kokotajlod29 Oct 2020 11:51 UTC
23 points
3 comments4 min readEA link

fic­tion about AI risk

Ann Garth12 Nov 2020 22:36 UTC
5 points
1 comment1 min readEA link

[Question] Donat­ing against Short Term AI risks

Jan-WillemvanPutten16 Nov 2020 12:23 UTC
5 points
9 comments1 min readEA link

How Rood­man’s GWP model trans­lates to TAI timelines

kokotajlod16 Nov 2020 14:11 UTC
21 points
0 comments2 min readEA link

Com­pet­i­tive Ethics

mwcvitkovic24 Nov 2020 1:00 UTC
18 points
7 comments4 min readEA link

Defend­ing against Ad­ver­sar­ial Poli­cies in Re­in­force­ment Learn­ing with Alter­nat­ing Training

sergeivolodin8 Dec 2020 18:38 UTC
3 points
0 comments13 min readEA link

Long-Term Fu­ture Fund: Ask Us Any­thing!

AdamGleave3 Dec 2020 13:44 UTC
88 points
154 comments1 min readEA link

[Question] Can we con­vince peo­ple to work on AI safety with­out con­vinc­ing them about AGI hap­pen­ing this cen­tury?

BrianTan26 Nov 2020 14:46 UTC
8 points
3 comments2 min readEA link

Cen­tre for the Study of Ex­is­ten­tial Risk Four Month Re­port June—Septem­ber 2020

HaydnBelfield2 Dec 2020 18:33 UTC
22 points
0 comments18 min readEA link

LessWrong is now a book, available for pre-or­der!

jacobjacob4 Dec 2020 20:42 UTC
48 points
1 comment10 min readEA link

De­fus­ing AGI Danger

Mark Xu24 Dec 2020 23:08 UTC
22 points
0 comments9 min readEA link
(www.alignmentforum.org)

Against GDP as a met­ric for timelines and take­off speeds

kokotajlod29 Dec 2020 17:50 UTC
41 points
6 comments14 min readEA link

Euro­pean Master’s Pro­grams in Ma­chine Learn­ing, Ar­tifi­cial In­tel­li­gence, and re­lated fields

Master Programs ML/AI17 Jan 2021 20:09 UTC
16 points
4 comments1 min readEA link

Birds, Brains, Planes, and AI: Against Ap­peals to the Com­plex­ity/​Mys­te­ri­ous­ness/​Effi­ciency of the Brain

kokotajlod18 Jan 2021 12:39 UTC
25 points
2 comments1 min readEA link

13 Re­cent Publi­ca­tions on Ex­is­ten­tial Risk (Jan 2021 up­date)

HaydnBelfield8 Feb 2021 12:42 UTC
7 points
2 comments10 min readEA link

Stu­art Rus­sell Hu­man Com­pat­i­ble AI Roundtable with Allan Dafoe, Rob Re­ich, & Ma­ri­etje Schaake

Mahendra Prasad11 Feb 2021 7:43 UTC
16 points
0 comments1 min readEA link

In­ter­view with Tom Chivers: “AI is a plau­si­ble ex­is­ten­tial risk, but it feels as if I’m in Pas­cal’s mug­ging”

felix.h21 Feb 2021 13:41 UTC
16 points
1 comment7 min readEA link

Work­ing at EA or­ga­ni­za­tions se­ries: Ma­chine In­tel­li­gence Re­search Institute

SoerenMind1 Nov 2015 12:49 UTC
8 points
0 commentsEA link

AI Fore­cast­ing Dic­tionary (Fore­cast­ing in­fras­truc­ture, part 1)

jacobjacob8 Aug 2019 13:16 UTC
18 points
0 comments5 min readEA link

AI Fore­cast­ing Re­s­olu­tion Coun­cil (Fore­cast­ing in­fras­truc­ture, part 2)

jacobjacob29 Aug 2019 17:43 UTC
28 points
0 comments3 min readEA link

AI Fore­cast­ing Ques­tion Database (Fore­cast­ing in­fras­truc­ture, part 3)

jacobjacob3 Sep 2019 14:57 UTC
23 points
1 comment4 min readEA link

Cri­tique of Su­per­in­tel­li­gence Part 1

Fods1213 Dec 2018 5:10 UTC
20 points
13 commentsEA link

Cri­tique of Su­per­in­tel­li­gence Part 2

Fods1213 Dec 2018 5:12 UTC
7 points
12 commentsEA link

Cri­tique of Su­per­in­tel­li­gence Part 3

Fods1213 Dec 2018 5:13 UTC
1 point
5 commentsEA link

Cri­tique of Su­per­in­tel­li­gence Part 4

Fods1213 Dec 2018 5:14 UTC
2 points
2 commentsEA link

Cri­tique of Su­per­in­tel­li­gence Part 5

Fods1213 Dec 2018 5:19 UTC
10 points
2 commentsEA link

Re­port on Semi-in­for­ma­tive Pri­ors for AI timelines (Open Philan­thropy)

Tom_Davidson26 Mar 2021 17:46 UTC
62 points
6 comments2 min readEA link

The flaws that make to­day’s AI ar­chi­tec­ture un­safe and a new ap­proach that could fix it

80000_Hours22 Jun 2020 22:15 UTC
3 points
0 comments86 min readEA link
(80000hours.org)

Three Bi­ases That Made Me Believe in AI Risk

beth​13 Feb 2019 23:22 UTC
36 points
20 comments3 min readEA link

But ex­actly how com­plex and frag­ile?

Katja_Grace13 Dec 2019 7:05 UTC
36 points
3 comments3 min readEA link
(meteuphoric.com)

Con­fused about AI re­search as a means of ad­dress­ing AI risk

reallyeli21 Feb 2019 0:07 UTC
31 points
15 comments1 min readEA link

Changes in fund­ing in the AI safety field

Sebastian_Farquhar3 Feb 2017 13:09 UTC
28 points
11 commentsEA link

AI Align­ment 2018-2019 Review

Habryka28 Jan 2020 21:14 UTC
28 points
0 comments6 min readEA link
(www.lesswrong.com)

A con­ver­sa­tion with Ro­hin Shah

AI Impacts12 Nov 2019 1:31 UTC
27 points
8 comments33 min readEA link
(aiimpacts.org)

“Tak­ing AI Risk Se­ri­ously” – Thoughts by An­drew Critch

Raemon19 Nov 2018 2:21 UTC
26 points
9 commentsEA link
(www.lesswrong.com)

Amanda Askell: AI safety needs so­cial scientists

EA Global4 Mar 2019 15:50 UTC
26 points
0 comments18 min readEA link
(www.youtube.com)

[Link] Thiel on GCRs

Milan_Griffes22 Jul 2019 20:47 UTC
28 points
11 comments1 min readEA link

Cortés, Pizarro, and Afonso as Prece­dents for Takeover

AI Impacts2 Mar 2020 12:25 UTC
26 points
17 comments11 min readEA link
(aiimpacts.org)

The first AI Safety Camp & onwards

remmelt7 Jun 2018 18:49 UTC
25 points
2 commentsEA link

In­tro to car­ing about AI al­ign­ment as an EA cause

So8res14 Apr 2017 0:42 UTC
24 points
12 commentsEA link

Im­pli­ca­tions of Quan­tum Com­put­ing for Ar­tifi­cial In­tel­li­gence al­ign­ment re­search (ABRIDGED)

Jsevillamol5 Sep 2019 14:56 UTC
24 points
4 comments2 min readEA link

Are Hu­mans ‘Hu­man Com­pat­i­ble’?

Matt Boyd6 Dec 2019 5:49 UTC
22 points
8 comments4 min readEA link

A re­sponse to Matthews on AI Risk

RyanCarey11 Aug 2015 12:58 UTC
11 points
17 commentsEA link

What can the prin­ci­pal-agent liter­a­ture tell us about AI risk?

Alexis Carlier10 Feb 2020 10:10 UTC
29 points
1 comment16 min readEA link

Sup­port­ing global co­or­di­na­tion in AI de­vel­op­ment: Why and how to con­tribute to in­ter­na­tional AI standards

pcihon17 Apr 2019 22:17 UTC
21 points
4 comments1 min readEA link

AI safety schol­ar­ships look worth-fund­ing (if other fund­ing is sane)

anon-a19 Nov 2019 0:59 UTC
21 points
6 comments2 min readEA link

Paul Chris­ti­ano: Cur­rent work in AI alignment

EA Global3 Apr 2020 7:06 UTC
16 points
0 comments23 min readEA link
(www.youtube.com)

AI Safety Ca­reer Bot­tle­necks Sur­vey Re­sponses Responses

Linda Linsefors28 May 2021 10:41 UTC
19 points
1 comment5 min readEA link
No comments.