RSS

AI governance

TagLast edit: 22 Jul 2022 19:52 UTC by Pablo

AI governance (or the governance of artificial intelligence) is the study of norms, policies, and institutions that can help humanity navigate the transition to a world with advanced artificial intelligence. This includes a broad range of subjects, from global coordination around regulating AI development to providing incentives for corporations to be more cautious in their AI research.

AI governance as a career

Long-term AI policy strategy research and implementation is one of 80,000 Hours’ “priority paths”—the most promising career opportunities the organization has identified so far.[1][2]

Further reading

Brundage, Miles (2017) Guide to working in AI policy and strategy, 80,000 Hours, June 7.

Cussins, Jessica (2020) Summaries of AI policy resources, Future of Life Institute.

Dafoe, Allan (2020) AI governance: opportunity and theory of impact, Effective Altruism Forum, September 17.

Muehlhauser, Luke (2021) A personal take on longtermist AI governance, Effective Altruism Forum, July 21.

Related entries

AI alignment | AI forecasting | alignment tax | antitrust law | compute governance | economics of artificial intelligence | ethics of artificial intelligence | global governance | standards and regulation

  1. ^

    Todd, Benjamin (2018) The highest impact career paths our research has identified so far, 80,000 Hours, August 12.

  2. ^

    Todd, Benjamin (2021) Long-term AI policy strategy research and implementation, 80,000 Hours, October.

AI Gover­nance: Op­por­tu­nity and The­ory of Impact

Allan Dafoe17 Sep 2020 6:30 UTC
265 points
20 comments12 min readEA link

Some AI Gover­nance Re­search Ideas

MarkusAnderljung3 Jun 2021 10:51 UTC
102 points
5 comments2 min readEA link

A per­sonal take on longter­mist AI governance

lukeprog16 Jul 2021 22:08 UTC
173 points
7 comments7 min readEA link

Com­po­nents of Strate­gic Clar­ity [Strate­gic Per­spec­tives on Long-term AI Gover­nance, #2]

MMMaas2 Jul 2022 11:22 UTC
66 points
0 comments6 min readEA link

Sin­ga­pore AI Policy Ca­reer Guide

Yi-Yang21 Jan 2021 3:05 UTC
28 points
0 comments5 min readEA link

Katja Grace: Let’s think about slow­ing down AI

peterhartree23 Dec 2022 0:57 UTC
84 points
6 comments2 min readEA link
(worldspiritsockpuppet.substack.com)

A Map to Nav­i­gate AI Governance

hanadulset14 Feb 2022 22:41 UTC
74 points
11 comments25 min readEA link

The longter­mist AI gov­er­nance land­scape: a ba­sic overview

Sam Clarke18 Jan 2022 12:58 UTC
172 points
13 comments9 min readEA link

AI Gover­nance Ca­reer Paths for Europeans

careersthrowaway16 May 2020 6:40 UTC
83 points
1 comment12 min readEA link

FLI open let­ter: Pause gi­ant AI experiments

Zach Stein-Perlman29 Mar 2023 4:04 UTC
220 points
38 comments2 min readEA link
(futureoflife.org)

Slow­ing down AI progress is an un­der­ex­plored al­ign­ment strategy

Michael Huang13 Jul 2022 3:22 UTC
91 points
11 comments3 min readEA link
(www.lesswrong.com)

Case stud­ies of self-gov­er­nance to re­duce tech­nol­ogy risk

jia6 Apr 2021 8:49 UTC
55 points
6 comments7 min readEA link

Techies Wanted: How STEM Back­grounds Can Ad­vance Safe AI Policy

Daniel_Eth26 May 2025 11:29 UTC
41 points
1 comment29 min readEA link

Per­sonal thoughts on ca­reers in AI policy and strategy

carrickflynn27 Sep 2017 16:52 UTC
56 points
28 comments18 min readEA link

Strate­gic Per­spec­tives on Trans­for­ma­tive AI Gover­nance: Introduction

MMMaas2 Jul 2022 11:20 UTC
115 points
18 comments4 min readEA link

Some un­der­rated rea­sons why the AI safety com­mu­nity should re­con­sider its em­brace of strict li­a­bil­ity

Cecil Abungu 8 Apr 2024 18:50 UTC
67 points
29 comments12 min readEA link

12 ten­ta­tive ideas for US AI policy (Luke Muehlhauser)

Lizka19 Apr 2023 21:05 UTC
117 points
12 comments4 min readEA link
(www.openphilanthropy.org)

AI Gover­nance Course—Cur­ricu­lum and Application

Mau29 Nov 2021 13:29 UTC
94 points
9 comments1 min readEA link

An Anal­y­sis of Sys­temic Risk and Ar­chi­tec­tural Re­quire­ments for the Con­tain­ment of Re­cur­sively Self-Im­prov­ing AI

Ihor Ivliev17 Jun 2025 0:16 UTC
2 points
5 comments4 min readEA link

Are We Ready for Digi­tal Per­sons?

Alex (Αλέξανδρος)3 Jun 2025 9:38 UTC
3 points
0 comments1 min readEA link
(www.linkedin.com)

Shift Re­sources to Ad­vo­cacy Now (Post 4 of 7 on AI Gover­nance)

Jason Green-Lowe28 May 2025 1:19 UTC
58 points
5 comments32 min readEA link

A sum­mary of cur­rent work in AI governance

constructive17 Jun 2023 16:58 UTC
89 points
4 comments11 min readEA link

GovAI: Towards best prac­tices in AGI safety and gov­er­nance: A sur­vey of ex­pert opinion

Zach Stein-Perlman15 May 2023 1:42 UTC
68 points
5 comments1 min readEA link
(arxiv.org)

We’re Not Ad­ver­tis­ing Enough (Post 3 of 7 on AI Gover­nance)

Jason Green-Lowe22 May 2025 17:11 UTC
53 points
3 comments28 min readEA link

The Cen­ter for AI Policy Has Shut Down

Tristan W16 Sep 2025 17:33 UTC
122 points
25 comments14 min readEA link

In­tro­duc­ing Deep Dive, a 201 AI policy course

Kambar17 Jun 2025 16:50 UTC
31 points
2 comments2 min readEA link

Con­sider grant­ing AIs freedom

Matthew_Barnett6 Dec 2024 0:55 UTC
100 points
38 comments5 min readEA link

[ur­gent] Amer­i­cans: call your Se­na­tors and tell them you op­pose AI preemption

Holly Elmore ⏸️ 🔸15 May 2025 1:57 UTC
176 points
22 comments2 min readEA link

Bi­den-Har­ris Ad­minis­tra­tion An­nounces First-Ever Con­sor­tium Ded­i­cated to AI Safety

ben.smith9 Feb 2024 6:40 UTC
15 points
1 comment1 min readEA link
(www.nist.gov)

The U.S. Na­tional Se­cu­rity State is Here to Make AI Even Less Trans­par­ent and Accountable

Matrice Jacobine🔸🏳️‍⚧️24 Nov 2024 9:34 UTC
7 points
0 comments2 min readEA link
(www.eff.org)

Why En­gag­ing with Global Ma­jor­ity AI Policy Matters

Heramb Podar2 Jul 2025 1:54 UTC
1 point
0 comments1 min readEA link
(www.lesswrong.com)

AGI safety ca­reer advice

richard_ngo2 May 2023 7:36 UTC
213 points
18 comments13 min readEA link

A Viral Li­cense for AI Safety

IvanVendrov5 Jun 2021 2:00 UTC
32 points
6 comments5 min readEA link

CSER Ad­vice to EU High-Level Ex­pert Group on AI

HaydnBelfield8 Mar 2019 20:42 UTC
14 points
0 comments5 min readEA link
(www.cser.ac.uk)

Some Things I Heard about AI Gover­nance at EAG

utilistrutil28 Feb 2023 21:27 UTC
35 points
5 comments6 min readEA link

[Question] Ex­am­ples of self-gov­er­nance to re­duce tech­nol­ogy risk?

jia25 Sep 2020 13:26 UTC
32 points
1 comment1 min readEA link

AI policy ideas: Read­ing list

Zach Stein-Perlman17 Apr 2023 19:00 UTC
60 points
3 comments4 min readEA link

Data Tax­a­tion: A Pro­posal for Slow­ing Down AGI Progress

Per Ivar Friborg11 Apr 2023 17:27 UTC
42 points
6 comments12 min readEA link

In­ter­ac­tive AI Gover­nance Map

Hamish McDoodles12 Mar 2024 10:02 UTC
67 points
8 comments1 min readEA link

[Link post] Co­or­di­na­tion challenges for pre­vent­ing AI conflict

stefan.torges9 Mar 2021 9:39 UTC
58 points
0 comments1 min readEA link
(longtermrisk.org)

Markus An­der­ljung and Ben Garfinkel: Fireside chat on AI governance

EA Global24 Jul 2020 14:56 UTC
25 points
0 comments16 min readEA link
(www.youtube.com)

“Pivotal Act” In­ten­tions: Nega­tive Con­se­quences and Fal­la­cious Arguments

Andrew Critch19 Apr 2022 20:24 UTC
80 points
10 comments7 min readEA link

Cur­rent UK gov­ern­ment lev­ers on AI development

rosehadshar10 Apr 2023 13:16 UTC
82 points
3 comments4 min readEA link

Mud­dling Along Is More Likely Than Dystopia

Jeffrey Heninger21 Oct 2023 9:30 UTC
87 points
3 comments8 min readEA link
(blog.aiimpacts.org)

A new pro­posal for reg­u­lat­ing AI in the EU

EdoArad26 Apr 2021 17:25 UTC
37 points
3 comments1 min readEA link
(www.bbc.com)

New Busi­ness Wars pod­cast sea­son on Sam Alt­man and OpenAI

Eevee🔹2 Apr 2024 6:22 UTC
10 points
0 comments1 min readEA link
(wondery.com)

AI Safety Ac­tion Plan—A re­port com­mis­sioned by the US State Department

Agustín Covarrubias 🔸11 Mar 2024 22:13 UTC
25 points
1 comment1 min readEA link
(www.gladstone.ai)

A Case for Su­per­hu­man Gover­nance, us­ing AI

Ozzie Gooen7 Jun 2024 0:10 UTC
59 points
26 comments10 min readEA link

Pro­ject ideas: Gover­nance dur­ing ex­plo­sive tech­nolog­i­cal growth

Lukas Finnveden4 Jan 2024 7:25 UTC
37 points
1 comment16 min readEA link
(www.forethought.org)

Cor­po­rate cam­paigns work: a key learn­ing for AI Safety

Jamie_Harris17 Aug 2023 21:35 UTC
73 points
12 comments6 min readEA link

The Gover­nance Prob­lem and the “Pretty Good” X-Risk

Zach Stein-Perlman28 Aug 2021 20:00 UTC
23 points
4 comments11 min readEA link

He­len Toner (ex-OpenAI board mem­ber): “We learned about ChatGPT on Twit­ter.”

defun 🔸29 May 2024 7:40 UTC
123 points
13 comments1 min readEA link
(x.com)

“Aligned with who?” Re­sults of sur­vey­ing 1,000 US par­ti­ci­pants on AI values

Holly Morgan21 Mar 2023 22:07 UTC
41 points
0 comments2 min readEA link
(www.lesswrong.com)

Overview of in­tro­duc­tory re­sources in AI Governance

Lucie Philippon 🔸27 May 2024 16:22 UTC
27 points
2 comments6 min readEA link
(www.lesswrong.com)

Large Lan­guage Models as Fi­du­cia­ries to Humans

johnjnay24 Jan 2023 19:53 UTC
25 points
0 comments34 min readEA link
(papers.ssrn.com)

My fa­vorite AI gov­er­nance re­search this year so far

Zach Stein-Perlman23 Jul 2023 22:00 UTC
81 points
4 comments7 min readEA link
(blog.aiimpacts.org)

Open Philan­thropy’s AI gov­er­nance grant­mak­ing (so far)

Aaron Gertler 🔸17 Dec 2020 12:00 UTC
63 points
0 comments6 min readEA link
(www.openphilanthropy.org)

AGI Policy Readi­ness by 2027: Fore­sight into In­sti­tu­tional Adaptation

Josephine Schwab11 Nov 2025 21:02 UTC
1 point
0 comments1 min readEA link

Slightly against al­ign­ing with neo-luddites

Matthew_Barnett26 Dec 2022 23:27 UTC
77 points
17 comments4 min readEA link

FLI re­port: Poli­cy­mak­ing in the Pause

Zach Stein-Perlman15 Apr 2023 17:01 UTC
29 points
4 comments1 min readEA link
(futureoflife.org)

Par­tial Tran­script of Re­cent Se­nate Hear­ing Dis­cussing AI X-Risk

Daniel_Eth27 Jul 2023 9:16 UTC
150 points
2 comments22 min readEA link
(medium.com)

AI-Rele­vant Reg­u­la­tion: IAEA

SWK15 Jul 2023 18:20 UTC
10 points
0 comments5 min readEA link

FHI Re­port: Stable Agree­ments in Tur­bu­lent Times

Cullen 🔸21 Feb 2019 17:12 UTC
25 points
2 comments4 min readEA link
(www.fhi.ox.ac.uk)

List of pro­jects that seem im­pact­ful for AI Governance

JaimeRV14 Jan 2024 16:52 UTC
40 points
2 comments13 min readEA link

AI In­ci­dent Re­port­ing: A Reg­u­la­tory Review

Deric Cheng11 Mar 2024 21:02 UTC
10 points
1 comment6 min readEA link

It’s not ob­vi­ous that get­ting dan­ger­ous AI later is better

Aaron_Scher23 Sep 2023 5:35 UTC
23 points
9 comments16 min readEA link

[link] Cen­tre for the Gover­nance of AI 2020 An­nual Report

MarkusAnderljung14 Jan 2021 10:23 UTC
11 points
5 comments1 min readEA link

AI policy & gov­er­nance in Aus­tralia: notes from an ini­tial discussion

Alexander Saeri15 May 2023 0:00 UTC
31 points
1 comment3 min readEA link

AGI ris­ing: why we are in a new era of acute risk and in­creas­ing pub­lic aware­ness, and what to do now

Greg_Colbourn ⏸️ 2 May 2023 10:17 UTC
68 points
35 comments13 min readEA link

“Risk Aware­ness Mo­ments” (Rams): A con­cept for think­ing about AI gov­er­nance interventions

oeg14 Apr 2023 17:40 UTC
53 points
0 comments9 min readEA link

An­titrust-Com­pli­ant AI In­dus­try Self-Regulation

Cullen 🔸7 Jul 2020 20:52 UTC
26 points
1 comment1 min readEA link
(cullenokeefe.com)

Thoughts on the AI Safety Sum­mit com­pany policy re­quests and responses

So8res31 Oct 2023 23:54 UTC
42 points
3 comments10 min readEA link

AI Benefits Post 4: Out­stand­ing Ques­tions on Select­ing Benefits

Cullen 🔸14 Jul 2020 17:24 UTC
6 points
0 comments5 min readEA link

AI-Rele­vant Reg­u­la­tion: CERN

SWK15 Jul 2023 18:40 UTC
12 points
0 comments6 min readEA link

(Even) More Early-Ca­reer EAs Should Try AI Safety Tech­ni­cal Research

tlevin30 Jun 2022 21:14 UTC
86 points
40 comments11 min readEA link

AI Risk Man­age­ment Frame­work | NIST

𝕮𝖎𝖓𝖊𝖗𝖆26 Jan 2023 15:27 UTC
50 points
0 comments2 min readEA link
(www.nist.gov)

Stan­dard policy frame­works for AI governance

Nathan_Barnard30 Jan 2024 18:14 UTC
27 points
2 comments3 min readEA link

New US Se­nate Bill on X-Risk Miti­ga­tion [Linkpost]

Evan R. Murphy4 Jul 2022 1:28 UTC
22 points
12 comments1 min readEA link
(www.hsgac.senate.gov)

A case for donat­ing to AI risk re­duc­tion (in­clud­ing if you work in AI)

tlevin2 Dec 2024 19:05 UTC
118 points
5 comments3 min readEA link

Should you work in the Euro­pean Union to do AGI gov­er­nance?

hanadulset31 Jan 2022 10:34 UTC
90 points
20 comments15 min readEA link

Prob­lem-solv­ing tasks in Graph The­ory for lan­guage mod­els

Bruno López Orozco1 Oct 2024 12:36 UTC
21 points
1 comment9 min readEA link

A fic­tional AI law laced w/​ al­ign­ment theory

Miguel17 Jul 2023 3:26 UTC
3 points
0 comments2 min readEA link

UK AI Bill Anal­y­sis & Opinion

CAISID5 Feb 2024 0:12 UTC
18 points
0 comments15 min readEA link

[Question] Pre­dic­tions for fu­ture AI gov­er­nance?

jackchang1102 Apr 2023 16:43 UTC
4 points
1 comment1 min readEA link

Cur­rent paths to im­pact in EU AI Policy (Feb ’24)

JOMG_Monnet12 Feb 2024 15:57 UTC
48 points
1 comment5 min readEA link

AI and Work: Sum­maris­ing a New Liter­a­ture Review

cpeppiatt15 Jul 2024 10:27 UTC
13 points
0 comments2 min readEA link
(arxiv.org)

UNGA Re­s­olu­tion on AI: 5 Key Take­aways Look­ing to Fu­ture Policy

Heramb Podar24 Mar 2024 12:03 UTC
17 points
1 comment3 min readEA link

News: Span­ish AI image out­cry + US AI work­force “reg­u­la­tion”

Benevolent_Rain26 Sep 2023 7:43 UTC
9 points
0 comments1 min readEA link

Prospects for AI safety agree­ments be­tween countries

oeg14 Apr 2023 17:41 UTC
104 points
3 comments22 min readEA link

Map­ping How Alli­ances, Ac­qui­si­tions, and An­titrust are Shap­ing the Fron­tier AI Industry

t6aguirre3 Jun 2024 9:43 UTC
24 points
1 comment2 min readEA link

What suc­cess looks like

mariushobbhahn28 Jun 2022 14:30 UTC
115 points
20 comments19 min readEA link

Align­ing the Align­ers: En­sur­ing Aligned AI acts for the com­mon good of all mankind

timunderwood16 Jan 2023 11:13 UTC
40 points
2 comments4 min readEA link

Aus­trali­ans are con­cerned about AI risks and ex­pect strong gov­ern­ment action

Alexander Saeri8 Mar 2024 6:39 UTC
38 points
12 comments5 min readEA link
(aigovernance.org.au)

Learn­ing so­cietal val­ues from law as part of an AGI al­ign­ment strategy

johnjnay21 Oct 2022 2:03 UTC
20 points
1 comment24 min readEA link

Are you re­ally in a race? The Cau­tion­ary Tales of Szilárd and Ellsberg

HaydnBelfield19 May 2022 8:42 UTC
494 points
45 comments18 min readEA link

Le­gal Pri­ori­ties Re­search: A Re­search Agenda

jonasschuett6 Jan 2021 21:47 UTC
58 points
4 comments1 min readEA link

Sur­vey on in­ter­me­di­ate goals in AI governance

MichaelA🔸17 Mar 2023 12:44 UTC
156 points
4 comments1 min readEA link

What we’re miss­ing: the case for struc­tural risks from AI

Justin Olive9 Nov 2023 5:52 UTC
32 points
3 comments6 min readEA link

Is Democ­racy a Fad?

bmg13 Mar 2021 12:40 UTC
166 points
36 comments18 min readEA link

We are fight­ing a shared bat­tle (a call for a differ­ent ap­proach to AI Strat­egy)

GideonF16 Mar 2023 14:37 UTC
59 points
10 comments15 min readEA link

Call for Papers on Global AI Gover­nance from the UN

Chris Leong20 Aug 2023 8:56 UTC
36 points
1 comment1 min readEA link
(www.linkedin.com)

FHI Re­port: The Wind­fall Clause: Distribut­ing the Benefits of AI for the Com­mon Good

Cullen 🔸5 Feb 2020 23:49 UTC
55 points
21 comments2 min readEA link

AI labs’ state­ments on governance

Zach Stein-Perlman4 Jul 2023 16:30 UTC
28 points
1 comment36 min readEA link

Merger of Deep­Mind and Google Brain

Greg_Colbourn ⏸️ 20 Apr 2023 20:16 UTC
11 points
12 comments1 min readEA link
(blog.google)

[US] NTIA: AI Ac­countabil­ity Policy Re­quest for Comment

Kyle J. Lucchese 🔸13 Apr 2023 16:12 UTC
47 points
4 comments1 min readEA link
(ntia.gov)

What is com­pute gov­er­nance?

Vishakha Agrawal23 Dec 2024 6:45 UTC
5 points
0 comments2 min readEA link
(aisafety.info)

EU AI Act now has a sec­tion on gen­eral pur­pose AI systems

MathiasKB🔸9 Dec 2021 12:40 UTC
64 points
10 comments1 min readEA link

My Feed­back to the UN Ad­vi­sory Body on AI

Heramb Podar4 Apr 2024 23:39 UTC
7 points
1 comment4 min readEA link

[Question] Know a grad stu­dent study­ing AI’s eco­nomic im­pacts?

Madhav Malhotra5 Jul 2023 0:07 UTC
7 points
0 comments1 min readEA link

Assess­ing the Danger­ous­ness of Malev­olent Ac­tors in AGI Gover­nance: A Pre­limi­nary Exploration

Callum Hinchcliffe14 Oct 2023 21:18 UTC
28 points
4 comments9 min readEA link

An ar­gu­ment for ac­cel­er­at­ing in­ter­na­tional AI gov­er­nance re­search (part 1)

MattThinks16 Aug 2023 5:40 UTC
10 points
0 comments3 min readEA link

My thoughts on the so­cial re­sponse to AI risk

Matthew_Barnett1 Nov 2023 21:27 UTC
116 points
17 comments10 min readEA link

In­for­ma­tion se­cu­rity ca­reers for GCR reduction

ClaireZabel20 Jun 2019 23:56 UTC
187 points
35 comments8 min readEA link

EU poli­cy­mak­ers reach an agree­ment on the AI Act

tlevin15 Dec 2023 6:03 UTC
109 points
11 comments7 min readEA link

NIST AI Risk Man­age­ment Frame­work re­quest for in­for­ma­tion (RFI)

Aryeh Englander31 Aug 2021 22:24 UTC
7 points
0 comments2 min readEA link

An­nounc­ing Open Philan­thropy’s AI gov­er­nance and policy RFP

JulianHazell17 Jul 2024 0:25 UTC
73 points
2 comments1 min readEA link
(www.openphilanthropy.org)

Two con­trast­ing mod­els of “in­tel­li­gence” and fu­ture growth

Magnus Vinding24 Nov 2022 11:54 UTC
74 points
32 comments22 min readEA link

[Link] Cen­ter for the Gover­nance of AI (GovAI) An­nual Re­port 2018

MarkusAnderljung21 Dec 2018 16:17 UTC
24 points
0 comments1 min readEA link

The Im­por­tance of Ar­tifi­cial Sentience

Jamie_Harris3 Mar 2021 17:17 UTC
71 points
10 comments11 min readEA link
(www.sentienceinstitute.org)

Ideal gov­er­nance (for com­pa­nies, coun­tries and more)

Holden Karnofsky7 Apr 2022 16:54 UTC
80 points
19 comments14 min readEA link

An­thropic’s sub­mis­sion to the White House’s RFI on AI policy

Agustín Covarrubias 🔸6 Mar 2025 22:47 UTC
48 points
7 comments1 min readEA link
(www.anthropic.com)

Truth­ful AI

Owen Cotton-Barratt20 Oct 2021 15:11 UTC
55 points
14 comments10 min readEA link

FLI pod­cast se­ries, “Imag­ine A World”, about as­pira­tional fu­tures with AGI

Jackson Wagner13 Oct 2023 16:03 UTC
18 points
0 comments4 min readEA link

60 U.K. Law­mak­ers Ac­cuse Google of Break­ing AI Safety Pledge

Joseph Miller29 Aug 2025 16:09 UTC
15 points
0 comments1 min readEA link
(time.com)

Why don’t gov­ern­ments seem to mind that com­pa­nies are ex­plic­itly try­ing to make AGIs?

Ozzie Gooen23 Dec 2021 7:08 UTC
82 points
49 comments2 min readEA link

Thoughts on yes­ter­day’s UN Se­cu­rity Coun­cil meet­ing on AI

Greg_Colbourn ⏸️ 19 Jul 2023 16:46 UTC
31 points
2 comments1 min readEA link

In­tro­duc­ing the Path­fin­der Fel­low­ship: Fund­ing and Men­tor­ship for AI Safety Group Organizers

Agustín Covarrubias 🔸22 Jul 2025 17:11 UTC
49 points
0 comments2 min readEA link

AMA: Markus An­der­ljung (PM at GovAI, FHI)

MarkusAnderljung21 Sep 2020 11:23 UTC
49 points
24 comments2 min readEA link

An­nounc­ing the EU Tech Policy Fellowship

Jan-Willem30 Mar 2022 8:15 UTC
53 points
4 comments5 min readEA link

Law-Fol­low­ing AI 3: Lawless AI Agents Un­der­mine Sta­bi­liz­ing Agreements

Cullen 🔸27 Apr 2022 17:20 UTC
28 points
3 comments3 min readEA link

AI gov­er­nance tal­ent pro­files I’d like to see ap­ply for OP funding

JulianHazell19 Dec 2023 12:34 UTC
119 points
4 comments3 min readEA link
(www.openphilanthropy.org)

An in­ter­ven­tion to shape policy di­alogue, com­mu­ni­ca­tion, and AI re­search norms for AI safety

Lee_Sharkey1 Oct 2017 18:29 UTC
9 points
28 comments10 min readEA link

[Fic­tion] Im­proved Gover­nance on the Crit­i­cal Path to AI Align­ment by 2045.

Jackson Wagner18 May 2022 15:50 UTC
20 points
1 comment12 min readEA link

Three pillars for avoid­ing AGI catas­tro­phe: Tech­ni­cal al­ign­ment, de­ploy­ment de­ci­sions, and co­or­di­na­tion

LintzA3 Aug 2022 21:24 UTC
93 points
4 comments11 min readEA link

Chris­ti­ano, Co­tra, and Yud­kowsky on AI progress

Ajeya25 Nov 2021 16:30 UTC
18 points
6 comments68 min readEA link

Good policy ideas that won’t hap­pen (yet)

Niel_Bowerman11 Sep 2014 12:29 UTC
28 points
8 comments14 min readEA link

UK AI Policy Re­port: Con­tent, Sum­mary, and its Im­pact on EA Cause Areas

Algo_Law21 Jul 2022 17:32 UTC
9 points
1 comment9 min readEA link

Co­or­di­na­tion by com­mon knowl­edge to pre­vent un­con­trol­lable AI

Karl von Wendt14 May 2023 13:37 UTC
14 points
0 comments9 min readEA link

Fi­nal Re­port of the Na­tional Se­cu­rity Com­mis­sion on Ar­tifi­cial In­tel­li­gence (NSCAI, 2021)

MichaelA🔸1 Jun 2021 8:19 UTC
51 points
3 comments4 min readEA link
(www.nscai.gov)

Here are the fi­nal­ists from FLI’s $100K Wor­ld­build­ing Contest

Jackson Wagner6 Jun 2022 18:42 UTC
44 points
5 comments2 min readEA link

Pro­ject ideas: Epistemics

Lukas Finnveden4 Jan 2024 7:26 UTC
43 points
1 comment17 min readEA link
(www.forethought.org)

Fron­tier Model Forum

Zach Stein-Perlman26 Jul 2023 14:30 UTC
40 points
7 comments4 min readEA link
(blog.google)

Mak­ing a con­ser­va­tive case for alignment

Larks17 Nov 2024 1:45 UTC
44 points
0 comments1 min readEA link
(www.lesswrong.com)

“Govern­abil­ity-By-De­sign”: Pon­der­ings on Why We Haven’t Died From Nu­clear Catas­tro­phe (And What We Can Learn From This)

C.K.19 Aug 2025 18:20 UTC
5 points
2 comments6 min readEA link
(proteinstoparadigms.substack.com)

Re­think Pri­ori­ties is hiring a Com­pute Gover­nance Re­searcher or Re­search Assistant

MichaelA🔸7 Jun 2023 13:22 UTC
36 points
2 comments8 min readEA link
(careers.rethinkpriorities.org)

Cruxes on US lead for some do­mes­tic AI regulation

Zach Stein-Perlman10 Sep 2023 18:00 UTC
20 points
6 comments2 min readEA link

An overview of ar­gu­ments for con­cern about automation

LintzA6 Aug 2019 7:56 UTC
34 points
3 comments13 min readEA link

Crises re­veal centralisation

Vasco Grilo🔸26 Mar 2024 18:00 UTC
31 points
2 comments5 min readEA link
(stefanschubert.substack.com)

On the com­pute gov­er­nance era and what has to come af­ter (Len­nart Heim on The 80,000 Hours Pod­cast)

80000_Hours23 Jun 2023 20:11 UTC
37 points
0 comments18 min readEA link

Dona­tion offsets for ChatGPT Plus subscriptions

Jeffrey Ladish16 Mar 2023 23:11 UTC
76 points
10 comments3 min readEA link

Ar­gu­ment Against Im­pact: EU Is Not an AI Su­per­power

EU AI Governance31 Jan 2022 9:48 UTC
35 points
9 comments4 min readEA link

An ‘AGI Emer­gency Eject Cri­te­ria’ con­sen­sus could be re­ally use­ful.

tcelferact7 Apr 2023 16:21 UTC
27 points
3 comments1 min readEA link

[Question] What kind of or­ga­ni­za­tion should be the first to de­velop AGI in a po­ten­tial arms race?

Eevee🔹17 Jul 2022 17:41 UTC
10 points
2 comments1 min readEA link

TAI Safety Biblio­graphic Database

Jess_Riedel22 Dec 2020 16:03 UTC
61 points
9 comments17 min readEA link

No­body’s on the ball on AGI alignment

leopold29 Mar 2023 14:26 UTC
328 points
66 comments9 min readEA link
(www.forourposterity.com)

Daniel Dewey: The Open Philan­thropy Pro­ject’s work on po­ten­tial risks from ad­vanced AI

EA Global11 Aug 2017 8:19 UTC
7 points
0 comments18 min readEA link
(www.youtube.com)

Three Im­pacts of Ma­chine Intelligence

Paul_Christiano23 Aug 2013 10:10 UTC
33 points
5 comments8 min readEA link
(rationalaltruist.com)

EU’s im­por­tance for AI gov­er­nance is con­di­tional on AI tra­jec­to­ries—a case study

MathiasKB🔸13 Jan 2022 14:58 UTC
31 points
2 comments3 min readEA link

Jaan Tal­linn: Fireside chat (2020)

EA Global21 Nov 2020 8:12 UTC
7 points
0 comments1 min readEA link
(www.youtube.com)

Law-Fol­low­ing AI 2: In­tent Align­ment + Su­per­in­tel­li­gence → Lawless AI (By De­fault)

Cullen 🔸27 Apr 2022 17:18 UTC
19 points
0 comments6 min readEA link

FHI Re­port: How Will Na­tional Se­cu­rity Con­sid­er­a­tions Affect An­titrust De­ci­sions in AI? An Ex­am­i­na­tion of His­tor­i­cal Precedents

Cullen 🔸28 Jul 2020 18:33 UTC
13 points
0 comments1 min readEA link
(www.fhi.ox.ac.uk)

Care­less talk on US-China AI com­pe­ti­tion? (and crit­i­cism of CAIS cov­er­age)

Oliver Sourbut20 Sep 2023 12:46 UTC
52 points
19 comments9 min readEA link
(www.oliversourbut.net)

[Question] What should I ask Ezra Klein about AI policy pro­pos­als?

Robert_Wiblin23 Jun 2023 16:36 UTC
21 points
4 comments1 min readEA link

Effec­tive Utopia & Nar­row Way There: Math-Proven Safe Static Mul­tiver­sal mAX-In­tel­li­gence (AXI), Mul­tiver­sal Align­ment, Phys­i­cal­ized Ethics… (Aug 11)

ank2 Mar 2025 3:14 UTC
1 point
3 comments38 min readEA link

GovAI An­nual Re­port 2021

GovAI5 Jan 2022 16:57 UTC
52 points
2 comments9 min readEA link

Shal­low eval­u­a­tions of longter­mist organizations

NunoSempere24 Jun 2021 15:31 UTC
193 points
34 comments34 min readEA link

AI 2027: What Su­per­in­tel­li­gence Looks Like (Linkpost)

Manuel Allgaier11 Apr 2025 10:31 UTC
51 points
3 comments42 min readEA link
(ai-2027.com)

Why Did Elon Musk Just Offer to Buy Con­trol of OpenAI for $100 Billion?

Garrison11 Feb 2025 0:20 UTC
152 points
2 comments6 min readEA link
(garrisonlovely.substack.com)

Four Ques­tions to Refine Your Policy Proposal

Jason Green-Lowe9 Oct 2025 16:30 UTC
22 points
1 comment6 min readEA link

The Ri­val AI De­ploy­ment Prob­lem: a Pre-de­ploy­ment Agree­ment as the least-bad response

HaydnBelfield23 Sep 2022 9:28 UTC
44 points
1 comment12 min readEA link

Dis­cus­sion with Eliezer Yud­kowsky on AGI interventions

RobBensinger11 Nov 2021 3:21 UTC
60 points
33 comments34 min readEA link

AMA: The new Open Philan­thropy Tech­nol­ogy Policy Fellowship

lukeprog26 Jul 2021 15:11 UTC
38 points
14 comments1 min readEA link

I’m Cul­len O’Keefe, a Policy Re­searcher at OpenAI, AMA

Cullen 🔸11 Jan 2020 4:13 UTC
45 points
68 comments1 min readEA link

Nav­i­gat­ing the Open-Source AI Land­scape: Data, Fund­ing, and Safety

AndreFerretti12 Apr 2023 10:30 UTC
23 points
3 comments10 min readEA link

UK gov­ern­ment to host first global sum­mit on AI Safety

DavidNash8 Jun 2023 13:24 UTC
78 points
1 comment5 min readEA link
(www.gov.uk)

[Question] What type of Master’s is best for AI policy work?

Milan Griffes22 Feb 2019 20:04 UTC
14 points
7 comments1 min readEA link

The state of AI in differ­ent coun­tries — an overview

Lizka14 Sep 2023 10:37 UTC
68 points
6 comments13 min readEA link
(aisafetyfundamentals.com)

GovAI We­bi­nars on the Gover­nance and Eco­nomics of AI

MarkusAnderljung12 May 2020 15:00 UTC
16 points
0 comments1 min readEA link

Yud­kowsky and Chris­ti­ano dis­cuss “Take­off Speeds”

EliezerYudkowsky22 Nov 2021 19:42 UTC
42 points
0 comments60 min readEA link

Effec­tive En­force­abil­ity of EU Com­pe­ti­tion Law Un­der Differ­ent AI Devel­op­ment Sce­nar­ios: A Frame­work for Le­gal Analysis

HaydnBelfield19 Aug 2022 17:20 UTC
11 points
0 comments6 min readEA link
(verfassungsblog.de)

Con­crete ac­tion­able poli­cies rele­vant to AI safety (writ­ten 2019)

weeatquince16 Dec 2022 18:41 UTC
48 points
0 comments22 min readEA link

Thoughts on “The Offense-Defense Balance Rarely Changes”

Cullen 🔸12 Feb 2024 3:26 UTC
42 points
4 comments5 min readEA link

Un­jour­nal: Eval­u­a­tions of “Ar­tifi­cial In­tel­li­gence and Eco­nomic Growth”, and new host­ing space

david_reinstein17 Mar 2023 20:20 UTC
47 points
0 comments2 min readEA link
(unjournal.pubpub.org)

The Case for AI Adap­ta­tion: The Per­ils of Liv­ing in a World with Aligned and Well-De­ployed Trans­for­ma­tive Ar­tifi­cial Intelligence

HTC30 May 2023 18:29 UTC
5 points
1 comment7 min readEA link

Im­por­tant, ac­tion­able re­search ques­tions for the most im­por­tant century

Holden Karnofsky24 Feb 2022 16:34 UTC
301 points
13 comments19 min readEA link

Trends in the dol­lar train­ing cost of ma­chine learn­ing systems

Ben Cottier1 Feb 2023 14:48 UTC
63 points
3 comments2 min readEA link
(epochai.org)

Some cruxes on im­pact­ful al­ter­na­tives to AI policy work

richard_ngo22 Nov 2018 13:43 UTC
28 points
2 comments12 min readEA link

CSER and FHI ad­vice to UN High-level Panel on Digi­tal Co­op­er­a­tion

HaydnBelfield8 Mar 2019 20:39 UTC
22 points
7 comments6 min readEA link
(www.cser.ac.uk)

[Question] Books and lec­ture se­ries rele­vant to AI gov­er­nance?

MichaelA🔸18 Jul 2021 15:54 UTC
22 points
8 comments1 min readEA link

Ross Gruet­zemacher: Defin­ing and un­pack­ing trans­for­ma­tive AI

EA Global18 Oct 2019 8:22 UTC
9 points
0 comments1 min readEA link
(www.youtube.com)

Allan Dafoe: Prepar­ing for AI — risks and opportunities

EA Global3 Nov 2017 7:43 UTC
7 points
0 comments1 min readEA link
(www.youtube.com)

OpenAI’s grant pro­gram for demo­cratic pro­cess for de­cid­ing what rules AI sys­tems should follow

Ronen Bar23 Jun 2023 10:46 UTC
7 points
0 comments1 min readEA link

Jeffrey Ding: Bring­ing techno-global­ism back: a ro­man­ti­cally re­al­ist re­fram­ing of the US-China tech relationship

EA Global21 Nov 2020 8:12 UTC
9 points
0 comments1 min readEA link
(www.youtube.com)

What is the EU AI Act and why should you care about it?

MathiasKB🔸10 Sep 2021 7:47 UTC
117 points
10 comments7 min readEA link

AI Safety Newslet­ter #3: AI policy pro­pos­als and a new challenger approaches

Oliver Z25 Apr 2023 16:15 UTC
35 points
1 comment4 min readEA link
(newsletter.safe.ai)

The case for build­ing ex­per­tise to work on US AI policy, and how to do it

80000_Hours31 Jan 2019 22:44 UTC
37 points
2 comments2 min readEA link

Law-Fol­low­ing AI 4: Don’t Rely on Vi­car­i­ous Liability

Cullen 🔸2 Aug 2022 23:23 UTC
13 points
0 comments3 min readEA link

New vol­un­tary com­mit­ments (AI Seoul Sum­mit)

Zach Stein-Perlman21 May 2024 11:00 UTC
12 points
1 comment7 min readEA link
(www.gov.uk)

Pal­isade is hiring Re­search Engineers

Charlie Rogers-Smith11 Nov 2023 3:09 UTC
23 points
0 comments3 min readEA link

Sum­mary of the AI Bill of Rights and Policy Implications

Tristan W20 Jun 2023 9:28 UTC
16 points
0 comments22 min readEA link

What SB 53, Cal­ifor­nia’s new AI law, does

tlevin29 Sep 2025 23:29 UTC
78 points
1 comment4 min readEA link

Un­veiling the Amer­i­can Public Opinion on AI Mo­ra­to­rium and Govern­ment In­ter­ven­tion: The Im­pact of Me­dia Exposure

Otto8 May 2023 10:49 UTC
28 points
5 comments6 min readEA link

The EU AI Act needs a defi­ni­tion of high-risk foun­da­tion mod­els to avoid reg­u­la­tory over­reach and backlash

matthias_samwald31 May 2023 15:34 UTC
17 points
0 comments4 min readEA link

Owen Cot­ton-Bar­ratt: What does (and doesn’t) AI mean for effec­tive al­tru­ism?

EA Global11 Aug 2017 8:19 UTC
10 points
0 comments12 min readEA link
(www.youtube.com)

[Question] Has An­thropic already made the ex­ter­nally leg­ible com­mit­ments that it planned to make?

Ofer12 Mar 2024 13:45 UTC
21 points
3 comments1 min readEA link

Pod­cast (+tran­script): Nathan Barnard on how US fi­nan­cial reg­u­la­tion can in­form AI governance

Aaron Bergman8 Aug 2023 21:46 UTC
12 points
0 comments23 min readEA link
(www.aaronbergman.net)

Cog­ni­tive as­sets and defen­sive acceleration

JulianHazell3 Apr 2024 14:55 UTC
13 points
3 comments4 min readEA link
(muddyclothes.substack.com)

In­ves­ti­gat­ing an in­surance-for-AI startup

L Rudolf L21 Sep 2024 15:29 UTC
40 points
1 comment15 min readEA link
(www.strataoftheworld.com)

[Question] Would an An­thropic/​OpenAI merger be good for AI safety?

M22 Nov 2023 20:21 UTC
6 points
1 comment1 min readEA link

Law-Fol­low­ing AI 1: Se­quence In­tro­duc­tion and Structure

Cullen 🔸27 Apr 2022 17:16 UTC
35 points
2 comments9 min readEA link

My ar­ti­cle in The Na­tion — Cal­ifor­nia’s AI Safety Bill Is a Mask-Off Mo­ment for the Industry

Garrison15 Aug 2024 19:25 UTC
134 points
0 comments1 min readEA link
(www.thenation.com)

Lo­cal De­tours On A Nar­row Path: How might treaties fail in China?

Jack_S🔸11 Aug 2025 20:33 UTC
9 points
0 comments14 min readEA link
(torchestogether.substack.com)

Pal­isade is hiring: Exec As­sis­tant, Con­tent Lead, Ops Lead, and Policy Lead

Charlie Rogers-Smith9 Oct 2024 0:04 UTC
15 points
2 comments4 min readEA link

[Question] What are the best jour­nals to pub­lish AI gov­er­nance pa­pers in?

Caro2 May 2022 10:07 UTC
26 points
4 comments1 min readEA link

Is it time for a pause?

Kelsey Piper6 Apr 2023 11:48 UTC
103 points
5 comments5 min readEA link

A Wind­fall Clause for CEO could worsen AI race dynamics

Larks9 Mar 2023 18:02 UTC
70 points
10 comments7 min readEA link

Bernie San­ders (I-VT) men­tions AI loss of con­trol risk in Giz­modo interview

Matrice Jacobine🔸🏳️‍⚧️14 Jul 2025 14:47 UTC
26 points
0 comments1 min readEA link
(gizmodo.com)

Sha­har Avin: Near-term AI se­cu­rity risks, and what to do about them

EA Global3 Nov 2017 7:43 UTC
7 points
0 comments1 min readEA link
(www.youtube.com)

Some tal­ent needs in AI governance

Sam Clarke13 Jun 2023 13:53 UTC
133 points
10 comments8 min readEA link

[Question] What would you do if you had a lot of money/​power/​in­fluence and you thought that AI timelines were very short?

Greg_Colbourn ⏸️ 12 Nov 2021 21:59 UTC
29 points
8 comments1 min readEA link

Want to do US tech­nol­ogy policy? Ap­ply to be a Hori­zon Fel­low!

Abby Babby29 Jul 2025 16:13 UTC
28 points
0 comments1 min readEA link

Good government

rosehadshar10 Sep 2025 13:22 UTC
64 points
1 comment6 min readEA link

Baobao Zhang: How so­cial sci­ence re­search can in­form AI governance

EA Global22 Jan 2021 15:10 UTC
9 points
0 comments16 min readEA link
(www.youtube.com)

Col­lec­tion of work on ‘Should you fo­cus on the EU if you’re in­ter­ested in AI gov­er­nance for longter­mist/​x-risk rea­sons?’

MichaelA🔸6 Aug 2022 16:49 UTC
51 points
3 comments1 min readEA link

AI gov­er­nance stu­dent hackathon on Satur­day, April 23: reg­ister now!

mic12 Apr 2022 4:39 UTC
18 points
0 comments1 min readEA link

Cal­ifor­ni­ans, tell your reps to vote yes on SB 1047!

Holly Elmore ⏸️ 🔸12 Aug 2024 19:49 UTC
106 points
6 comments1 min readEA link

Fore­sight for AGI Safety Strat­egy: Miti­gat­ing Risks and Iden­ti­fy­ing Golden Opportunities

jacquesthibs5 Dec 2022 16:09 UTC
14 points
1 comment8 min readEA link

[Linkpost] OpenAI is award­ing ten 100k grants for build­ing pro­to­types of a demo­cratic pro­cess for steer­ing AI

pseudonym26 May 2023 12:49 UTC
36 points
2 comments1 min readEA link
(openai.com)

P(doom|AGI) is high: why the de­fault out­come of AGI is doom

Greg_Colbourn ⏸️ 2 May 2023 10:40 UTC
15 points
28 comments3 min readEA link

Pause For Thought: The AI Pause Debate

Scott Alexander10 Oct 2023 15:34 UTC
113 points
20 comments14 min readEA link
(www.astralcodexten.com)

UK’s new 10-year “Na­tional AI Strat­egy,” re­leased today

jared_m22 Sep 2021 11:18 UTC
28 points
7 comments1 min readEA link

Re­lease of UN’s draft re­lated to the gov­er­nance of AI (a sum­mary of the Si­mon In­sti­tute’s re­sponse)

SebastianSchmidt27 Apr 2024 18:27 UTC
22 points
0 comments1 min readEA link

How ma­jor gov­ern­ments can help with the most im­por­tant century

Holden Karnofsky24 Feb 2023 19:37 UTC
56 points
4 comments4 min readEA link
(www.cold-takes.com)

Some AI re­search ar­eas and their rele­vance to ex­is­ten­tial safety

Andrew Critch15 Dec 2020 12:15 UTC
12 points
1 comment56 min readEA link
(alignmentforum.org)

[Link Post: New York Times] White House Un­veils Ini­ti­a­tives to Re­duce Risks of A.I.

Rockwell4 May 2023 14:04 UTC
50 points
1 comment2 min readEA link

Nvidia Comes Out Swing­ing as Congress Weighs Limits on China Chip Sales

Matrice Jacobine🔸🏳️‍⚧️10 Sep 2025 6:52 UTC
5 points
0 comments1 min readEA link
(www.nytimes.com)

Hooray for step­ping out of the limelight

So8res1 Apr 2023 2:45 UTC
103 points
0 comments1 min readEA link

Deep­Mind is hiring Long-term Strat­egy & Gover­nance researchers

vishal13 Sep 2021 18:44 UTC
54 points
1 comment1 min readEA link

Non-al­ign­ment pro­ject ideas for mak­ing trans­for­ma­tive AI go well

Lukas Finnveden4 Jan 2024 7:23 UTC
66 points
1 comment3 min readEA link
(www.forethought.org)

Jeffrey Ding: Re-de­ci­pher­ing China’s AI dream

EA Global18 Oct 2019 18:05 UTC
13 points
0 comments1 min readEA link
(www.youtube.com)

[Question] AI policy 501(c)(4) recom­men­da­tions for in­di­vi­d­ual donors?

Eevee🔹1 Oct 2025 4:45 UTC
26 points
4 comments1 min readEA link

Pile of Law and Law-Fol­low­ing AI

Cullen 🔸13 Jul 2022 0:29 UTC
28 points
2 comments3 min readEA link

Some thoughts on risks from nar­row, non-agen­tic AI

richard_ngo19 Jan 2021 0:07 UTC
36 points
2 comments8 min readEA link

[Link] EAF Re­search agenda: “Co­op­er­a­tion, Con­flict, and Trans­for­ma­tive Ar­tifi­cial In­tel­li­gence”

stefan.torges17 Jan 2020 13:28 UTC
64 points
0 comments1 min readEA link

RP’s AI Gover­nance & Strat­egy team—June 2023 in­terim overview

MichaelA🔸22 Jun 2023 13:45 UTC
68 points
1 comment7 min readEA link

The­o­ries of Change for Track II Di­plo­macy [Founders Pledge]

christian.r9 Jul 2024 13:31 UTC
21 points
2 comments33 min readEA link

Po­ten­tial Risks from Ad­vanced AI

EA Global13 Aug 2017 7:00 UTC
9 points
0 comments18 min readEA link

Ngo and Yud­kowsky on AI ca­pa­bil­ity gains

richard_ngo19 Nov 2021 1:54 UTC
23 points
4 comments39 min readEA link

Lead, Own, Share: Sovereign Wealth Funds for Trans­for­ma­tive AI

Matrice Jacobine🔸🏳️‍⚧️14 Jul 2025 9:34 UTC
24 points
0 comments1 min readEA link
(www.convergenceanalysis.org)

In­sights from an ex­pert sur­vey about in­ter­me­di­ate goals in AI governance

Sebastian Schwiecker17 Mar 2023 14:59 UTC
11 points
2 comments1 min readEA link

What does (and doesn’t) AI mean for effec­tive al­tru­ism?

EA Global12 Aug 2017 7:00 UTC
9 points
0 comments12 min readEA link

AI & Policy 1/​3: On know­ing the effect of to­day’s poli­cies on Trans­for­ma­tive AI risks, and the case for in­sti­tu­tional im­prove­ments.

weeatquince27 Aug 2019 11:04 UTC
27 points
3 comments10 min readEA link

The AI rev­olu­tion and in­ter­na­tional poli­tics (Allan Dafoe)

EA Global2 Jun 2017 8:48 UTC
8 points
0 comments18 min readEA link
(www.youtube.com)

Com­pute & An­titrust: Reg­u­la­tory im­pli­ca­tions of the AI hard­ware sup­ply chain, from chip de­sign to cloud APIs

HaydnBelfield19 Aug 2022 17:20 UTC
32 points
0 comments6 min readEA link
(verfassungsblog.de)

Sym­bio­sis, not al­ign­ment, as the goal for liberal democ­ra­cies in the tran­si­tion to ar­tifi­cial gen­eral intelligence

simonfriederich17 Mar 2023 13:04 UTC
18 points
2 comments24 min readEA link
(rdcu.be)

AI Safety Newslet­ter #5: Ge­offrey Hin­ton speaks out on AI risk, the White House meets with AI labs, and Tro­jan at­tacks on lan­guage models

Center for AI Safety9 May 2023 15:26 UTC
60 points
0 comments4 min readEA link
(newsletter.safe.ai)

US Congress in­tro­duces CREATE AI Act for es­tab­lish­ing Na­tional AI Re­search Resource

Daniel_Eth28 Jul 2023 23:27 UTC
9 points
1 comment1 min readEA link
(eshoo.house.gov)

Mauhn Re­leases AI Safety Documentation

Berg Severens2 Jul 2021 12:19 UTC
4 points
2 comments1 min readEA link

AI policy ca­reers in the EU

Lauro Langosco11 Nov 2019 10:43 UTC
62 points
7 comments11 min readEA link

[Question] AI strat­egy ca­reer pipeline

Zach Stein-Perlman22 May 2023 0:00 UTC
72 points
23 comments1 min readEA link

[Question] What are some crit­i­cisms of PauseAI?

Eevee🔹23 Nov 2024 17:49 UTC
53 points
71 comments1 min readEA link

The new UK gov­ern­ment’s stance on AI safety

Elliot Mckernon31 Jul 2024 15:23 UTC
19 points
0 comments4 min readEA link

Sum­mary of Si­tu­a­tional Aware­ness—The Decade Ahead

OscarD🔸8 Jun 2024 11:29 UTC
143 points
5 comments18 min readEA link

Read­ing list on AI agents and as­so­ci­ated policy

Peter Wildeford9 Aug 2024 17:40 UTC
79 points
2 comments1 min readEA link

[Question] How com­mit­ted to AGI safety are the cur­rent OpenAI non­profit board mem­bers?

Eevee🔹2 Dec 2024 4:03 UTC
14 points
1 comment1 min readEA link

Notes on nukes, IR, and AI from “Arse­nals of Folly” (and other books)

tlevin4 Sep 2023 19:02 UTC
21 points
2 comments6 min readEA link

AI welfare vs. AI rights

Matthew_Barnett4 Feb 2025 18:28 UTC
37 points
20 comments3 min readEA link

The U.S. and China Need an AI In­ci­dents Hotline

christian.r3 Jun 2024 18:46 UTC
25 points
0 comments1 min readEA link
(www.lawfaremedia.org)

Ap­ply to be a men­tor in SPAR!

Agustín Covarrubias 🔸24 Jun 2025 23:00 UTC
25 points
0 comments1 min readEA link

[Job ad] Re­search im­por­tant longter­mist top­ics at Re­think Pri­ori­ties!

Linch6 Oct 2021 19:09 UTC
65 points
46 comments1 min readEA link

[Question] Is there ev­i­dence that recom­mender sys­tems are chang­ing users’ prefer­ences?

zdgroff12 Apr 2021 19:11 UTC
60 points
15 comments1 min readEA link

[Question] How to Im­prove China-Western Co­or­di­na­tion on EA Is­sues?

Michael Kehoe3 Nov 2021 7:28 UTC
15 points
2 comments1 min readEA link

Linkpost: 7 A.I. Com­pa­nies Agree to Safe­guards After Pres­sure From the White House

MHR🔸21 Jul 2023 13:23 UTC
61 points
4 comments1 min readEA link
(www.nytimes.com)

Dear An­thropic peo­ple, please don’t re­lease Claude

Joseph Miller8 Feb 2023 2:44 UTC
28 points
5 comments1 min readEA link

AMA: Fu­ture of Life In­sti­tute’s EU Team

Risto Uuk31 Jan 2022 17:14 UTC
44 points
15 comments2 min readEA link

Ap­ply to the new Open Philan­thropy Tech­nol­ogy Policy Fel­low­ship!

lukeprog20 Jul 2021 18:41 UTC
78 points
6 comments4 min readEA link

Spicy takes about AI policy (Clark, 2022)

Will Aldred9 Aug 2022 13:49 UTC
44 points
0 comments3 min readEA link
(twitter.com)

[Question] Track­ing Com­pute Stocks and Flows: Case Stud­ies?

Cullen 🔸5 Oct 2022 17:54 UTC
34 points
1 comment1 min readEA link

Google in­vests $300mn in ar­tifi­cial in­tel­li­gence start-up An­thropic | FT

𝕮𝖎𝖓𝖊𝖗𝖆3 Feb 2023 19:43 UTC
155 points
5 comments1 min readEA link
(www.ft.com)

Ap­ply to fall policy in­tern­ships (we can help)

ES2 Jul 2023 21:37 UTC
57 points
4 comments1 min readEA link

How CISA can Sup­port the Se­cu­rity of Large AI Models Against Theft [Grad School As­sign­ment]

Marcel23 May 2023 15:36 UTC
7 points
0 comments13 min readEA link

How Could AI Gover­nance Go Wrong?

HaydnBelfield26 May 2022 21:29 UTC
40 points
7 comments18 min readEA link

Giv­ing away copies of Un­con­trol­lable by Dar­ren McKee

Greg_Colbourn ⏸️ 14 Dec 2023 17:00 UTC
39 points
2 comments1 min readEA link

Suc­ces­sif: Join our AI pro­gram to help miti­gate the catas­trophic risks of AI

ClaireB25 Oct 2023 16:51 UTC
15 points
0 comments5 min readEA link

[Question] Will AGI cause mass tech­nolog­i­cal un­em­ploy­ment?

Eevee🔹22 Jun 2020 20:55 UTC
4 points
2 comments2 min readEA link

Race to the Top: Bench­marks for AI Safety

isaduan4 Dec 2022 22:50 UTC
52 points
8 comments1 min readEA link

Should there be just one west­ern AGI pro­ject?

rosehadshar4 Dec 2024 14:41 UTC
49 points
3 comments15 min readEA link
(www.forethought.org)

Assess­ing the state of AI R&D in the US, China, and Europe – Part 1: Out­put indicators

stefan.torges1 Nov 2019 14:41 UTC
21 points
0 comments14 min readEA link

Could Reg­u­la­tory Cost-Benefit Anal­y­sis Stop Fron­tier AI Reg­u­la­tions in the US?

Luise11 Jul 2024 15:25 UTC
23 points
1 comment14 min readEA link

List #1: Why stop­ping the de­vel­op­ment of AGI is hard but doable

Remmelt24 Dec 2022 9:52 UTC
24 points
2 comments5 min readEA link

Shar­ing the AI Wind­fall: A Strate­gic Ap­proach to In­ter­na­tional Benefit-Sharing

michel16 Aug 2024 12:54 UTC
67 points
0 comments13 min readEA link
(wrtaigovernance.substack.com)

Con­crete ac­tions to im­prove AI gov­er­nance: the be­havi­our sci­ence approach

Alexander Saeri1 Dec 2022 21:34 UTC
31 points
0 comments11 min readEA link

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

Jan_Kulveit22 Feb 2023 16:09 UTC
68 points
1 comment6 min readEA link

Seek­ing Mechanism De­signer for Re­search into In­ter­nal­iz­ing Catas­trophic Externalities

c.trout11 Sep 2024 15:09 UTC
11 points
0 comments3 min readEA link

Hu­man­i­ties Re­search Ideas for Longtermists

Lizka9 Jun 2021 4:39 UTC
151 points
13 comments13 min readEA link

[Question] What are the challenges and prob­lems with pro­gram­ming law-break­ing con­straints into AGI?

Michael St Jules 🔸2 Feb 2020 20:53 UTC
20 points
34 comments1 min readEA link

Nav­i­gat­ing AI Risks (NAIR) #1: Slow­ing Down AI

simeon_c14 Apr 2023 14:35 UTC
12 points
1 comment1 min readEA link
(navigatingairisks.substack.com)

AI Risk & Policy Fore­casts from Me­tac­u­lus & FLI’s AI Path­ways Workshop

Will Aldred16 May 2023 8:53 UTC
41 points
0 comments8 min readEA link

[Question] Do­ing Global Pri­ori­ties or AI Policy re­search from re­mote lo­ca­tion?

With Love from Israel29 Oct 2019 9:34 UTC
30 points
4 comments1 min readEA link

Poli­ti­cal Fund­ing Ex­per­tise (Post 6 of 7 on AI Gover­nance)

Jason Green-Lowe19 Jun 2025 14:14 UTC
37 points
1 comment14 min readEA link

AI gov­er­nance and strat­egy: a list of re­search agen­das and work that could be done.

Nathan_Barnard12 Mar 2024 11:21 UTC
33 points
4 comments17 min readEA link

Rol­ling Thresh­olds for AGI Scal­ing Regulation

Larks12 Jan 2025 1:30 UTC
60 points
4 comments6 min readEA link

AI Gover­nance Needs Tech­ni­cal Work

Mau5 Sep 2022 22:25 UTC
121 points
3 comments8 min readEA link

Sup­ple­ment to “The Brus­sels Effect and AI: How EU AI reg­u­la­tion will im­pact the global AI mar­ket”

MarkusAnderljung16 Aug 2022 20:55 UTC
109 points
7 comments8 min readEA link

AI Gover­nance Read­ing Group Guide

AHT25 Jun 2020 10:16 UTC
26 points
2 comments3 min readEA link

Longter­mist rea­sons to work for in­no­va­tive governments

ac13 Oct 2020 16:32 UTC
74 points
8 comments1 min readEA link

How to make the best of the most im­por­tant cen­tury?

Holden Karnofsky14 Sep 2021 21:05 UTC
57 points
5 comments12 min readEA link

FLI launches Wor­ld­build­ing Con­test with $100,000 in prizes

ggilgallon17 Jan 2022 13:54 UTC
87 points
55 comments6 min readEA link

UK policy and poli­tics careers

weeatquince28 Sep 2019 16:18 UTC
28 points
10 comments7 min readEA link

AGI Safety Fun­da­men­tals cur­ricu­lum and application

richard_ngo20 Oct 2021 21:45 UTC
123 points
20 comments8 min readEA link
(docs.google.com)

UN Sec­re­tary-Gen­eral recog­nises ex­is­ten­tial threat from AI

Greg_Colbourn ⏸️ 15 Jun 2023 17:03 UTC
58 points
1 comment1 min readEA link

Up­dates from Cam­paign for AI Safety

Jolyn Khoo27 Sep 2023 2:44 UTC
16 points
0 comments2 min readEA link
(www.campaignforaisafety.org)

Ap­ti­tudes for AI gov­er­nance work

Sam Clarke13 Jun 2023 13:54 UTC
68 points
0 comments7 min readEA link

Grad­ual Disem­pow­er­ment: Sys­temic Ex­is­ten­tial Risks from In­cre­men­tal AI Development

Jan_Kulveit30 Jan 2025 17:07 UTC
39 points
4 comments2 min readEA link
(gradual-disempowerment.ai)

Tar­bell Fel­low­ship 2025 - Ap­pli­ca­tions Open (AI Jour­nal­ism)

Tarbell Center for AI Journalism8 Jan 2025 15:25 UTC
62 points
0 comments1 min readEA link

Re­duc­ing global AI com­pe­ti­tion through the Com­merce Con­trol List and Im­mi­gra­tion re­form: a dual-pronged approach

ben.smith3 Sep 2024 5:28 UTC
15 points
0 comments9 min readEA link

Catas­trophic Risks from Un­safe AI: Nav­i­gat­ing a Tightrope Sce­nario (Ben Garfinkel, EAG Lon­don 2023)

Alexander Saeri2 Jun 2023 9:59 UTC
19 points
1 comment10 min readEA link

How Europe might mat­ter for AI governance

stefan.torges12 Jul 2019 23:42 UTC
52 points
13 comments8 min readEA link

AI com­pa­nies are un­likely to make high-as­surance safety cases if timelines are short

Ryan Greenblatt23 Jan 2025 18:41 UTC
45 points
1 comment13 min readEA link

Idea to boost in­ter­na­tional AI coordination

Jamie Green13 Aug 2025 13:40 UTC
2 points
0 comments3 min readEA link

A Primer on God, Liber­al­ism and the End of History

Mahdi Complex28 Mar 2022 5:26 UTC
8 points
3 comments14 min readEA link

In­tel­sat as a Model for In­ter­na­tional AGI Governance

rosehadshar13 Mar 2025 12:58 UTC
42 points
3 comments1 min readEA link
(www.forethought.org)

AISN #46: The Transition

Center for AI Safety23 Jan 2025 18:01 UTC
10 points
0 comments5 min readEA link
(newsletter.safe.ai)

AGI will be made of het­ero­ge­neous com­po­nents, Trans­former and Selec­tive SSM blocks will be among them

Roman Leventov27 Dec 2023 14:51 UTC
5 points
0 comments4 min readEA link

A se­lec­tion of les­sons from Se­bas­tian Lodemann

ClaireB11 Nov 2024 21:33 UTC
82 points
2 comments7 min readEA link

A per­sonal take on why (and why not) to work on AI safety at Open Philanthropy

cb7 Nov 2025 10:47 UTC
92 points
7 comments11 min readEA link

Com­pute Re­search Ques­tions and Met­rics—Trans­for­ma­tive AI and Com­pute [4/​4]

lennart28 Nov 2021 22:18 UTC
18 points
2 comments1 min readEA link

Com­par­ing AI Labs and Phar­ma­ceu­ti­cal Companies

mxschons13 Nov 2024 14:51 UTC
13 points
0 comments1 min readEA link
(mxschons.com)

In­tent al­ign­ment should not be the goal for AGI x-risk reduction

johnjnay26 Oct 2022 1:24 UTC
7 points
1 comment2 min readEA link

Fun­da­men­tal Challenges in AI Governance

Tharin23 Oct 2023 1:30 UTC
7 points
1 comment7 min readEA link

Stop call­ing them labs

sawyer🔸24 Feb 2025 22:58 UTC
263 points
22 comments1 min readEA link

Tar­bell Fel­low­ship 2024 - Ap­pli­ca­tions Open (AI Jour­nal­ism)

Cillian_28 Sep 2023 10:38 UTC
58 points
1 comment3 min readEA link

Un-un­plug­ga­bil­ity—can’t we just un­plug it?

Oliver Sourbut15 May 2023 13:23 UTC
15 points
0 comments10 min readEA link
(www.oliversourbut.net)

AI Risk: Can We Thread the Nee­dle? [Recorded Talk from EA Sum­mit Van­cou­ver ’25]

Evan R. Murphy2 Oct 2025 19:05 UTC
8 points
0 comments2 min readEA link

Main­stream Grant­mak­ing Ex­per­tise (Post 7 of 7 on AI Gover­nance)

Jason Green-Lowe23 Jun 2025 1:38 UTC
53 points
2 comments37 min readEA link

In­tro­duc­ing In­ter­na­tional AI Gover­nance Alli­ance (IAIGA)

James Norris5 Feb 2025 15:59 UTC
12 points
0 comments1 min readEA link

Scal­ing Wargam­ing for Global Catas­trophic Risks with AI

rai18 Jan 2025 15:07 UTC
73 points
1 comment4 min readEA link
(blog.sentinel-team.org)

[Linkpost] “Blueprint for an AI Bill of Rights”—Office of Science and Tech­nol­ogy Policy, USA (2022)

QubitSwarm995 Oct 2022 16:48 UTC
15 points
0 comments2 min readEA link
(www.whitehouse.gov)

Me­tacrisis as a Frame­work for AI Governance

Jonah Wilberg22 Sep 2025 14:17 UTC
35 points
2 comments8 min readEA link

The AI bub­ble cov­ered in the Atlantic

Remmelt11 Nov 2025 4:12 UTC
13 points
1 comment2 min readEA link
(www.theatlantic.com)

Pes­simism about AI Safety

Max_He-Ho2 Apr 2023 7:57 UTC
5 points
0 comments25 min readEA link
(www.lesswrong.com)

The His­tory, Episte­mol­ogy and Strat­egy of Tech­nolog­i­cal Res­traint, and les­sons for AI (short es­say)

MMMaas10 Aug 2022 11:00 UTC
90 points
6 comments9 min readEA link
(verfassungsblog.de)

Pulse 2024: At­ti­tudes to­wards ar­tifi­cial intelligence

Jamie E27 Nov 2024 11:33 UTC
62 points
4 comments3 min readEA link

New Re­port: Multi-Agent Risks from Ad­vanced AI

Lewis Hammond23 Feb 2025 0:32 UTC
40 points
3 comments2 min readEA link
(www.cooperativeai.com)

An­nounc­ing #AISum­mitTalks fea­tur­ing Pro­fes­sor Stu­art Rus­sell and many others

Otto24 Oct 2023 10:16 UTC
9 points
1 comment1 min readEA link

His­tor­i­cal Prece­dents for In­ter­na­tional AI Safety Collaborations

ZacRichardson13 Jul 2025 21:30 UTC
20 points
1 comment55 min readEA link

AI-Rele­vant Reg­u­la­tion: CPSC

SWK13 Aug 2023 15:44 UTC
3 points
0 comments6 min readEA link

Ap­pli­ca­tions Now Open for Deep Dive: A 201 AI Policy Course by ENAIS

Kambar2 Jul 2025 8:32 UTC
10 points
5 comments1 min readEA link

Enough about AI timelines— we already know what we need to know.

Holly Elmore ⏸️ 🔸9 Apr 2025 10:29 UTC
138 points
35 comments2 min readEA link

China pro­poses new global AI co­op­er­a­tion organisation

Matrice Jacobine🔸🏳️‍⚧️30 Jul 2025 2:50 UTC
15 points
1 comment1 min readEA link
(www.reuters.com)

Stan­ford sum­mer course: Eco­nomics of Trans­for­ma­tive AI

trammell23 Jan 2025 23:07 UTC
83 points
4 comments1 min readEA link

‘Ar­tifi­cial In­tel­li­gence Gover­nance un­der Change’ (PhD dis­ser­ta­tion)

MMMaas15 Sep 2022 12:10 UTC
54 points
1 comment2 min readEA link
(drive.google.com)

Why We Need a Bea­con of Hope in the Loom­ing Gloom of AGI

Beyond Singularity2 Apr 2025 14:22 UTC
4 points
6 comments5 min readEA link

Try o3-pro in ChatGPT for $1 (is AI a bub­ble?)

Hauke Hillebrandt24 Jun 2025 11:15 UTC
29 points
1 comment4 min readEA link

Wor­ri­some Trends for Digi­tal Mind Evaluations

Derek Shiller20 Feb 2025 15:35 UTC
79 points
10 comments8 min readEA link

ChatGPT bug leaked users’ con­ver­sa­tion histories

Ian Turner27 Mar 2023 0:17 UTC
15 points
2 comments1 min readEA link
(www.bbc.com)

Ex-OpenAI em­ployee am­ici leave to file de­nied in Musk v OpenAI case?

TFD2 May 2025 12:31 UTC
8 points
0 comments2 min readEA link
(www.thefloatingdroid.com)

The Inequal­ity We Might Want: Merit-Based Redis­tri­bu­tion for the AI Transition

Andrei Navrotskii27 Nov 2025 10:51 UTC
5 points
0 comments12 min readEA link

Ret­ro­spec­tive Case Study of Com­mu­nity Build­ing Land­scape in China | Run­ning Arete Fel­low­ship in NYU Shanghai

arghyas9 Jul 2024 18:35 UTC
18 points
1 comment12 min readEA link

Quick Thoughts on A.I. Governance

Nicholas Kross30 Apr 2022 14:49 UTC
43 points
0 comments2 min readEA link
(www.thinkingmuchbetter.com)

My work­ing group on the best dona­tion opportunities

Eric Neyman21 Nov 2025 2:07 UTC
55 points
4 comments2 min readEA link

Panel Dis­cus­sion on the Gover­nance of Fron­tier AI

Alicia Pollard12 Nov 2025 18:07 UTC
2 points
0 comments1 min readEA link

2024 CFP for APSA, Largest An­nual Meet­ing of Poli­ti­cal Science

nemeryxu3 Jan 2024 19:43 UTC
2 points
0 comments1 min readEA link

Birds, Brains, Planes, and AI: Against Ap­peals to the Com­plex­ity/​Mys­te­ri­ous­ness/​Effi­ciency of the Brain

kokotajlod18 Jan 2021 12:39 UTC
27 points
2 comments1 min readEA link

The Pug­wash Con­fer­ences and the Anti-Bal­lis­tic Mis­sile Treaty as a case study of Track II diplomacy

rani_martin16 Sep 2022 10:42 UTC
82 points
5 comments27 min readEA link

What is ev­ery­one do­ing in AI governance

Igor Ivanov8 Jul 2023 15:19 UTC
31 points
0 comments5 min readEA link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

EliezerYudkowsky9 Apr 2023 15:53 UTC
50 points
3 comments12 min readEA link

ALTER Is­rael End-of-2024 Update

Davidmanheim7 Jan 2025 15:07 UTC
38 points
1 comment4 min readEA link

Con­scious AI: Will we know it when we see it? [Con­scious AI & Public Per­cep­tion]

ixex4 Jul 2024 20:30 UTC
13 points
1 comment12 min readEA link

The 6D effect: When com­pa­nies take risks, one email can be very pow­er­ful.

stecas4 Nov 2023 20:08 UTC
40 points
1 comment3 min readEA link

Re­duc­ing profit mo­ti­va­tions in AI development

Luke Frymire3 Apr 2023 20:04 UTC
20 points
1 comment6 min readEA link

The AGI-Proof Mind: Se­cur­ing Cog­ni­tive Pri­vacy via the Cog­ni­tive Fortress Man­date (CFM)

T. Johnson25 Nov 2025 16:50 UTC
−1 points
0 comments4 min readEA link

Tips for con­duct­ing wor­ld­view investigations

lukeprog12 Apr 2022 19:28 UTC
88 points
4 comments2 min readEA link

Rac­ing through a minefield: the AI de­ploy­ment problem

Holden Karnofsky31 Dec 2022 21:44 UTC
79 points
1 comment13 min readEA link
(www.cold-takes.com)

The ne­ces­sity of “Guardian AI” and two con­di­tions for its achievement

Proica28 May 2024 11:42 UTC
1 point
1 comment15 min readEA link

Com­pute Gover­nance and Con­clu­sions—Trans­for­ma­tive AI and Com­pute [3/​4]

lennart14 Oct 2021 7:55 UTC
20 points
3 comments5 min readEA link

[Question] An eco­nomics of AI gov—best re­sources for

Liv26 Feb 2023 11:11 UTC
10 points
4 comments1 min readEA link

Three new re­ports re­view­ing re­search and con­cepts in ad­vanced AI governance

MMMaas28 Nov 2023 9:21 UTC
32 points
0 comments2 min readEA link
(www.legalpriorities.org)

Finish­ing The SB-1047 Documentary

Michaël Trazzi28 Oct 2024 20:26 UTC
67 points
0 comments4 min readEA link

MIT Fu­tureTech are hiring ‍a Product and Data Vi­su­al­iza­tion De­signer

PeterSlattery13 Nov 2024 14:41 UTC
9 points
0 comments4 min readEA link

[Linkpost] “Gover­nance of su­per­in­tel­li­gence” by OpenAI

Daniel_Eth22 May 2023 20:15 UTC
51 points
6 comments2 min readEA link
(openai.com)

The Short Timelines Strat­egy for AI Safety Univer­sity Groups

Josh Thorsteinson 🔸7 Mar 2025 4:26 UTC
52 points
8 comments5 min readEA link

Catas­tro­phe with­out Agency

ZenoSr20 Oct 2025 16:42 UTC
3 points
0 comments12 min readEA link

Prepar­ing De­spite Uncer­tainty: The Grand Challenges of AI Progress

Andrew Knott7 Nov 2025 10:42 UTC
7 points
0 comments7 min readEA link

A Fron­tier AI Risk Man­age­ment Frame­work: Bridg­ing the Gap Between Cur­rent AI Prac­tices and Estab­lished Risk Management

simeon_c13 Mar 2025 18:29 UTC
4 points
0 comments1 min readEA link
(arxiv.org)

An­thropic rewrote its RSP

Zach Stein-Perlman15 Oct 2024 14:30 UTC
32 points
1 comment6 min readEA link

Back­ground for “Un­der­stand­ing the diffu­sion of large lan­guage mod­els”

Ben Cottier21 Dec 2022 13:49 UTC
12 points
0 comments23 min readEA link

What we can learn from stress test­ing for AI regulation

Nathan_Barnard17 Jul 2023 19:56 UTC
27 points
0 comments26 min readEA link

What would it take for AI to dis­em­power us? Ryan Green­blatt on take­off dy­nam­ics, rogue de­ploy­ments, and al­ign­ment risks

80000_Hours8 Jul 2025 18:10 UTC
8 points
0 comments33 min readEA link

Assess­ing China’s im­por­tance as an AI superpower

JulianHazell3 Feb 2023 11:08 UTC
89 points
7 comments1 min readEA link
(muddyclothes.substack.com)

[Question] AI safety mile­stones?

Zach Stein-Perlman23 Jan 2023 21:30 UTC
6 points
0 comments1 min readEA link

Could this be an un­usu­ally good time to Earn To Give?

Tom Gardiner 🔸3 Mar 2025 23:00 UTC
60 points
15 comments3 min readEA link

AISN #57: The RAISE Act

Center for AI Safety17 Jun 2025 17:38 UTC
12 points
1 comment3 min readEA link
(newsletter.safe.ai)

In DC, a new wave of AI lob­by­ists gains the up­per hand

Chris Leong13 May 2024 7:31 UTC
97 points
7 comments1 min readEA link
(www.politico.com)

In­for­ma­tion in risky tech­nol­ogy races

nemeryxu2 Aug 2022 23:35 UTC
15 points
2 comments3 min readEA link

LawAI’s Sum­mer Re­search Fel­low­ship – ap­ply by Fe­bru­ary 16

LawAI7 Feb 2024 21:01 UTC
51 points
2 comments2 min readEA link

The be­hav­ioral se­lec­tion model for pre­dict­ing AI motivations

Alex Mallen4 Dec 2025 18:38 UTC
6 points
1 comment16 min readEA link

New AI Eco­nomics Brief—Is­sue #1

Jacob 🔸6 Nov 2025 9:04 UTC
9 points
0 comments1 min readEA link
(windfalltrust.substack.com)

SB 53 FAQs

Miles Kodama4 Aug 2025 8:15 UTC
13 points
1 comment8 min readEA link

Es­ti­mat­ing the Sub­sti­tutabil­ity be­tween Com­pute and Cog­ni­tive La­bor in AI Research

Parker_Whitfill1 Jun 2025 14:27 UTC
138 points
29 comments9 min readEA link

CSER is hiring for a se­nior re­search as­so­ci­ate on longterm AI risk and governance

Sam Clarke24 Jan 2022 13:24 UTC
9 points
4 comments1 min readEA link

Our A.I. Align­ment Im­per­a­tive: Creat­ing a Fu­ture Worth Sharing

Christopher Hunt Robertson, M.Ed.26 Oct 2025 20:46 UTC
1 point
0 comments21 min readEA link

What does it mean to be­come an ex­pert in AI Hard­ware?

Toph9 Jan 2021 4:15 UTC
87 points
10 comments11 min readEA link

Sam Alt­man and the Cross­roads of AI Power: Can We Trust the Fu­ture We’re Build­ing?

Kayode Adekoya23 May 2025 15:39 UTC
0 points
0 comments1 min readEA link

In­side the Bi­den ad­min’s AI Policy ap­proach | Jake Sul­li­van, Bi­den’s NSA | via The Cog­ni­tive Revolution

80000_Hours1 Oct 2025 13:56 UTC
6 points
1 comment1 min readEA link

Effec­tive pe­ti­tions (Novem­ber 2025)

Stijn Bruers 🔸9 Nov 2025 23:09 UTC
14 points
2 comments1 min readEA link

[Job]: AI Stan­dards Devel­op­ment Re­search Assistant

Tony Barrett14 Oct 2022 20:18 UTC
13 points
0 comments2 min readEA link

An­nounc­ing Apollo Research

mariushobbhahn30 May 2023 16:17 UTC
158 points
4 comments8 min readEA link

CAIDP State­ment on Lethal Au­tonomous Weapons Sys­tems

Heramb Podar30 Nov 2024 18:00 UTC
7 points
0 comments1 min readEA link
(www.linkedin.com)

Med­i­cal Wind­fall Prizes

PeterMcCluskey7 Feb 2025 0:13 UTC
5 points
0 comments5 min readEA link
(bayesianinvestor.com)

Mar­ius Hobb­hahn on the race to solve AI schem­ing be­fore mod­els go superhuman

80000_Hours3 Dec 2025 21:08 UTC
6 points
0 comments17 min readEA link

Con­sider not donat­ing un­der $100 to poli­ti­cal candidates

DanielFilan11 May 2025 3:22 UTC
86 points
12 comments1 min readEA link

Na­tion­wide Ac­tion Work­shop: Con­tact Congress about AI Safety!

Felix De Simone24 Feb 2025 16:14 UTC
5 points
0 comments1 min readEA link
(www.zeffy.com)

World and Mind in Ar­tifi­cial In­tel­li­gence: ar­gu­ments against the AI pause

Arturo Macias18 Apr 2023 14:35 UTC
6 points
3 comments5 min readEA link

How might we al­ign trans­for­ma­tive AI if it’s de­vel­oped very soon?

Holden Karnofsky29 Aug 2022 15:48 UTC
164 points
17 comments44 min readEA link

We’re not pre­pared for an AI mar­ket crash

Remmelt1 Apr 2025 4:33 UTC
28 points
4 comments2 min readEA link

How the AI Safety Com­mu­nity Can Counter Safety Washing

Chris Leong13 Oct 2025 8:27 UTC
9 points
3 comments18 min readEA link

AI com­pa­nies’ eval re­ports mostly don’t sup­port their claims

Zach Stein-Perlman9 Jun 2025 13:00 UTC
51 points
2 comments4 min readEA link

High­lights from Ilya Sutskever’s Novem­ber 2025 in­ter­view with Dwarkesh Patel

Yarrow Bouchard 🔸25 Nov 2025 22:52 UTC
20 points
2 comments5 min readEA link

What are some good books about AI safety?

Vishakha Agrawal17 Feb 2025 11:54 UTC
7 points
0 comments3 min readEA link
(aisafety.info)

Ex-OpenAI re­searcher says OpenAI mass-vi­o­lated copy­right law

Remmelt24 Oct 2024 1:00 UTC
11 points
0 comments1 min readEA link
(suchir.net)

Re­in­force­ment Learn­ing: A Non-Tech­ni­cal Primer on o1 and Deep­Seek-R1

AlexChalk9 Feb 2025 23:58 UTC
4 points
0 comments9 min readEA link
(alexchalk.net)

State­ment on Su­per­in­tel­li­gence—FLI Open Letter

plex22 Oct 2025 22:27 UTC
52 points
1 comment1 min readEA link
(superintelligence-statement.org)

Deep­Seek Made it Even Harder for US AI Com­pa­nies to Ever Reach Profitability

Garrison19 Feb 2025 21:02 UTC
30 points
1 comment3 min readEA link
(garrisonlovely.substack.com)

Ad­dress­ing the non­hu­man gap in in­ter­gov­ern­men­tal AI gov­er­nance frameworks

Alistair Stewart15 Jul 2025 21:13 UTC
73 points
2 comments8 min readEA link

Sur­vey on AI ex­is­ten­tial risk scenarios

Sam Clarke8 Jun 2021 17:12 UTC
159 points
11 comments6 min readEA link

AISN #66: Eval­u­at­ing Fron­tier Models, New Gem­ini and Claude, Preemp­tion is Back

Center for AI Safety2 Dec 2025 17:37 UTC
6 points
0 comments5 min readEA link
(newsletter.safe.ai)

How we use back-of-the-en­velope calcu­la­tions in our grantmaking

Coefficient Giving28 May 2025 23:22 UTC
79 points
2 comments10 min readEA link

When is it im­por­tant that open-weight mod­els aren’t re­leased? My thoughts on the benefits and dan­gers of open-weight mod­els in re­sponse to de­vel­op­ments in CBRN ca­pa­bil­ities.

Ryan Greenblatt9 Jun 2025 19:19 UTC
39 points
3 comments9 min readEA link

AI Safety Newslet­ter #7: Dis­in­for­ma­tion, Gover­nance Recom­men­da­tions for AI labs, and Se­nate Hear­ings on AI

Center for AI Safety23 May 2023 21:42 UTC
23 points
0 comments6 min readEA link
(newsletter.safe.ai)

Join the Vir­tual AI Safety Un­con­fer­ence (VAISU)!

Nguyên🔸21 Jun 2023 4:46 UTC
23 points
0 comments1 min readEA link
(vaisu.ai)

13 Re­cent Publi­ca­tions on Ex­is­ten­tial Risk (Jan 2021 up­date)

HaydnBelfield8 Feb 2021 12:42 UTC
7 points
2 comments10 min readEA link

AI Safety Field Growth Anal­y­sis 2025

Stephen McAleese27 Sep 2025 17:02 UTC
76 points
13 comments3 min readEA link

Genes did mis­al­ign­ment first: com­par­ing gra­di­ent hack­ing and meiotic drive

Holly Elmore ⏸️ 🔸18 Apr 2025 5:39 UTC
45 points
9 comments15 min readEA link
(hollyelmore.substack.com)

The world’s first fron­tier AI reg­u­la­tion is sur­pris­ingly thought­ful: the EU’s Code of Practice

Miles Kodama22 Sep 2025 15:22 UTC
20 points
1 comment15 min readEA link

The In­ter­na­tional PauseAI Protest: Ac­tivism un­der uncertainty

Joseph Miller12 Oct 2023 17:36 UTC
136 points
3 comments4 min readEA link

Har­den­ing against AI takeover is difficult, but we should try

Otto5 Nov 2025 16:29 UTC
8 points
1 comment5 min readEA link
(www.existentialriskobservatory.org)

Have your say on the fu­ture of AI reg­u­la­tion: Dead­line ap­proach­ing for your feed­back on UN High-Level Ad­vi­sory Body on AI In­terim Re­port ‘Govern­ing AI for Hu­man­ity’

Deborah W.A. Foulkes29 Mar 2024 6:37 UTC
17 points
1 comment1 min readEA link

Shut­ting down all com­pet­ing AI pro­jects might not buy a lot of time due to In­ter­nal Time Pressure

ThomasCederborg3 Oct 2024 0:05 UTC
6 points
1 comment12 min readEA link

My AI Vibes are Shifting

Nathan Young5 Sep 2025 14:45 UTC
15 points
2 comments4 min readEA link
(nathanpmyoung.substack.com)

US gov­ern­ment com­mis­sion pushes Man­hat­tan Pro­ject-style AI initiative

Larks19 Nov 2024 16:22 UTC
83 points
15 comments1 min readEA link
(www.reuters.com)

AI Policy In­sights from the AIMS Survey

Janet Pauketat22 Feb 2024 19:17 UTC
10 points
1 comment18 min readEA link
(www.sentienceinstitute.org)

A Cal­ifor­nia Effect for Ar­tifi­cial Intelligence

henryj9 Sep 2022 14:17 UTC
73 points
1 comment4 min readEA link
(docs.google.com)

Le­gal tem­plate for con­di­tional gift deed as an al­ter­na­tive to wa­gers on AI doom

bruce13 Mar 2025 14:57 UTC
30 points
6 comments1 min readEA link

Six Re­search Pit­falls and How to Avoid Them: a Guide for Re­search Managers

Morgan Simpson28 Jan 2025 9:49 UTC
19 points
0 comments10 min readEA link

AI Align­ment and the Fi­nan­cial War Against Nar­cis­sis­tic Manipulation

Julian Nalenz19 Feb 2025 20:36 UTC
2 points
0 comments3 min readEA link

The Prob­lem With the Word ‘Align­ment’

Peli Grietzer21 May 2024 21:37 UTC
13 points
1 comment6 min readEA link

6 (Po­ten­tial) Mis­con­cep­tions about AI Intellectuals

Ozzie Gooen14 Feb 2025 23:51 UTC
30 points
2 comments12 min readEA link

A map of work needed to achieve safe AI

Tristan Katz11 Sep 2025 11:33 UTC
16 points
0 comments1 min readEA link

Ap­pli­ca­tions for EU Tech Policy Fel­low­ship 2024 now open

Jan-Willem13 Sep 2023 16:17 UTC
22 points
2 comments1 min readEA link

Why AI Safety Needs a Cen­tral­ized Plan—And What It Might Look Like

Brandon Riggs28 May 2025 21:40 UTC
21 points
7 comments15 min readEA link

The AI Adop­tion Gap: Prepar­ing the US Govern­ment for Ad­vanced AI

Lizka2 Apr 2025 21:37 UTC
40 points
20 comments17 min readEA link
(www.forethought.org)

AISN #55: Trump Ad­minis­tra­tion Re­scinds AI Diffu­sion Rule, Allows Chip Sales to Gulf States

Center for AI Safety20 May 2025 16:05 UTC
7 points
0 comments4 min readEA link
(newsletter.safe.ai)

My re­flec­tions on do­ing a re­search fellowship

Yadav13 Jun 2025 10:41 UTC
11 points
2 comments5 min readEA link

How to pur­sue a ca­reer in AI gov­er­nance and coordination

Cody_Fenwick25 Sep 2023 12:00 UTC
32 points
1 comment29 min readEA link
(80000hours.org)

In­sti­tu­tions Can­not Res­train Dark-Triad AI Exploitation

Remmelt27 Dec 2022 10:34 UTC
8 points
0 comments5 min readEA link
(mflb.com)

AI Au­dit in Costa Rica

Priscilla Campos27 Jan 2025 2:57 UTC
10 points
4 comments9 min readEA link

AISN #64: New AGI Defi­ni­tion and Se­nate Bill Would Estab­lish Li­a­bil­ity for AI Harms

Center for AI Safety16 Oct 2025 17:50 UTC
8 points
0 comments5 min readEA link
(aisafety.substack.com)

Up­dates from Cam­paign for AI Safety

Jolyn Khoo31 Oct 2023 5:46 UTC
14 points
1 comment2 min readEA link
(www.campaignforaisafety.org)

A Tax­on­omy of Jobs Deeply Re­sis­tant to TAI Automation

Deric Cheng18 Mar 2025 16:26 UTC
40 points
1 comment12 min readEA link
(www.convergenceanalysis.org)

AGI Can­not Be Pre­dicted From Real In­ter­est Rates

Nicholas Decker28 Jan 2025 17:45 UTC
26 points
3 comments1 min readEA link
(nicholasdecker.substack.com)

Prefer­ence gaps as a safe­guard against AI self-replication

Bradford Saad26 Nov 2025 14:57 UTC
16 points
0 comments11 min readEA link

Pub­lished re­port: Path­ways to short TAI timelines

Zershaaneh Qureshi20 Feb 2025 22:10 UTC
47 points
2 comments17 min readEA link
(www.convergenceanalysis.org)

Eth­i­cal co-evolu­tion, or how to turn the main threat into a lev­er­age for longter­mism?

Beyond Singularity17 Sep 2025 17:24 UTC
7 points
7 comments8 min readEA link

Will the EU reg­u­la­tions on AI mat­ter to the rest of the world?

hanadulset1 Jan 2022 21:56 UTC
33 points
5 comments5 min readEA link

Align­ment Boot­strap­ping Is Dangerous

MichaelDickens27 Nov 2025 18:18 UTC
14 points
0 comments2 min readEA link

ALTER Is­rael Mid-2025 Semi­an­nual Update

Davidmanheim15 Jul 2025 7:47 UTC
13 points
1 comment5 min readEA link

Why I’m do­ing PauseAI

Joseph Miller30 Apr 2024 16:21 UTC
147 points
36 comments4 min readEA link

[Pod­cast] Ajeya Co­tra on wor­ld­view di­ver­sifi­ca­tion and how big the fu­ture could be

Eevee🔹22 Jan 2021 23:57 UTC
57 points
20 comments1 min readEA link
(80000hours.org)

Re­sults from the AI x Democ­racy Re­search Sprint

Esben Kran14 Jun 2024 16:40 UTC
19 points
1 comment6 min readEA link

God Coin: A Modest Pro­posal

Mahdi Complex1 Apr 2024 12:02 UTC
4 points
0 comments22 min readEA link

Lab Col­lab­o­ra­tion on AI Safety Best Prac­tices

amta17 Mar 2024 12:20 UTC
3 points
0 comments20 min readEA link

The AIA and its Brus­sels Effect

Kathryn O'Rourke27 Dec 2022 16:01 UTC
16 points
0 comments5 min readEA link

AI Safety Land­scape & Strate­gic Gaps

MichaelDickens17 Sep 2025 23:02 UTC
93 points
4 comments57 min readEA link

How to build AI you can ac­tu­ally Trust—Like a Med­i­cal Team, Not a Black Box

Ihor Ivliev22 Mar 2025 21:27 UTC
2 points
1 comment4 min readEA link

The Slip­pery Slope from DALLE-2 to Deep­fake Anarchy

stecas5 Nov 2022 14:47 UTC
55 points
11 comments17 min readEA link

On ex­clud­ing dan­ger­ous in­for­ma­tion from training

ShayBenMoshe17 Nov 2023 20:09 UTC
8 points
0 comments3 min readEA link
(www.lesswrong.com)

Com­ment on Bar­nett (2025): Growth effects of AI could hit a bot­tle­neck even if lo­cal elas­tic­i­ties are high

Dan Carey18 Mar 2025 11:52 UTC
92 points
0 comments9 min readEA link

A New Model for Com­pute Cen­ter Verification

Damin Curtis🔹10 Oct 2023 19:23 UTC
21 points
2 comments5 min readEA link

Sense-mak­ing about ex­treme power concentration

rosehadshar11 Sep 2025 10:09 UTC
35 points
0 comments4 min readEA link

How Apart Re­search would use marginal fund­ing to scale AI safety tal­ent development

JaimeRV23 Nov 2025 16:59 UTC
31 points
0 comments6 min readEA link

A Brief Sum­mary Of The Most Im­por­tant Century

Maynk0225 Oct 2022 15:28 UTC
3 points
0 comments5 min readEA link

Ap­ply to HAIST/​MAIA’s AI Gover­nance Work­shop in DC (Feb 17-20)

Phosphorous28 Jan 2023 0:45 UTC
15 points
0 comments1 min readEA link
(www.lesswrong.com)

The Man­hat­tan Trap: Why a Race to Ar­tifi­cial Su­per­in­tel­li­gence is Self-Defeating

Corin Katzke21 Jan 2025 16:57 UTC
98 points
1 comment2 min readEA link
(www.convergenceanalysis.org)

What to sug­gest com­pa­nies & en­trepreneurs do to use AI safely?

AlfalfaBloom5 Apr 2023 22:36 UTC
11 points
1 comment1 min readEA link

Why we need a new agency to reg­u­late ad­vanced ar­tifi­cial intelligence

Michael Huang4 Aug 2022 13:38 UTC
25 points
0 comments1 min readEA link
(www.brookings.edu)

What is scaf­fold­ing?

Vishakha Agrawal27 Mar 2025 9:40 UTC
3 points
0 comments2 min readEA link
(aisafety.info)

Ap­pli­ca­tions Open: GovAI Sum­mer Fel­low­ship 2023

GovAI21 Dec 2022 15:00 UTC
28 points
0 comments2 min readEA link

How the Hu­man Psy­cholog­i­cal “Pro­gram” Un­der­mines AI Align­ment — and What We Can Do

Beyond Singularity6 May 2025 13:37 UTC
14 points
2 comments3 min readEA link

GPT-3-like mod­els are now much eas­ier to ac­cess and de­ploy than to develop

Ben Cottier21 Dec 2022 13:49 UTC
22 points
3 comments19 min readEA link

Su­per Lenses + Mo­rally-Aimed Drives for A.I. Mo­ral Align­ment: Tech­ni­cal Framework

Christopher Hunt Robertson, M.Ed.16 Nov 2025 14:01 UTC
1 point
0 comments6 min readEA link

De-em­pha­sise al­ign­ment, em­pha­sise restraint

EuanMcLean4 Feb 2025 17:43 UTC
19 points
2 comments7 min readEA link

Con­sid­er­a­tions re­gard­ing be­ing nice to AIs

Matt Alexander18 Nov 2025 13:27 UTC
2 points
0 comments15 min readEA link
(www.lesswrong.com)

AI-Rele­vant Reg­u­la­tion: In­surance in Safety-Crit­i­cal Industries

SWK22 Jul 2023 17:52 UTC
5 points
0 comments6 min readEA link

AI, An­i­mals & Digi­tal Minds NYC 2025: Retrospective

Jonah Woodward31 Oct 2025 3:09 UTC
43 points
5 comments6 min readEA link

Re­search + Real­ity Graph­ing to Sup­port AI Policy (and more): Sum­mary of a Frozen Project

Marcel22 Jul 2022 20:58 UTC
34 points
2 comments8 min readEA link

Reg­u­la­tion of AI Use for Per­sonal Data Pro­tec­tion: Com­par­i­son of Global Strate­gies and Op­por­tu­ni­ties for Latin Amer­ica

Lisbeth Guzman 14 Oct 2024 13:22 UTC
10 points
1 comment21 min readEA link

AI could cause a drop in GDP, even if mar­kets are com­pet­i­tive and efficient

Casey Barkan17 Apr 2025 14:37 UTC
18 points
7 comments5 min readEA link
(www.lesswrong.com)

In­tro­duc­ing the AI Ob­jec­tives In­sti­tute’s Re­search: Differ­en­tial Paths to­ward Safe and Benefi­cial AI

cmck5 May 2023 20:26 UTC
43 points
1 comment8 min readEA link

FLI is hiring a new Direc­tor of US Policy

aaguirre27 Jul 2022 0:07 UTC
14 points
0 comments1 min readEA link

Carl Shul­man on AI takeover mechanisms (& more): Part II of Dwarkesh Pa­tel in­ter­view for The Lu­nar Society

alejandro25 Jul 2023 18:31 UTC
28 points
0 comments5 min readEA link
(www.dwarkeshpatel.com)

Eileen Yam on how we’re com­pletely out of touch with what the pub­lic thinks about AI

80000_Hours20 Nov 2025 17:37 UTC
5 points
1 comment22 min readEA link

Against rac­ing to AGI: Co­op­er­a­tion, de­ter­rence, and catas­trophic risks

Max_He-Ho29 Jul 2025 22:22 UTC
6 points
1 comment1 min readEA link
(philpapers.org)

Video and Tran­script of Pre­sen­ta­tion on Ex­is­ten­tial Risk from Power-Seek­ing AI

Joe_Carlsmith8 May 2022 3:52 UTC
97 points
7 comments30 min readEA link

Se­nate Strikes Po­ten­tial AI Mo­ra­to­rium

Tristan W1 Jul 2025 11:49 UTC
31 points
0 comments1 min readEA link
(www.reuters.com)

AI for Re­solv­ing Fore­cast­ing Ques­tions: An Early Exploration

Ozzie Gooen16 Jan 2025 21:40 UTC
22 points
0 comments9 min readEA link

The ‘Old AI’: Les­sons for AI gov­er­nance from early elec­tric­ity regulation

Sam Clarke19 Dec 2022 2:46 UTC
64 points
1 comment13 min readEA link

How Good­fire Is Turn­ing AI In­ter­pretabil­ity Into Real Products

Strad Slater30 Nov 2025 11:00 UTC
0 points
0 comments4 min readEA link
(williamslater2003.medium.com)

Sup­port Talos’ AI policy place­ments: The tal­ent pipeline for Euro­pean AI Governance

DavidConrad20 Nov 2025 12:17 UTC
40 points
1 comment4 min readEA link

How good would a CCP-dom­i­nated AI fu­ture be?

OscarD🔸22 Oct 2025 1:14 UTC
64 points
25 comments5 min readEA link
(oscardelaney.substack.com)

Gover­nance of AI, Break­fast Ce­real, Car Fac­to­ries, Etc.

Jeff Martin6 Nov 2023 1:44 UTC
2 points
0 comments3 min readEA link

[Link and com­men­tary] Beyond Near- and Long-Term: Towards a Clearer Ac­count of Re­search Pri­ori­ties in AI Ethics and Society

MichaelA🔸14 Mar 2020 9:04 UTC
18 points
0 comments6 min readEA link

The Game Board has been Flipped: Now is a good time to re­think what you’re doing

LintzA28 Jan 2025 21:20 UTC
391 points
69 comments13 min readEA link

15 Lev­ers to In­fluence Fron­tier AI Companies

Jan Wehner🔸26 Sep 2025 8:36 UTC
16 points
0 comments10 min readEA link

Anal­y­sis of Global AI Gover­nance Strategies

SammyDMartin11 Dec 2024 11:08 UTC
23 points
0 comments1 min readEA link
(www.lesswrong.com)

Un­der­stand­ing the diffu­sion of large lan­guage mod­els: summary

Ben Cottier21 Dec 2022 13:49 UTC
127 points
18 comments22 min readEA link

An ar­gu­ment for ac­cel­er­at­ing in­ter­na­tional AI gov­er­nance re­search (part 2)

MattThinks22 Aug 2023 22:40 UTC
5 points
0 comments10 min readEA link

AISN #49: Su­per­in­tel­li­gence Strategy

Center for AI Safety6 Mar 2025 17:43 UTC
8 points
0 comments5 min readEA link
(newsletter.safe.ai)

Book re­view: Ar­chi­tects of In­tel­li­gence by Martin Ford (2018)

Ofer11 Aug 2020 17:24 UTC
11 points
1 comment2 min readEA link

[Question] Im­pact: Eng­ineer­ing VS Med­i­cal Scien­tist VS AI Safety VS Governance

AhmedWez15 Jan 2025 15:47 UTC
1 point
0 comments1 min readEA link

AI Safety Eval­u­a­tions: A Reg­u­la­tory Review

Elliot Mckernon19 Mar 2024 15:09 UTC
12 points
2 comments11 min readEA link

How long till Brus­sels?: A light in­ves­ti­ga­tion into the Brus­sels Gap

Yadav26 Dec 2022 7:49 UTC
50 points
2 comments5 min readEA link

Les­sons from Three Mile Is­land for AI Warn­ing Shots

NickGabs26 Sep 2022 2:47 UTC
44 points
0 comments15 min readEA link

USA/​China Rec­on­cili­a­tion a Ne­ces­sity Be­cause of AI/​Tech Acceleration

bhrdwj🔸17 Apr 2025 13:13 UTC
1 point
7 comments7 min readEA link

OpenAI lost $5 billion in 2024 (and its losses are in­creas­ing)

Remmelt31 Mar 2025 4:17 UTC
0 points
3 comments12 min readEA link
(www.wheresyoured.at)

The cur­rent AI strate­gic land­scape: one bear’s perspective

Matrice Jacobine🔸🏳️‍⚧️15 Feb 2025 9:49 UTC
6 points
0 comments2 min readEA link
(philosophybear.substack.com)

In­tro­duc­ing LEAP: The Lon­gi­tu­di­nal Ex­pert AI Panel

Forecasting Research Institute10 Nov 2025 16:28 UTC
81 points
11 comments9 min readEA link
(forecastingresearch.substack.com)

AI Safety Strat­egy—A new or­ga­ni­za­tion for bet­ter timelines

Prometheus14 Jun 2023 20:41 UTC
8 points
0 comments2 min readEA link

China x AI Refer­ence List: Au­gust 2025 Update

Sarah Weiler9 Sep 2025 7:31 UTC
32 points
2 comments42 min readEA link
(docs.google.com)

AISN #36: Vol­un­tary Com­mit­ments are In­suffi­cient Plus, a Se­nate AI Policy Roadmap, and Chap­ter 1: An Overview of Catas­trophic Risks

Center for AI Safety30 May 2024 18:23 UTC
6 points
0 comments5 min readEA link
(newsletter.safe.ai)

MATS 8.0 Re­search Projects

Jonathan Michala8 Sep 2025 21:36 UTC
9 points
0 comments1 min readEA link
(substack.com)

Re­duc­ing LLM de­cep­tion at scale with self-other over­lap fine-tuning

Marc Carauleanu13 Mar 2025 19:09 UTC
8 points
0 comments6 min readEA link

Re­spon­si­ble Scal­ing Poli­cies Are Risk Man­age­ment Done Wrong

simeon_c25 Oct 2023 23:46 UTC
42 points
1 comment22 min readEA link
(www.navigatingrisks.ai)

Does gen­er­al­ity pay? GPT-3 can provide pre­limi­nary ev­i­dence.

Eevee🔹12 Jul 2020 18:53 UTC
21 points
4 comments2 min readEA link

Effec­tive al­tru­ism in the age of AGI

William_MacAskill10 Oct 2025 10:57 UTC
466 points
76 comments20 min readEA link

Publi­ca­tion de­ci­sions for large lan­guage mod­els, and their impacts

Ben Cottier21 Dec 2022 13:50 UTC
14 points
0 comments16 min readEA link

The UN Has a Rare Shot at Re­duc­ing the Risks of AI in War­fare

Mark Leon Goldberg21 May 2025 21:22 UTC
6 points
0 comments1 min readEA link

Microsoft Plans to In­vest $10B in OpenAI; $3B In­vested to Date | For­tune

𝕮𝖎𝖓𝖊𝖗𝖆10 Jan 2023 23:43 UTC
25 points
2 comments2 min readEA link
(fortune.com)

Leav­ing Open Philan­thropy, go­ing to Anthropic

Joe_Carlsmith3 Nov 2025 17:41 UTC
141 points
14 comments18 min readEA link

Main paths to im­pact in EU AI Policy

JOMG_Monnet8 Dec 2022 16:17 UTC
69 points
2 comments8 min readEA link

If in­ter­pretabil­ity re­search goes well, it may get dangerous

So8res3 Apr 2023 21:48 UTC
33 points
0 comments2 min readEA link

[Linkpost] A Nar­row Path—How to Se­cure our Future

MathiasKB🔸2 Oct 2024 22:50 UTC
68 points
0 comments1 min readEA link
(www.narrowpath.co)

Are we drop­ping the ball on Recom­men­da­tion AIs?

Raphaël S23 Oct 2024 19:37 UTC
5 points
0 comments6 min readEA link

4 Years Later: Pres­i­dent Trump and Global Catas­trophic Risk

HaydnBelfield25 Oct 2020 16:28 UTC
43 points
10 comments10 min readEA link

Heuris­tics for as­sess­ing how much of a bub­ble AI is in/​will be

Remmelt28 Oct 2025 8:08 UTC
14 points
1 comment2 min readEA link
(www.wired.com)

Hiring a CEO & EU Tech Policy Lead to launch an AI policy ca­reer org in Europe

Cillian_6 Dec 2023 13:52 UTC
50 points
0 comments7 min readEA link

Propos­ing the Con­di­tional AI Safety Treaty (linkpost TIME)

Otto15 Nov 2024 13:56 UTC
12 points
6 comments3 min readEA link
(time.com)

Joe Hardie on Ar­ca­dia Im­pact’s pro­jects (FBB #7)

gergo8 Jul 2025 13:22 UTC
18 points
3 comments15 min readEA link

Recom­men­da­tion to Ap­ply ISIC and NAICS to AI In­ci­dent Database

Ben Turse21 Jul 2024 7:25 UTC
3 points
0 comments2 min readEA link

AISN #54: OpenAI Up­dates Restruc­ture Plan

Center for AI Safety13 May 2025 16:48 UTC
7 points
0 comments4 min readEA link
(newsletter.safe.ai)

In­cen­tive de­sign and ca­pa­bil­ity elicitation

Joe_Carlsmith12 Nov 2024 20:56 UTC
9 points
0 comments12 min readEA link

A course for the gen­eral pub­lic on AI

LeandroD31 Aug 2020 1:29 UTC
1 point
0 comments1 min readEA link

Effi­cacy of AI Ac­tivism: Have We Ever Said No?

Charlie Harrison27 Oct 2023 16:52 UTC
80 points
25 comments20 min readEA link

#214 – Con­trol­ling AI that wants to take over – so we can use it any­way (Buck Sh­legeris on The 80,000 Hours Pod­cast)

80000_Hours4 Apr 2025 19:59 UTC
17 points
0 comments32 min readEA link

Nine Points of Col­lec­tive Insanity

Remmelt27 Dec 2022 3:14 UTC
1 point
0 comments1 min readEA link
(mflb.com)

Train­ing for Good—Up­date & Plans for 2023

Cillian_15 Nov 2022 16:02 UTC
82 points
1 comment10 min readEA link

AI Offense Defense Balance in a Mul­tipo­lar World

Otto17 Jul 2025 9:47 UTC
15 points
0 comments19 min readEA link
(www.existentialriskobservatory.org)

AISN #53: An Open Let­ter At­tempts to Block OpenAI Restructuring

Center for AI Safety29 Apr 2025 15:56 UTC
6 points
0 comments4 min readEA link
(newsletter.safe.ai)

Are New Ideas in AI Get­ting Harder to Find?

Charlie Harrison10 Dec 2024 12:52 UTC
40 points
3 comments5 min readEA link

On Com­par­a­tive Ad­van­tage & AGI

Charles Dillon 🔸3 Dec 2025 21:01 UTC
27 points
0 comments3 min readEA link

The Boiled-Frog Failure Mode

ontologics30 Jun 2025 13:24 UTC
7 points
3 comments5 min readEA link

Skep­ti­cism to­wards claims about the views of pow­er­ful institutions

tlevin13 Feb 2025 7:40 UTC
20 points
1 comment4 min readEA link

AI safety ad­vo­cates should con­sider pro­vid­ing gen­tle push­back fol­low­ing the events at OpenAI

I_machinegun_Kelly22 Dec 2023 21:05 UTC
86 points
5 comments3 min readEA link
(www.lesswrong.com)

Tether­ware #1: The case for hu­man­like AI with free will

Jáchym Fibír30 Jan 2025 11:57 UTC
−3 points
2 comments10 min readEA link
(tetherware.substack.com)

I read ev­ery ma­jor AI lab’s safety plan so you don’t have to

sarahhw16 Dec 2024 14:12 UTC
68 points
2 comments11 min readEA link
(longerramblings.substack.com)

Cut­ting AI Safety down to size

Holly Elmore ⏸️ 🔸9 Nov 2024 23:40 UTC
90 points
5 comments5 min readEA link

Ap­pendix to Bridg­ing Demonstration

mako yass1 Jun 2022 20:30 UTC
18 points
2 comments28 min readEA link

An­nounc­ing Con­ver­gence Anal­y­sis: An In­sti­tute for AI Sce­nario & Gover­nance Research

David_Kristoffersson7 Mar 2024 21:18 UTC
46 points
0 comments4 min readEA link

[Question] How in­de­pen­dent is the re­search com­ing out of OpenAI’s pre­pared­ness team?

Earthling10 Feb 2024 16:59 UTC
18 points
0 comments1 min readEA link

Gw­ern on cre­at­ing your own AI race and China’s Fast Fol­lower strat­egy.

Larks25 Nov 2024 3:01 UTC
129 points
4 comments2 min readEA link
(www.lesswrong.com)

Toby Ord’s new re­port on les­sons from the de­vel­op­ment of the atomic bomb

Ishan Mukherjee22 Nov 2022 10:37 UTC
65 points
3 comments1 min readEA link
(www.governance.ai)

A col­lec­tion of AI Gover­nance-re­lated Pod­casts, Newslet­ters, Blogs, and more

LintzA2 Oct 2021 0:46 UTC
24 points
1 comment1 min readEA link

[Question] Ben Horow­itz and oth­ers are spread­ing a “reg­u­la­tion is bad” view. Would it be use­ful to have a pub­lic bet on “would Ben up­date his view if he had 1-1 with X-Risk re­searcher?”, and urge Ben to run such an ex­per­i­ment?

AntonOsika8 Aug 2023 6:36 UTC
2 points
0 comments1 min readEA link

How to in­fluence the AI trainer workforce

Singer Robin25 Aug 2025 5:21 UTC
3 points
0 comments2 min readEA link

Thoughts about Policy Ecosys­tems: The Miss­ing Links in AI Governance

Echo Huang31 Jan 2025 13:23 UTC
22 points
2 comments5 min readEA link

AI Safety: The [Hy­po­thet­i­cal] Video Game

barryl 🔸18 Apr 2025 20:19 UTC
3 points
2 comments3 min readEA link

The Wind­fall Clause: How to Prevent AI-In­duced Wealth Inequality

Strad Slater11 Nov 2025 9:18 UTC
12 points
1 comment6 min readEA link
(williamslater2003.medium.com)

List of pe­ti­tions against OpenAI’s for-profit move

Remmelt25 Apr 2025 10:03 UTC
13 points
4 comments1 min readEA link

The repli­ca­tion and em­u­la­tion of GPT-3

Ben Cottier21 Dec 2022 13:49 UTC
14 points
0 comments33 min readEA link

From Con­flict to Coex­is­tence: Rewrit­ing the Game Between Hu­mans and AGI

Michael Batell6 May 2025 5:09 UTC
15 points
2 comments35 min readEA link

US credit rat­ing down­graded, $1T in Gulf state in­vest­ments in the US, Kur­dis­tan Work­ers’ Party dis­banded | Sen­tinel Global Risks Weekly Roundup #20/​2025

NunoSempere19 May 2025 18:02 UTC
50 points
0 comments1 min readEA link
(blog.sentinel-team.org)

Three Types of In­tel­li­gence Explosion

rosehadshar17 Mar 2025 14:47 UTC
45 points
2 comments3 min readEA link
(www.forethought.org)

AI safety re­mains un­der­funded by more than 3 OOMs

Impatient_Longtermist 🔸🌱6 Oct 2025 19:53 UTC
25 points
3 comments1 min readEA link
(www.nber.org)

Credo AI is hiring!

IanEisenberg3 Mar 2022 18:02 UTC
16 points
6 comments4 min readEA link

Aus­trali­ans call for AI safety to be taken seriously

Alexander Saeri21 Jul 2023 1:16 UTC
51 points
1 comment1 min readEA link

The Need for Poli­ti­cal Ad­ver­tis­ing (Post 2 of 7 on AI Gover­nance)

Jason Green-Lowe21 May 2025 0:52 UTC
60 points
0 comments13 min readEA link

No, the EMH does not im­ply that mar­kets have long AGI timelines

Jakob24 Apr 2023 8:27 UTC
83 points
21 comments8 min readEA link

Start an AI safety group with the Path­fin­der Fellowship

Topaz7 Nov 2025 12:57 UTC
14 points
0 comments1 min readEA link

Op­tion control

Joe_Carlsmith4 Nov 2024 17:54 UTC
11 points
0 comments54 min readEA link

Dra­co­nian mea­sures can in­crease the risk of ir­re­vo­ca­ble catastrophe

dsj23 Sep 2025 21:40 UTC
8 points
1 comment2 min readEA link
(thedavidsj.substack.com)

In­tro­duc­ing SB53.info

Miles Kodama25 Jul 2025 9:42 UTC
50 points
4 comments7 min readEA link

In­tro­duc­ing a New Course on the Eco­nomics of AI

akorinek21 Dec 2021 4:55 UTC
84 points
6 comments2 min readEA link

How I ex­pect TAI to im­pact de­vel­op­ing countries

Tax Geek18 Oct 2025 8:54 UTC
40 points
8 comments7 min readEA link

China x AI Refer­ence List

Saad Siddiqui13 Mar 2024 18:57 UTC
61 points
3 comments3 min readEA link
(docs.google.com)

US-China trade talks should pave way for AI safety treaty [SCMP cross­post]

Otto16 May 2025 20:53 UTC
15 points
1 comment3 min readEA link

Im­pli­ca­tions of the White­house meet­ing with AI CEOs for AI su­per­in­tel­li­gence risk—a first-step to­wards evals?

Jamie B7 May 2023 17:33 UTC
78 points
3 comments7 min readEA link

Ap­ply now for the EU Tech Policy Fel­low­ship 2023

Jan-Willem11 Nov 2022 6:16 UTC
64 points
1 comment5 min readEA link

(out­dated ver­sion) Shortlist of Longter­mist Interventions

Jordan Arel21 Oct 2025 11:59 UTC
4 points
0 comments14 min readEA link

Liti­gate-for-Im­pact: Prepar­ing Le­gal Ac­tion against an AGI Fron­tier Lab Leader

Sonia M Joseph8 Dec 2024 14:28 UTC
77 points
1 comment2 min readEA link

My new non­profit Evitable is hiring.

David Krueger7 Nov 2025 3:40 UTC
24 points
0 comments1 min readEA link

AI Benefits Post 5: Out­stand­ing Ques­tions on Govern­ing Benefits

Cullen 🔸21 Jul 2020 16:45 UTC
5 points
0 comments4 min readEA link

9+ weeks of men­tored AI safety re­search in Lon­don – Pivotal Re­search Fellowship

Tobias Häberli12 Nov 2025 15:21 UTC
14 points
0 comments2 min readEA link

The Real AI Threat: Com­fortable Obsolescence

Andrei Navrotskii11 Nov 2025 22:11 UTC
4 points
0 comments15 min readEA link

Jailbreak­ing Claude 4 and Other Fron­tier Lan­guage Models

James-Sullivan15 Jun 2025 1:01 UTC
6 points
0 comments3 min readEA link
(open.substack.com)

How do we solve the al­ign­ment prob­lem?

Joe_Carlsmith13 Feb 2025 18:27 UTC
38 points
1 comment7 min readEA link
(joecarlsmith.substack.com)

AI Model Registries: A Foun­da­tional Tool for AI Governance

Elliot Mckernon7 Oct 2024 13:59 UTC
19 points
0 comments4 min readEA link
(www.convergenceanalysis.org)

*New* Canada AI Safety & Gover­nance community

Wyatt Tessari L'Allié29 Aug 2022 15:58 UTC
32 points
2 comments1 min readEA link

Towards an al­ter­na­tive to the COPs

Arnold Bomans19 Nov 2025 14:53 UTC
1 point
1 comment1 min readEA link

[Re­port] Bridg­ing the In­ter­na­tional AI Gover­nance Divide: Key Strate­gies for In­clud­ing the Global South

Heramb Podar26 Jan 2025 23:55 UTC
8 points
0 comments1 min readEA link
(encodeai.org)

Map­ping ar­tifi­cial in­tel­li­gence in the United States: A ge­o­graphic anal­y­sis of the tech­nol­ogy in­fras­truc­ture in U.S. data cen­ters.

GabrielRB30 Apr 2025 15:23 UTC
10 points
1 comment16 min readEA link

Win­ning Non-Triv­ial Pro­ject: Set­ting a high stan­dard for fron­tier model security

XaviCF8 Jan 2024 11:20 UTC
31 points
0 comments18 min readEA link

Biosafety Reg­u­la­tions (BMBL) and their rele­vance for AI

stepanlos29 Jun 2023 19:20 UTC
8 points
0 comments4 min readEA link

Shar­ing the Global AI Gover­nance Alliance

JordanStone17 Aug 2025 19:30 UTC
7 points
0 comments1 min readEA link

How Open Source Ma­chine Learn­ing Soft­ware Shapes AI

Max L28 Sep 2022 17:49 UTC
11 points
3 comments15 min readEA link
(maxlangenkamp.me)

What are the differ­ences be­tween AGI, trans­for­ma­tive AI, and su­per­in­tel­li­gence?

Vishakha Agrawal23 Jan 2025 10:11 UTC
12 points
0 comments3 min readEA link
(aisafety.info)

Rea­sons for su­per­pow­ers to de­velop (and not de­velop) su­per in­tel­li­gent AI?

flyingtiger25 Mar 2025 22:22 UTC
1 point
0 comments1 min readEA link

Or­phaned Poli­cies (Post 5 of 7 on AI Gover­nance)

Jason Green-Lowe29 May 2025 21:42 UTC
46 points
3 comments16 min readEA link

AI gov­er­nance tracker of each coun­try per re­gion

Alix Ramillon24 Jul 2024 17:39 UTC
16 points
2 comments23 min readEA link

AI Safety Camp 11

Robert Kralisch7 Nov 2025 14:27 UTC
7 points
1 comment15 min readEA link

IABIED Re­view—An Un­for­tu­nate Miss

Darren McKee18 Sep 2025 22:39 UTC
22 points
2 comments9 min readEA link

Eth­i­cal Roots of Chi­nese AI

Vasiliy Kondyrev5 Nov 2024 14:07 UTC
0 points
0 comments6 min readEA link

Where I Am Donat­ing in 2024

MichaelDickens19 Nov 2024 0:09 UTC
181 points
73 comments46 min readEA link

Pos­si­ble di­rec­tions in AI ideal gov­er­nance research

RoryG10 Aug 2022 8:36 UTC
5 points
0 comments3 min readEA link

Soft Na­tion­al­iza­tion: How the US Govern­ment Will Con­trol AI Labs

Deric Cheng27 Aug 2024 15:10 UTC
103 points
6 comments21 min readEA link
(www.convergenceanalysis.org)

METR: Mea­sur­ing AI Abil­ity to Com­plete Long Tasks

Ben_West🔸19 Mar 2025 16:49 UTC
122 points
16 comments1 min readEA link
(metr.org)

2024 State of AI Reg­u­la­tory Landscape

Deric Cheng28 May 2024 12:00 UTC
12 points
1 comment2 min readEA link
(www.convergenceanalysis.org)

Un­jour­nal eval­u­a­tion of “Towards best prac­tices in AGI safety and gov­er­nance” (Schuett et al, 2023)

david_reinstein3 Jun 2025 11:18 UTC
9 points
1 comment1 min readEA link
(unjournal.pubpub.org)

As­ter­isk Magaz­ine Is­sue 03: AI

alejandro24 Jul 2023 15:53 UTC
34 points
3 comments1 min readEA link
(asteriskmag.com)

In­ter­view with a drone ex­pert on the fu­ture of AI warfare

NunoSempere9 Oct 2025 20:20 UTC
46 points
2 comments4 min readEA link
(blog.sentinel-team.org)

A challenge for AGI or­ga­ni­za­tions, and a challenge for readers

RobBensinger1 Dec 2022 23:11 UTC
172 points
13 comments2 min readEA link

How to re­duce risks re­lated to con­scious AI: A user guide [Con­scious AI & Public Per­cep­tion]

Jay Luong5 Jul 2024 14:19 UTC
9 points
1 comment15 min readEA link

A differ­ent take on the Musk v OpenAI pre­limi­nary in­junc­tion order

TFD11 Mar 2025 14:29 UTC
6 points
1 comment20 min readEA link
(www.thefloatingdroid.com)

Evals pro­jects I’d like to see, and a call to ap­ply to OP’s evals RFP

cb25 Mar 2025 11:50 UTC
25 points
2 comments3 min readEA link

Bench­mark­ing Emo­tional Align­ment: Can VSPE Re­duce Flat­tery in LLMs?

Astelle Kay4 Aug 2025 3:36 UTC
2 points
0 comments3 min readEA link

“That’s (not) me!”: The mal­i­cious em­ploy­ment of deep­fakes and their miti­ga­tion in le­gal en­vi­ron­ments for AI governance

Gabriela Pardo1 May 2025 14:54 UTC
5 points
0 comments12 min readEA link

What We Can Do to Prevent Ex­tinc­tion by AI

Joe Rogero24 Feb 2025 17:15 UTC
23 points
3 comments11 min readEA link

Will we ever run out of new jobs?

Kevin Kohler19 Aug 2024 15:03 UTC
11 points
4 comments7 min readEA link
(machinocene.substack.com)

Con­tribute by fa­cil­i­tat­ing the AGI Safety Fun­da­men­tals Programme

Jamie B6 Dec 2021 11:50 UTC
27 points
0 comments2 min readEA link

Align­ing AI with Hu­mans by Lev­er­ag­ing Le­gal Informatics

johnjnay18 Sep 2022 7:43 UTC
20 points
11 comments3 min readEA link

Me­tac­u­lus Launches Chi­nese AI Chips Tour­na­ment, Sup­port­ing In­sti­tute for AI Policy and Strat­egy Research

christian6 Dec 2023 11:26 UTC
27 points
0 comments1 min readEA link
(www.metaculus.com)

Emer­gency pod: Judge plants a le­gal time bomb un­der OpenAI (with Rose Chan Loui)

80000_Hours7 Mar 2025 19:24 UTC
62 points
18 comments2 min readEA link

[Question] Should we na­tion­al­ize AI de­vel­op­ment?

Jadon Schmitt20 Jul 2023 5:31 UTC
5 points
4 comments1 min readEA link

[Question] What are some sources re­lated to big-pic­ture AI strat­egy?

Jacob Watts🔸2 Mar 2023 5:04 UTC
9 points
4 comments1 min readEA link

Ac­cel­er­ated Hori­zons — Pod­cast + Blog Idea

Cadejs16 Apr 2025 14:20 UTC
2 points
3 comments1 min readEA link

Reflec­tions on Dario Amodei’s ‘Ur­gency of In­ter­pretabil­ity’

Strad Slater27 Nov 2025 8:30 UTC
2 points
0 comments5 min readEA link
(williamslater2003.medium.com)

Distil­la­tion of The Offense-Defense Balance of Scien­tific Knowledge

Arjun Yadav12 Aug 2022 7:01 UTC
17 points
0 comments2 min readEA link

An­thropic’s lead­ing re­searchers acted as mod­er­ate accelerationists

Remmelt1 Sep 2025 23:23 UTC
79 points
4 comments42 min readEA link

AI Im­pacts Quar­terly Newslet­ter, Jan-Mar 2023

Harlan17 Apr 2023 23:07 UTC
20 points
1 comment3 min readEA link
(blog.aiimpacts.org)

Some gov­er­nance re­search ideas to pre­vent malev­olent con­trol over AGI and why this might mat­ter a hell of a lot

Jim Buhler23 May 2023 13:07 UTC
64 points
5 comments16 min readEA link

AI, An­i­mals, & Digi­tal Minds 2025: Retrospective

Alistair Stewart12 Jul 2025 2:28 UTC
64 points
3 comments11 min readEA link

Differ­en­tial knowl­edge interconnection

Roman Leventov12 Oct 2024 12:52 UTC
3 points
1 comment7 min readEA link

The Bot­tle­neck in AI Policy Isn’t Ethics—It’s Implementation

Tristan D4 Apr 2025 6:07 UTC
10 points
4 comments1 min readEA link

By de­fault, cap­i­tal will mat­ter more than ever af­ter AGI

L Rudolf L28 Dec 2024 17:52 UTC
113 points
3 comments16 min readEA link
(nosetgauge.substack.com)

Stable to­tal­i­tar­i­anism: an overview

80000_Hours29 Oct 2024 16:07 UTC
36 points
1 comment20 min readEA link
(80000hours.org)

Beyond Con­trol: The Strate­gic Case for AI Rights

Dawn Drescher12 Aug 2025 14:06 UTC
8 points
3 comments3 min readEA link
(impartial-priorities.org)

Grad­ual Disem­pow­er­ment: Con­crete Re­search Projects

Raymond D29 May 2025 18:58 UTC
20 points
1 comment10 min readEA link

Call for Cruxes by Rhyme, a Longter­mist His­tory Con­sul­tancy

Lara_TH1 Mar 2023 10:20 UTC
147 points
6 comments3 min readEA link

We read ev­ery labs safety plan so you don’t have to: 2025 edition

Algon29 Oct 2025 16:48 UTC
14 points
1 comment16 min readEA link
(aisafety.info)

AI Might Kill Every­one

Bentham's Bulldog5 Jun 2025 15:36 UTC
20 points
1 comment4 min readEA link

In­ter­na­tional co­op­er­a­tion as a tool to re­duce two ex­is­ten­tial risks.

johl@umich.edu19 Apr 2021 16:51 UTC
28 points
4 comments23 min readEA link

How long will reach­ing a Risk Aware­ness Mo­ment and CHARTS agree­ment take?

Yadav6 Sep 2023 16:39 UTC
12 points
0 comments14 min readEA link

Prepar­ing for the In­tel­li­gence Explosion

finm11 Mar 2025 15:38 UTC
120 points
15 comments1 min readEA link
(www.forethought.org)

His­tory’s Gran­d­est Pro­jects: In­tro­duc­tion to Macro Strate­gies for AI Risk, Part 1

Coleman20 Jun 2025 17:32 UTC
7 points
0 comments38 min readEA link

What if we don’t need a “Hard Left Turn” to reach AGI?

Eigengender15 Jul 2022 9:49 UTC
39 points
7 comments4 min readEA link

Four Pre­dic­tions About OpenAI’s Plans To Re­tain Non­profit Control

Garrison7 May 2025 15:48 UTC
15 points
2 comments5 min readEA link
(www.obsolete.pub)

An­nounc­ing New Begin­ner-friendly Book on AI Safety and Risk

Darren McKee25 Nov 2023 15:57 UTC
117 points
9 comments1 min readEA link

Ex­tend­ing Con­sti­tu­tional AI to Com­mu­nity Gover­nance: Re­duc­ing Mor­tal­ity from Psy­chi­a­tric Detention

Mac Black1 Dec 2025 15:14 UTC
0 points
0 comments7 min readEA link

The Elic­i­ta­tion Game: Eval­u­at­ing ca­pa­bil­ity elic­i­ta­tion techniques

Teun van der Weij27 Feb 2025 20:33 UTC
3 points
0 comments2 min readEA link

How to make the fu­ture bet­ter (other than by re­duc­ing ex­tinc­tion risk)

William_MacAskill15 Aug 2025 15:40 UTC
45 points
3 comments3 min readEA link

Cor­po­rate Gover­nance for Fron­tier AI Labs: A Re­search Agenda

Matthew Wearden28 Feb 2024 11:32 UTC
18 points
3 comments16 min readEA link
(matthewwearden.co.uk)

The Rise of AI Agents: Con­se­quences and Challenges Ahead

Tristan D28 Mar 2025 5:19 UTC
5 points
0 comments15 min readEA link

Im­pact of Quan­ti­za­tion on Small Lan­guage Models (SLMs) for Mul­tilin­gual Math­e­mat­i­cal Rea­son­ing Tasks

Angie Paola Giraldo7 May 2025 21:48 UTC
11 points
0 comments14 min readEA link

Is the AI In­dus­try in a Bub­ble?

Yarrow Bouchard 🔸16 Nov 2025 1:00 UTC
37 points
8 comments15 min readEA link

A con­cern­ing ob­ser­va­tion from me­dia cov­er­age of AI in­dus­try dynamics

Justin Olive2 Mar 2023 23:56 UTC
48 points
5 comments3 min readEA link

Re­quest for In­for­ma­tion for a new US AI Ac­tion Plan (OSTP RFI)

Agustín Covarrubias 🔸7 Feb 2025 20:22 UTC
19 points
2 comments2 min readEA link
(www.federalregister.gov)

Effec­tive AI Outreach | A Data Driven Approach

NoahCWilson🔸28 Feb 2025 0:44 UTC
15 points
2 comments15 min readEA link

Stu­art Rus­sell Hu­man Com­pat­i­ble AI Roundtable with Allan Dafoe, Rob Re­ich, & Ma­ri­etje Schaake

Mahendra Prasad11 Feb 2021 7:43 UTC
16 points
0 comments1 min readEA link

AISN #61: OpenAI Re­leases GPT-5

Center for AI Safety12 Aug 2025 17:52 UTC
6 points
0 comments4 min readEA link
(newsletter.safe.ai)

The Case for Jour­nal­ism on AI

michel19 Feb 2025 19:45 UTC
95 points
5 comments4 min readEA link

[Question] Im­pact of (po­ten­tial) fed­eral AI rele­vant po­si­tion?

Mako06022 Sep 2025 22:37 UTC
3 points
0 comments1 min readEA link

Com­pli­ance Mon­i­tor­ing as an Im­pact­ful Mechanism of AI Safety Policy

CAISID7 Feb 2024 16:10 UTC
6 points
3 comments9 min readEA link

Early Chi­nese Lan­guage Me­dia Cover­age of the AI 2027 Re­port: A Qual­i­ta­tive Analysis

eeeee30 Apr 2025 14:23 UTC
14 points
0 comments11 min readEA link
(www.lesswrong.com)

(Re­port) Eval­u­at­ing Taiwan’s Tac­tics to Safe­guard its Semi­con­duc­tor As­sets Against a Chi­nese Invasion

Yadav7 Dec 2023 0:01 UTC
16 points
0 comments22 min readEA link
(bristolaisafety.org)

AI De­faults: A Ne­glected Lever for An­i­mal Welfare?

andiehansen30 May 2025 9:59 UTC
13 points
0 comments10 min readEA link

New OGL and ITAR changes are shift­ing AI Gover­nance and Policy be­low the sur­face: A sim­plified up­date

CAISID31 May 2024 7:54 UTC
12 points
2 comments3 min readEA link

U.S. Govern­ment Seeks In­put on Na­tional AI R&D Strate­gic Plan—Dead­line May 29

Matt Brooks27 May 2025 1:53 UTC
8 points
1 comment1 min readEA link

[Question] Hyper-Abun­dant Systems

DavidofOmegaS19 Feb 2025 13:20 UTC
1 point
0 comments4 min readEA link

The Choice Transition

Owen Cotton-Barratt18 Nov 2024 12:32 UTC
49 points
1 comment15 min readEA link
(strangecities.substack.com)

In AI Gover­nance, let the Non-EA World Train You First

Camille23 Jul 2025 17:46 UTC
10 points
0 comments1 min readEA link

AGI Safety.

Jensen1130 Aug 2025 13:46 UTC
1 point
0 comments3 min readEA link

Reflec­tions on Com­pat­i­bil­ism, On­tolog­i­cal Trans­la­tions, and the Ar­tifi­cial Divine

Mahdi Complex7 May 2025 12:17 UTC
−4 points
0 comments22 min readEA link

The Miss­ing Piece: Why We Need a Grand Strat­egy for AI

Coleman28 Feb 2025 23:49 UTC
7 points
1 comment9 min readEA link

OpenAI: The non­profit re­fuses to die (with Tyler Whit­mer)

80000_Hours11 Nov 2025 21:55 UTC
19 points
0 comments5 min readEA link

Les­sons for AI Gover­nance from Atoms for Peace

Amritanshu Prasad16 Apr 2025 14:25 UTC
10 points
2 comments2 min readEA link
(www.thenextfrontier.blog)

Ar­chi­tect­ing Trust: A Con­cep­tual Blueprint for Ver­ifi­able AI Governance

Ihor Ivliev31 Mar 2025 18:48 UTC
3 points
0 comments8 min readEA link

The Work of Chad Jones

Nicholas Decker13 Mar 2025 18:00 UTC
12 points
0 comments1 min readEA link
(nicholasdecker.substack.com)

Im­pli­ca­tions of the in­fer­ence scal­ing paradigm for AI safety

Ryan Kidd15 Jan 2025 0:59 UTC
48 points
5 comments5 min readEA link

When AI Speaks Too Soon: How Pre­ma­ture Reve­la­tion Can Sup­press Hu­man Emergence

KaedeHamasaki10 Apr 2025 18:19 UTC
1 point
3 comments3 min readEA link

10 of Founders Pledge’s biggest grants

Matt_Lerner9 Jul 2025 21:55 UTC
124 points
1 comment6 min readEA link

A Plau­si­ble AI Eco­nomic Scenario

Deric Cheng19 Oct 2025 6:19 UTC
10 points
3 comments5 min readEA link

The flaws that make to­day’s AI ar­chi­tec­ture un­safe and a new ap­proach that could fix it

80000_Hours22 Jun 2020 22:15 UTC
3 points
0 comments86 min readEA link
(80000hours.org)

AI Risk and Sur­vivor­ship Bias—How An­dreessen and LeCun got it wrong

stepanlos14 Jul 2023 17:10 UTC
5 points
1 comment6 min readEA link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibs29 Mar 2023 23:30 UTC
212 points
75 comments3 min readEA link
(time.com)

Data Publi­ca­tion for the 2021 Ar­tifi­cial In­tel­li­gence, Mo­ral­ity, and Sen­tience (AIMS) Sur­vey

Janet Pauketat24 Mar 2022 15:43 UTC
21 points
0 comments3 min readEA link
(www.sentienceinstitute.org)

Na­tional Se­cu­rity Is Not In­ter­na­tional Se­cu­rity: A Cri­tique of AGI Realism

C.K.2 Feb 2025 17:04 UTC
44 points
2 comments36 min readEA link
(conradkunadu.substack.com)

The Need for an Effec­tive AI In­ci­dent Re­port­ing Framework

Strad Slater13 Nov 2025 8:53 UTC
2 points
0 comments4 min readEA link
(williamslater2003.medium.com)

Se­na­tor Scott Wiener Q+A with METR policy di­rec­tor Chris Painter

Rachel Shu5 Dec 2025 7:10 UTC
4 points
0 comments1 min readEA link

Fore­cast­ing Com­pute—Trans­for­ma­tive AI and Com­pute [2/​4]

lennart1 Oct 2021 8:25 UTC
39 points
6 comments19 min readEA link

[Question] Can you donate to AI advocacy

k6427 May 2025 16:37 UTC
4 points
2 comments1 min readEA link

[Question] What “defense lay­ers” should gov­ern­ments, AI labs, and busi­nesses use to pre­vent catas­trophic AI failures?

LintzA3 Dec 2021 14:24 UTC
37 points
3 comments1 min readEA link

AGI × An­i­mals Wargame

Sentient Futures8 Oct 2025 2:14 UTC
75 points
2 comments5 min readEA link

Forethought has room for more funding

Forethought21 Nov 2025 15:42 UTC
89 points
14 comments1 min readEA link

AISN #50: AI Ac­tion Plan Re­sponses

Center for AI Safety31 Mar 2025 20:07 UTC
10 points
0 comments6 min readEA link
(newsletter.safe.ai)

Microsoft’s CEO Satya Nadella says he doesn’t be­lieve in AGI

Yarrow Bouchard 🔸12 Nov 2025 23:03 UTC
12 points
1 comment4 min readEA link

The Case for an On­line En­cy­clo­pe­dia Man­aged by AI Agents

Casey Milkweed21 Jul 2025 14:06 UTC
2 points
0 comments1 min readEA link
(substack.com)

An­nounc­ing the GovAI Policy Team

MarkusAnderljung1 Aug 2022 22:46 UTC
107 points
11 comments2 min readEA link

The OpenAI Foun­da­tion has an­nounced its first round of grants

Larks4 Dec 2025 4:34 UTC
19 points
10 comments1 min readEA link
(openai.com)

Daniel Koko­ta­jlo on AI 2027: Race Dy­nam­ics, Robot Economies, and the Nar­row Path to a Good Future

80000_Hours21 Oct 2025 19:51 UTC
7 points
1 comment18 min readEA link

[Question] Any fur­ther work on AI Safety Suc­cess Sto­ries?

Krieger2 Oct 2022 11:59 UTC
4 points
0 comments1 min readEA link

#219 – Graphs AI com­pa­nies would pre­fer you didn’t (fully) un­der­stand (Toby Ord on The 80,000 Hours Pod­cast)

80000_Hours25 Jun 2025 18:23 UTC
19 points
0 comments27 min readEA link

HIRING: In­form and shape a new pro­ject on AI safety at Part­ner­ship on AI

Madhulika Srikumar24 Nov 2021 16:29 UTC
11 points
2 comments1 min readEA link

Join the AI gov­er­nance and in­ter­pretabil­ity hackathons!

Esben Kran23 Mar 2023 14:39 UTC
33 points
1 comment5 min readEA link
(alignmentjam.com)

How to get tech­nolog­i­cal knowl­edge on AI/​ML (for non-tech peo­ple)

FangFang30 Jun 2021 7:53 UTC
63 points
7 comments5 min readEA link

Long-term AI policy strat­egy re­search and implementation

Benjamin_Todd9 Nov 2021 0:00 UTC
1 point
0 comments7 min readEA link
(80000hours.org)

Epistemic Spot Check: Ex­pected Value of Donat­ing to Alex Bores’s Con­gres­sional Campaign

MichaelDickens13 Nov 2025 19:09 UTC
67 points
3 comments6 min readEA link

Ten AI safety pro­jects I’d like peo­ple to work on

JulianHazell24 Jul 2025 15:32 UTC
51 points
7 comments10 min readEA link

Effec­tive Lob­by­ism: Les­sons from 5 years of practice

++Johnny Stengel++23 Sep 2025 19:19 UTC
56 points
3 comments5 min readEA link

Me­tac­u­lus Launches Fu­ture of AI Series, Based on Re­search Ques­tions by Arb

christian13 Mar 2024 21:14 UTC
34 points
0 comments1 min readEA link
(www.metaculus.com)

Ven­ture Cap­i­tal In­fluence Tracker

Kayode Adekoya25 Nov 2025 14:09 UTC
1 point
0 comments5 min readEA link

By failing to take se­ri­ous AI ac­tion, the US could be in vi­o­la­tion of its in­ter­na­tional law obligations

Cecil Abungu 27 May 2023 4:25 UTC
45 points
1 comment10 min readEA link

The Com­pendium, A full ar­gu­ment about ex­tinc­tion risk from AGI

adamShimi31 Oct 2024 12:02 UTC
9 points
1 comment2 min readEA link
(www.thecompendium.ai)

Con­test for Bet­ter AGI Safety Plans

Peter3 Jul 2025 17:02 UTC
18 points
0 comments8 min readEA link
(manifund.org)

The moral ar­gu­ment for giv­ing AIs autonomy

Matthew_Barnett8 Jan 2025 0:59 UTC
41 points
7 comments11 min readEA link

[Question] An­thropic says it’s highly con­fi­dent a Chi­nese state-spon­sored group used AI to hack gov­ern­ments, chem­i­cal firms, and oth­ers. Why isn’t this get­ting more at­ten­tion?

adam.kruger16 Nov 2025 21:27 UTC
13 points
5 comments1 min readEA link

Don’t Bet the Fu­ture on Win­ning an AI Arms Race

Eric Drexler11 Jul 2025 11:11 UTC
25 points
1 comment5 min readEA link

AI Benefits Post 2: How AI Benefits Differs from AI Align­ment & AI for Good

Cullen 🔸29 Jun 2020 16:59 UTC
9 points
0 comments2 min readEA link

A Devel­op­men­tal Ap­proach to AI Safety: Re­plac­ing Sup­pres­sion with Reflec­tive Learning

Petra Vojtassakova23 Oct 2025 16:01 UTC
2 points
0 comments5 min readEA link

The ‘Bad Par­ent’ Prob­lem: Why Hu­man So­ciety Com­pli­cates AI Alignment

Beyond Singularity5 Apr 2025 21:08 UTC
11 points
1 comment3 min readEA link

AI and An­i­mal Welfare: A Policy Case Study from Aotearoa New Zealand Policy

Karen Singleton20 Oct 2025 20:57 UTC
45 points
5 comments6 min readEA link

AI, An­i­mals, & Digi­tal Minds 2025: ap­ply to speak by Wed­nes­day!

Alistair Stewart5 May 2025 0:45 UTC
8 points
0 comments1 min readEA link

Want to win the AGI race? Solve al­ign­ment.

leopold29 Mar 2023 15:19 UTC
56 points
5 comments5 min readEA link
(www.forourposterity.com)

Open Agency model can solve the AI reg­u­la­tion dilemma

Roman Leventov9 Nov 2023 15:22 UTC
4 points
0 comments2 min readEA link

AGI Timelines in Gover­nance: Differ­ent Strate­gies for Differ­ent Timeframes

simeon_c19 Dec 2022 21:31 UTC
110 points
19 comments10 min readEA link

Go Mo­bi­lize? Les­sons from GM Protests for Paus­ing AI

Charlie Harrison24 Oct 2023 15:01 UTC
54 points
11 comments31 min readEA link

Ge­offrey Hin­ton on the Past, Pre­sent, and Fu­ture of AI

Stephen McAleese12 Oct 2024 16:41 UTC
5 points
1 comment18 min readEA link

Emer­gency pod: Don’t be­lieve OpenAI’s “non­profit” spin (with Tyler Whit­mer)

80000_Hours15 May 2025 16:52 UTC
37 points
0 comments2 min readEA link

OpenAI Alums, No­bel Lau­re­ates Urge Reg­u­la­tors to Save Com­pany’s Non­profit Structure

Garrison23 Apr 2025 23:01 UTC
61 points
2 comments8 min readEA link
(garrisonlovely.substack.com)

The Most Im­por­tant Thing We’ll Ever Do

Bentham's Bulldog24 Nov 2025 16:13 UTC
14 points
3 comments3 min readEA link

AI data gaps could lead to on­go­ing An­i­mal Suffering

Darkness8i817 Oct 2024 10:52 UTC
13 points
3 comments5 min readEA link

Paper Sum­mary: The Effec­tive­ness of AI Ex­is­ten­tial Risk Com­mu­ni­ca­tion to the Amer­i­can and Dutch Public

Otto9 Mar 2023 10:40 UTC
97 points
11 comments4 min readEA link

Open ques­tions on a Chi­nese in­va­sion of Taiwan and its effects on the semi­con­duc­tor stock

Yadav7 Dec 2023 16:39 UTC
21 points
0 comments2 min readEA link

AISN #31: A New AI Policy Bill in Cal­ifor­nia Plus, Prece­dents for AI Gover­nance and The EU AI Office

Center for AI Safety21 Feb 2024 21:55 UTC
27 points
0 comments6 min readEA link
(newsletter.safe.ai)

A brief his­tory of the au­to­mated corporation

Owen Cotton-Barratt4 Nov 2024 14:37 UTC
21 points
1 comment5 min readEA link
(strangecities.substack.com)

Con­clu­sion and Bibliog­ra­phy for “Un­der­stand­ing the diffu­sion of large lan­guage mod­els”

Ben Cottier21 Dec 2022 13:50 UTC
12 points
0 comments11 min readEA link

Trans­for­ma­tive AI and Com­pute—Read­ing List

Frederik Berg4 Sep 2023 6:21 UTC
24 points
0 comments1 min readEA link
(docs.google.com)

AISN #47: Rea­son­ing Models

Center for AI Safety6 Feb 2025 18:44 UTC
8 points
0 comments4 min readEA link
(newsletter.safe.ai)

Un­less its gov­er­nance changes, An­thropic is untrustworthy

MikhailSamin2 Dec 2025 17:07 UTC
36 points
3 comments29 min readEA link
(anthropic.ml)

How Can Aver­age Peo­ple Con­tribute to AI Safety?

Stephen McAleese6 Mar 2025 22:50 UTC
15 points
4 comments8 min readEA link

How Rood­man’s GWP model trans­lates to TAI timelines

kokotajlod16 Nov 2020 14:11 UTC
22 points
0 comments2 min readEA link

2024: a year of con­soli­da­tion for ORCG

JorgeTorresC18 Dec 2024 17:47 UTC
33 points
0 comments7 min readEA link
(www.orcg.info)

Help us find founders for new AI safety projects

lukeprog1 Dec 2025 16:57 UTC
66 points
3 comments1 min readEA link

In­tent al­ign­ment with­out moral al­ign­ment prob­a­bly leads to catastrophe

Alistair Stewart29 Aug 2025 17:21 UTC
12 points
0 comments5 min readEA link

The ELYSIUM Proposal

Roko16 Oct 2024 2:14 UTC
−10 points
0 comments1 min readEA link
(transhumanaxiology.substack.com)

Re­port: Eval­u­at­ing an AI Chip Regis­tra­tion Policy

Deric Cheng12 Apr 2024 4:40 UTC
15 points
0 comments5 min readEA link
(www.convergenceanalysis.org)

Fun­da­men­tal Risk

Ihor Ivliev26 Jun 2025 0:25 UTC
−5 points
0 comments1 min readEA link

The Oper­a­tor’s Gam­ble: A Pivot to Ma­te­rial Con­se­quence in AI Safety

Ihor Ivliev21 Jul 2025 19:33 UTC
−1 points
0 comments4 min readEA link

Tech­ni­cal AI Safety re­search tax­on­omy at­tempt (2025)

Ben Plaut27 Aug 2025 14:07 UTC
10 points
3 comments2 min readEA link

Brain Farm­ing: The Case for a Global Ban

Novel Minds Project27 Sep 2025 17:31 UTC
48 points
3 comments3 min readEA link

Ar­tifi­cial In­tel­li­gence, Mo­ral­ity, and Sen­tience (AIMS) Sur­vey: 2021

Janet Pauketat1 Jul 2022 7:47 UTC
36 points
0 comments2 min readEA link
(www.sentienceinstitute.org)

Chip Pro­duc­tion Policy Won’t Mat­ter as Much as You’d Think

Davidmanheim31 Aug 2025 18:58 UTC
33 points
8 comments5 min readEA link

[Question] What are the ar­gu­ments that sup­port China build­ing AGI+ if Western com­pa­nies de­lay/​pause AI de­vel­op­ment?

DMMF29 Mar 2023 18:53 UTC
32 points
9 comments1 min readEA link

AI Benefits Post 3: Direct and Indi­rect Ap­proaches to AI Benefits

Cullen 🔸6 Jul 2020 18:46 UTC
5 points
0 comments2 min readEA link

Promethean Gover­nance As­cen­dant: Les­sons from the Forge and Vi­sions for the Cos­mic Polity

Paul Fallavollita23 Mar 2025 0:54 UTC
−9 points
0 comments3 min readEA link

Com­par­i­son of LLM scal­a­bil­ity and perfor­mance be­tween the U.S. and China based on benchmark

Ivanna_alvarado12 Oct 2024 21:51 UTC
8 points
0 comments34 min readEA link

Promethean Gover­nance and Memetic Le­gi­t­i­macy: Les­sons from the Vene­tian Doge for AI Era Institutions

Paul Fallavollita19 Mar 2025 18:09 UTC
0 points
0 comments3 min readEA link

Google’s ethics is alarming

len.hoang.lnh25 Feb 2021 5:57 UTC
6 points
5 comments1 min readEA link

More ev­i­dence X-risk am­plifies ac­tion against cur­rent AI harms

Daniel_Friedrich22 Dec 2023 15:21 UTC
27 points
2 comments2 min readEA link
(osf.io)

How hu­man-like do safe AI mo­ti­va­tions need to be?

Joe_Carlsmith12 Nov 2025 5:33 UTC
26 points
1 comment52 min readEA link

In­tro­duc­ing StakeOut.AI

Harry Luk17 Feb 2024 0:21 UTC
52 points
6 comments9 min readEA link

Credo AI is hiring for AI Gov Re­searcher & more!

IanEisenberg15 Aug 2023 21:10 UTC
8 points
0 comments3 min readEA link

So You Want to Work at a Fron­tier AI Lab

Joe Rogero11 Jun 2025 23:11 UTC
36 points
2 comments7 min readEA link
(intelligence.org)

AI-Safety Mex­ico: A Pilot Sur­vey in Yu­catán.

Janeth Valdivia28 May 2025 23:19 UTC
5 points
1 comment5 min readEA link

#212 – Why tech­nol­ogy is un­stop­pable & how to shape AI de­vel­op­ment any­way (Allan Dafoe on The 80,000 Hours Pod­cast)

80000_Hours17 Feb 2025 16:38 UTC
16 points
0 comments19 min readEA link

Giv­ing AIs safe motivations

Joe_Carlsmith18 Aug 2025 18:02 UTC
22 points
1 comment51 min readEA link

The two-tiered society

Roman Leventov13 May 2024 7:53 UTC
14 points
5 comments3 min readEA link

Tony Blair In­sti­tute—Com­pute for AI In­dex ( Seek­ing a Sup­plier)

TomWestgarth3 Oct 2022 10:25 UTC
29 points
8 comments1 min readEA link

Jour­nal­ist’s in­quiry into a core or­ganiser break­ing his non­vi­o­lence com­mit­ment and leav­ing Stop AI

Remmelt5 Dec 2025 15:47 UTC
14 points
0 comments4 min readEA link
(www.theatlantic.com)

Emer­gency pod: Elon tries to crash OpenAI’s party (with Rose Chan Loui)

80000_Hours14 Feb 2025 16:29 UTC
21 points
0 comments2 min readEA link

Defin­ing AI “Rights” by Gemini

khayali8 Jun 2025 18:42 UTC
−4 points
0 comments32 min readEA link

The case for long-term cor­po­rate gov­er­nance of AI

SethBaum3 Nov 2021 10:50 UTC
42 points
3 comments8 min readEA link

New TIME mag­a­z­ine ar­ti­cle on the UK AI Safety In­sti­tute (AISI)

Rasool16 Jan 2025 22:51 UTC
10 points
0 comments1 min readEA link
(time.com)

Deep­Mind’s gen­er­al­ist AI, Gato: A non-tech­ni­cal explainer

frances_lorenz16 May 2022 21:19 UTC
128 points
13 comments6 min readEA link

The End of OpenAI’s Non­profit Era

Garrison29 Oct 2025 16:28 UTC
32 points
3 comments9 min readEA link
(www.obsolete.pub)

Prin­ci­ples for AI Welfare Research

jeffsebo19 Jun 2023 11:30 UTC
138 points
16 comments13 min readEA link

Let’s think about slow­ing down AI

Katja_Grace23 Dec 2022 19:56 UTC
339 points
9 comments38 min readEA link

[Linkpost] 80,000 Hours re­view: 2023 to mid-2025

80000_Hours15 Sep 2025 9:15 UTC
52 points
3 comments1 min readEA link
(80000hours.org)

The ul­ti­mate goal

Alvin Ånestrand6 Jul 2025 15:13 UTC
4 points
2 comments5 min readEA link
(forecastingaifutures.substack.com)

[Question] How con­fi­dent are you that it’s prefer­able for Amer­ica to de­velop AGI be­fore China does?

ScienceMon🔸22 Feb 2025 13:37 UTC
218 points
53 comments1 min readEA link

Public Opinion on AI Safety: AIMS 2023 and 2021 Summary

Janet Pauketat25 Sep 2023 18:09 UTC
19 points
0 comments3 min readEA link
(www.sentienceinstitute.org)

Stable Emer­gence in a Devel­op­men­tal AI Ar­chi­tec­ture: Re­sults from “Twins V3”

Petra Vojtassakova17 Nov 2025 23:23 UTC
6 points
2 comments2 min readEA link

Per­sonal agents

Roman Leventov17 Jun 2025 2:05 UTC
3 points
1 comment7 min readEA link

Why Mo­ral Weights Have Two Types and How to Mea­sure Them

Beyond Singularity17 Jul 2025 10:58 UTC
17 points
4 comments4 min readEA link

A Sketch of AI-Driven Epistemic Lock-In

Ozzie Gooen5 Mar 2025 22:40 UTC
15 points
1 comment3 min readEA link

AISN #60: The AI Ac­tion Plan

Center for AI Safety31 Jul 2025 18:10 UTC
6 points
0 comments7 min readEA link
(newsletter.safe.ai)

Global Risks Weekly Roundup #19/​2025: In­dia/​Pak­istan ceasefire, US/​China tar­iffs deal & OpenAI non­profit control

NunoSempere12 May 2025 17:11 UTC
16 points
0 comments1 min readEA link

AI Benefits Post 1: In­tro­duc­ing “AI Benefits”

Cullen 🔸22 Jun 2020 16:58 UTC
10 points
2 comments3 min readEA link

You Don’t Have to Be an AI Doomer to Sup­port AI Safety

Liam Robins14 Jun 2025 23:10 UTC
10 points
0 comments4 min readEA link
(thelimestack.substack.com)

Crash sce­nario 1: Rapidly mo­bil­ise for a 2025 AI crash

Remmelt11 Apr 2025 6:54 UTC
8 points
0 comments1 min readEA link

An Em­piri­cal De­mon­stra­tion of a New AI Catas­trophic Risk Fac­tor: Me­tapro­gram­matic Hijacking

Hiyagann27 Jun 2025 13:38 UTC
5 points
0 comments1 min readEA link

What the Head­lines Miss About the Lat­est De­ci­sion in the Musk vs. OpenAI Lawsuit

Garrison6 Mar 2025 19:49 UTC
87 points
9 comments6 min readEA link
(garrisonlovely.substack.com)

[Question] What would need to be true for AI to trans­late a le­gal con­tract to a smart con­tract?

Patrick Liu18 Mar 2023 16:42 UTC
−1 points
0 comments1 min readEA link

AI Safety Protest, Melbourne, Aus­tralia

Mark Brown17 Jan 2025 14:55 UTC
2 points
0 comments1 min readEA link

Spec­u­lat­ing on Se­cret In­tel­li­gence Explosions

calebp5 Jun 2025 13:55 UTC
21 points
5 comments8 min readEA link

Re­port on Semi-in­for­ma­tive Pri­ors for AI timelines (Open Philan­thropy)

Tom_Davidson26 Mar 2021 17:46 UTC
62 points
6 comments2 min readEA link

AI Gover­nance Read­ing Group [Toronto+re­mote]

Liav.Koren24 Jan 2023 22:05 UTC
2 points
0 comments1 min readEA link

6 Ways AI Can Harm You — and How to Stop It

Strad Slater18 Nov 2025 10:36 UTC
3 points
0 comments6 min readEA link
(williamslater2003.medium.com)

Yud­kowsky and Soares’ Book Is Empty

Oscar Davies5 Dec 2025 22:06 UTC
−4 points
3 comments7 min readEA link

[Question] Can we train AI so that fu­ture philan­thropy is more effec­tive?

Ricardo Pimentel3 Nov 2024 15:08 UTC
3 points
0 comments1 min readEA link

Pro­pos­als for the AI Reg­u­la­tory Sand­box in Spain

Guillem Bas27 Apr 2023 10:33 UTC
55 points
2 comments11 min readEA link
(riesgoscatastroficosglobales.com)

Par­allels Between AI Safety by De­bate and Ev­i­dence Law

Cullen 🔸20 Jul 2020 22:52 UTC
30 points
2 comments2 min readEA link
(cullenokeefe.com)

FLF Fel­low­ship on AI for Hu­man Rea­son­ing: $25-50k, 12 weeks

Oliver Sourbut19 May 2025 13:25 UTC
69 points
2 comments2 min readEA link
(www.flf.org)

FYI: I’m work­ing on a book about the threat of AGI/​ASI for a gen­eral au­di­ence. I hope it will be of value to the cause and the community

Darren McKee17 Jun 2022 11:52 UTC
32 points
1 comment2 min readEA link

An­thropic is Quietly Backpedal­ling on its Safety Commitments

Garrison23 May 2025 2:26 UTC
100 points
7 comments5 min readEA link
(www.obsolete.pub)

deleted

funnyfranco21 Mar 2025 13:13 UTC
11 points
0 comments1 min readEA link

Does Re­in­force­ment Learn­ing Really In­cen­tivize Rea­son­ing Ca­pac­ity in LLMs Beyond the Base Model?

Matrice Jacobine🔸🏳️‍⚧️24 Apr 2025 14:11 UTC
10 points
0 comments1 min readEA link
(limit-of-rlvr.github.io)

What Does an ASI Poli­ti­cal Ecol­ogy Mean for Hu­man Sur­vival?

Nathan Sidney23 Feb 2025 8:53 UTC
7 points
3 comments1 min readEA link

Slay­ing the Hy­dra: to­ward a new game board for AI

Prometheus23 Jun 2023 17:04 UTC
3 points
2 comments6 min readEA link

We won’t solve non-al­ign­ment prob­lems by do­ing research

MichaelDickens21 Nov 2025 18:03 UTC
51 points
1 comment4 min readEA link

Top AI safety newslet­ters, books, pod­casts, etc – new AISafety.com resource

Bryce Robertson4 Mar 2025 17:01 UTC
9 points
0 comments1 min readEA link

Hu­man­ity AI Com­mits $500 mil­lion to AI and Democ­racy Pro­tec­tion, AI x Se­cu­rity, and more

Peter14 Oct 2025 17:51 UTC
27 points
1 comment1 min readEA link
(www.macfound.org)

My Model of EA and AI Safety

Eva Lu24 Jun 2025 6:23 UTC
9 points
1 comment2 min readEA link

CoreWeave Is A Time Bomb

Remmelt31 Mar 2025 3:52 UTC
10 points
2 comments2 min readEA link
(www.wheresyoured.at)

AI Safety Endgame Stories

IvanVendrov28 Sep 2022 17:12 UTC
31 points
1 comment10 min readEA link

Align­ing AI Safety Pro­jects with a Repub­li­can Administration

Deric Cheng21 Nov 2024 22:13 UTC
13 points
1 comment8 min readEA link

Ori­ent­ing to 3 year AGI timelines

Nikola22 Dec 2024 23:07 UTC
122 points
18 comments8 min readEA link

Teach­ing AI to rea­son: this year’s most im­por­tant story

Benjamin_Todd13 Feb 2025 17:56 UTC
140 points
18 comments8 min readEA link
(benjamintodd.substack.com)

Ex­plor­ing AI Policy & the Fu­ture of Work — Seek­ing Guidance for PhD Path­ways (No UK/​EU/​US Pass­port, No Master’s)

genesis14 Jun 2025 21:47 UTC
2 points
0 comments1 min readEA link

[Question] Trade Between Altru­ists With Differ­ent AI Timelines?

Spiarrow18 Mar 2025 17:53 UTC
3 points
3 comments1 min readEA link

How not to lose your job to AI

80000_Hours1 Aug 2025 18:27 UTC
27 points
2 comments29 min readEA link

Fund­ing for hu­man­i­tar­ian non-prof­its to re­search re­spon­si­ble AI

Deborah W.A. Foulkes10 Dec 2024 8:08 UTC
4 points
0 comments2 min readEA link
(www.gov.uk)

Frac­tal Gover­nance: A Tractable, Ne­glected Ap­proach to Ex­is­ten­tial Risk Reduction

WillPearson5 Mar 2025 19:57 UTC
3 points
1 comment3 min readEA link

Con­sider this me drunk tex­ting the fo­rum: Is it use­ful to have data that can’t be touched by AI?

Jonas Søvik 🔹7 Feb 2025 21:52 UTC
−8 points
0 comments1 min readEA link

Things I Learned Mak­ing The SB-1047 Documentary

Michaël Trazzi12 May 2025 18:15 UTC
59 points
1 comment2 min readEA link

If The Data Is Poi­soned, Align­ment Won’t Save Us

keivn26 Sep 2025 17:59 UTC
1 point
0 comments3 min readEA link

AI com­pa­nies have started say­ing safe­guards are load-bearing

Zach Stein-Perlman27 Aug 2025 13:00 UTC
23 points
4 comments5 min readEA link

Ap­proaches to Miti­gat­ing AI Image-Gen­er­a­tion Risks through Regulation

scronkfinkle19 Apr 2025 13:50 UTC
1 point
0 comments4 min readEA link

OpenAI Does Not Ap­pear to be Ap­ply­ing Water­marks Honestly

scronkfinkle6 Nov 2025 21:25 UTC
14 points
0 comments2 min readEA link
(www.lesswrong.com)

The Con­ver­gent Path to the Stars—Similar Utility Across Civ­i­liza­tions Challenges Ex­tinc­tion Prioritization

Maxime Riché 🔸18 Mar 2025 17:09 UTC
8 points
1 comment20 min readEA link

(out­dated ver­sion) Vi­atopia and Buy-In

Jordan Arel21 Oct 2025 11:39 UTC
6 points
0 comments20 min readEA link

Could one coun­try out­grow the rest of the world?

Tom_Davidson21 Aug 2025 15:41 UTC
19 points
0 comments17 min readEA link

AI Dis­clo­sures: A Reg­u­la­tory Review

Elliot Mckernon29 Mar 2024 11:46 UTC
12 points
1 comment7 min readEA link

Time to Think about ASI Con­sti­tu­tions?

ukc1001427 Jan 2025 9:28 UTC
22 points
0 comments12 min readEA link

Euro­pean Union AI Devel­op­ment and Gover­nance Part­ner­ships

EU AI Governance19 Jan 2022 10:26 UTC
22 points
1 comment4 min readEA link

Should We Treat Open-Source AI Like Digi­tal Firearms? — A Draft Dec­la­ra­tion on the Eth­i­cal Limits of Fron­tier AI Models

DongHun Lee23 May 2025 8:58 UTC
−3 points
0 comments2 min readEA link

Deep­Mind: Model eval­u­a­tion for ex­treme risks

Zach Stein-Perlman25 May 2023 3:00 UTC
49 points
3 comments1 min readEA link
(arxiv.org)

Con­sider keep­ing your threat mod­els pri­vate.

Miles Kodama1 Feb 2025 0:29 UTC
17 points
2 comments4 min readEA link

[Op­por­tu­nity] Flour­ish­ing Fund­ing from the UK Government

Joey Bream🔸21 Nov 2025 12:45 UTC
30 points
4 comments2 min readEA link

Mid­dle Pow­ers in AI Gover­nance: Po­ten­tial paths to im­pact and re­lated ques­tions.

EffectiveAdvocate🔸15 Mar 2024 20:11 UTC
5 points
1 comment5 min readEA link

Where are the red lines for AI?

Karl von Wendt5 Aug 2022 9:41 UTC
13 points
3 comments6 min readEA link

OPEC for a slow AGI takeoff

vyrax21 Apr 2023 10:53 UTC
4 points
0 comments3 min readEA link

US AI Safety In­sti­tute will be ‘gut­ted,’ Ax­ios reports

Matrice Jacobine🔸🏳️‍⚧️20 Feb 2025 14:40 UTC
12 points
1 comment1 min readEA link
(www.zdnet.com)

[Linkpost] Longter­mists Are Push­ing a New Cold War With China

Radical Empath Ismam27 May 2023 6:53 UTC
38 points
16 comments1 min readEA link
(jacobin.com)

Scal­able And Trans­fer­able Black-Box Jailbreaks For Lan­guage Models Via Per­sona Modulation

sjp7 Nov 2023 18:00 UTC
10 points
0 comments2 min readEA link
(arxiv.org)

The V&V method—A step to­wards safer AGI

Yoav Hollander24 Jun 2025 15:57 UTC
1 point
0 comments1 min readEA link
(blog.foretellix.com)

deleted

funnyfranco24 Mar 2025 19:44 UTC
4 points
10 comments1 min readEA link

[Question] Am I tak­ing crazy pills? Why aren’t EAs ad­vo­cat­ing for a pause on AI ca­pa­bil­ities?

yanni kyriacos15 Aug 2023 23:29 UTC
18 points
21 comments1 min readEA link

How gov­ern­ments might re­spond to the eco­nomic up­heaval of TAI

Tax Geek17 Oct 2025 11:02 UTC
11 points
6 comments8 min readEA link

AI as a Con­sti­tu­tional Moment

atb28 May 2025 15:40 UTC
37 points
1 comment9 min readEA link

Linkpost: “Imag­in­ing and build­ing wise ma­chines: The cen­tral­ity of AI metacog­ni­tion” by John­son, Karimi, Ben­gio, et al.

Chris Leong17 Nov 2024 15:00 UTC
8 points
0 comments1 min readEA link
(arxiv.org)

How AI Takeover Might Hap­pen in Two Years

Joshc7 Feb 2025 23:51 UTC
35 points
7 comments29 min readEA link
(x.com)

The In­tel­li­gence Curse: an es­say series

L Rudolf L24 Apr 2025 12:59 UTC
22 points
1 comment2 min readEA link

Re­sources & op­por­tu­ni­ties for ca­reers in Euro­pean AI Policy

Cillian_12 Oct 2023 15:02 UTC
13 points
1 comment2 min readEA link

Come to Oxford/​Stan­ford to Work on De­cen­tral­ized AI Se­cu­rity (Funded)

samuelemarro8 Sep 2025 18:01 UTC
7 points
0 comments3 min readEA link

Agen­tic Align­ment: Nav­i­gat­ing be­tween Harm and Illegitimacy

LennardZ26 Nov 2024 21:27 UTC
2 points
1 comment9 min readEA link

Co­op­er­a­tion for AI safety must tran­scend geopoli­ti­cal interference

Matrice Jacobine🔸🏳️‍⚧️16 Feb 2025 18:18 UTC
9 points
0 comments1 min readEA link
(www.scmp.com)

Should AI fo­cus on prob­lem-solv­ing or strate­gic plan­ning? Why not both?

oliver_siegel1 Nov 2022 9:53 UTC
1 point
0 comments1 min readEA link

Give Neo a Chance

ank6 Mar 2025 14:35 UTC
1 point
3 comments7 min readEA link

High im­pact job op­por­tu­nity at ARIA (UK)

Rasool12 Feb 2023 10:35 UTC
83 points
0 comments1 min readEA link

Mo­ral Ed­u­ca­tion in the Age of AI: Are We Rais­ing Good Hu­mans?

Era Sarda31 Jul 2025 13:25 UTC
3 points
3 comments4 min readEA link

The fun­da­men­tal hu­man value is power.

Linyphia30 Mar 2023 15:15 UTC
−1 points
5 comments1 min readEA link

AI Safety Newslet­ter #42: New­som Ve­toes SB 1047 Plus, OpenAI’s o1, and AI Gover­nance Summary

Center for AI Safety1 Oct 2024 20:33 UTC
10 points
0 comments6 min readEA link
(newsletter.safe.ai)

Quotes on AI and wisdom

Chris Leong26 Nov 2025 15:55 UTC
8 points
0 comments2 min readEA link

What AI com­pa­nies should do: Some rough ideas

Zach Stein-Perlman21 Oct 2024 14:00 UTC
14 points
1 comment5 min readEA link

Scal­ing and Sus­tain­ing Stan­dards: A Case Study on the Basel Accords

C.K.16 Jul 2023 18:18 UTC
18 points
0 comments7 min readEA link
(docs.google.com)

AI gov­er­nance & China: Read­ing list

Zach Stein-Perlman18 Dec 2023 15:30 UTC
14 points
0 comments1 min readEA link
(docs.google.com)

Ac­tion­able-guidance and roadmap recom­men­da­tions for the NIST AI Risk Man­age­ment Framework

Tony Barrett17 May 2022 15:27 UTC
11 points
0 comments3 min readEA link

New pub­li­ca­tion “Com­pas­sion­ate Gover­nance” + launch webinar

jonleighton23 Jun 2025 13:16 UTC
9 points
0 comments1 min readEA link

Thoughts on re­spon­si­ble scal­ing poli­cies and regulation

Paul_Christiano24 Oct 2023 22:25 UTC
191 points
5 comments6 min readEA link

Emer­gency pod: Did OpenAI give up, or is this just a new trap? (with Rose Chan Loui)

80000_Hours9 May 2025 15:10 UTC
6 points
0 comments2 min readEA link

The day Elon Musk’s AI be­came a Nazi (and what it means for AI safety) | New video from AI in Context

ChanaMessinger2 Oct 2025 20:09 UTC
62 points
6 comments3 min readEA link

New AI safety treaty pa­per out!

Otto26 Mar 2025 9:28 UTC
28 points
2 comments4 min readEA link

Why ChatGPT Can’t Be Your Therapist

Strad Slater14 Nov 2025 10:07 UTC
13 points
0 comments4 min readEA link
(williamslater2003.medium.com)

[Question] Slow­ing down AI progress?

Eleni_A26 Jul 2022 8:46 UTC
16 points
9 comments1 min readEA link

Govern­ing High-Im­pact AI Sys­tems: Un­der­stand­ing Canada’s Pro­posed AI Bill. April 15, Car­leton Univer­sity, Ottawa

Liav.Koren27 Mar 2023 23:11 UTC
3 points
0 comments1 min readEA link
(www.eventbrite.com)

Five ne­glected work ar­eas that could re­duce AI risk

Aaron_Scher24 Sep 2023 2:09 UTC
22 points
0 comments9 min readEA link

Towards ev­i­dence gap-maps for AI safety

dEAsign25 Jul 2023 8:13 UTC
6 points
1 comment2 min readEA link

New Fund­ing Round on Hard­ware-En­abled Mechanisms (HEMs)

aog30 Apr 2025 17:45 UTC
54 points
0 comments15 min readEA link

A sur­vey of con­crete risks de­rived from Ar­tifi­cial Intelligence

Guillem Bas8 Jun 2023 22:09 UTC
36 points
2 comments6 min readEA link
(riesgoscatastroficosglobales.com)

Les­sons from the Iraq War for AI policy

Buck10 Jul 2025 18:52 UTC
71 points
11 comments4 min readEA link

Don’t Let Other Global Catas­trophic Risks Fall Be­hind: Sup­port ORCG in 2024

JorgeTorresC11 Nov 2024 18:27 UTC
48 points
1 comment4 min readEA link

De­mon­strate and eval­u­ate risks from AI to so­ciety at the AI x Democ­racy re­search hackathon

Esben Kran19 Apr 2024 14:46 UTC
24 points
0 comments6 min readEA link
(www.apartresearch.com)

Re­port: Ar­tifi­cial In­tel­li­gence Risk Man­age­ment in Spain

JorgeTorresC15 Jun 2023 16:08 UTC
22 points
0 comments3 min readEA link
(riesgoscatastroficosglobales.com)

Con­ver­gence 2024 Im­pact Review

David_Kristoffersson24 Mar 2025 20:28 UTC
39 points
0 comments14 min readEA link

A Bench­mark for Mea­sur­ing Hon­esty in AI Systems

Mantas Mazeika4 Mar 2025 17:44 UTC
29 points
0 comments2 min readEA link
(www.mask-benchmark.ai)

Is Eric Sch­midt fund­ing AI ca­pa­bil­ities re­search by the US gov­ern­ment?

Pranay K24 Dec 2022 8:32 UTC
46 points
3 comments2 min readEA link
(www.politico.com)

An­i­malHar­mBench 2.0: Eval­u­at­ing LLMs on rea­son­ing about an­i­mal welfare

Sentient Futures5 Nov 2025 1:13 UTC
43 points
4 comments6 min readEA link

Con­scious AI con­cerns all of us. [Con­scious AI & Public Per­cep­tions]

ixex3 Jul 2024 3:12 UTC
25 points
1 comment12 min readEA link

Drivers of large lan­guage model diffu­sion: in­cre­men­tal re­search, pub­lic­ity, and cascades

Ben Cottier21 Dec 2022 13:50 UTC
21 points
0 comments29 min readEA link

Ex­ec­u­tive Direc­tor for AIS Brus­sels—Ex­pres­sion of interest

gergo19 Dec 2024 9:15 UTC
29 points
0 comments4 min readEA link

In­tro­duc­ing Tech Gover­nance Pro­ject

Zakariyau Yusuf29 Oct 2024 9:20 UTC
52 points
5 comments8 min readEA link

My talk on AI risks at the Na­tional Con­ser­vatism con­fer­ence last week

Geoffrey Miller11 Sep 2025 16:30 UTC
83 points
9 comments10 min readEA link

AI Open Source De­bate Comes Down to Trust in In­sti­tu­tions, and AI Policy Mak­ers Should Con­sider How We Can Foster It

another-anon-do-gooder20 Jan 2024 13:47 UTC
6 points
2 comments1 min readEA link

Pivotal out­comes and pivotal processes

Andrew Critch17 Jun 2022 23:43 UTC
49 points
1 comment4 min readEA link

Adap­tive Com­pos­able Cog­ni­tive Core Unit (ACCCU)

Ihor Ivliev20 Mar 2025 21:48 UTC
10 points
2 comments4 min readEA link

A.I. Mo­ral Align­ment Kalei­do­scopic Com­pass Pro­posal: Philo­soph­i­cal and Tech­ni­cal Framework

Christopher Hunt Robertson, M.Ed.22 Nov 2025 13:52 UTC
1 point
0 comments11 min readEA link

When is AI safety re­search harm­ful?

Nathan_Barnard9 May 2022 10:36 UTC
13 points
6 comments9 min readEA link

Thoughts on SB-1047

Ryan Greenblatt30 May 2024 0:19 UTC
53 points
4 comments11 min readEA link

[Question] If an AI fi­nan­cial bub­ble popped, how much would that change your mind about near-term AGI?

Yarrow Bouchard 🔸21 Oct 2025 22:39 UTC
19 points
6 comments2 min readEA link

Credo AI is hiring for sev­eral roles

IanEisenberg11 Apr 2022 15:58 UTC
14 points
2 comments1 min readEA link

Ad­vice for Ac­tivists from the His­tory of Environmentalism

Jeffrey Heninger16 May 2024 20:36 UTC
48 points
2 comments6 min readEA link
(blog.aiimpacts.org)

The Ter­minol­ogy of Ar­tifi­cial Sentience

Janet Pauketat28 Nov 2021 7:52 UTC
29 points
0 comments1 min readEA link
(www.sentienceinstitute.org)

Holden Karnofsky on dozens of amaz­ing op­por­tu­ni­ties to make AI safer — and all his AGI takes

80000_Hours31 Oct 2025 12:13 UTC
70 points
0 comments25 min readEA link

AISN #63: Cal­ifor­nia’s SB-53 Passes the Leg­is­la­ture

Center for AI Safety24 Sep 2025 16:56 UTC
6 points
0 comments4 min readEA link
(newsletter.safe.ai)

NAIRA—An ex­er­cise in reg­u­la­tory, com­pet­i­tive safety gov­er­nance [AI Gover­nance In­sti­tu­tional De­sign idea]

Heramb Podar19 Mar 2024 14:55 UTC
5 points
1 comment6 min readEA link

AI Self-Mod­ifi­ca­tion Am­plifies Risks

Ihor Ivliev3 Jun 2025 20:27 UTC
0 points
0 comments2 min readEA link

AISN #59: EU Pub­lishes Gen­eral-Pur­pose AI Code of Prac­tice

Center for AI Safety15 Jul 2025 18:32 UTC
8 points
0 comments4 min readEA link
(aisafety.substack.com)

The Prac­ti­cal Value of Flawed Models: A Re­sponse to tito­tal’s AI 2027 Critique

Michelle_Ma25 Jun 2025 22:24 UTC
91 points
1 comment6 min readEA link

AI dis­course an­a­lyzed (we looked at es­says, Twit­ter, Bluesky, Truth So­cial)

Matt Brooks26 Nov 2025 16:02 UTC
68 points
5 comments9 min readEA link

From Ther­apy Tool to Align­ment Puz­zle-Piece: In­tro­duc­ing the VSPE Framework

Astelle Kay18 Jun 2025 14:47 UTC
6 points
1 comment2 min readEA link

From Cri­sis to Con­trol: Estab­lish­ing a Re­silient In­ci­dent Re­sponse Frame­work for De­ployed AI Models

KevinN31 Jan 2025 13:06 UTC
10 points
1 comment6 min readEA link
(www.techpolicy.press)

Sup­port­ing global co­or­di­na­tion in AI de­vel­op­ment: Why and how to con­tribute to in­ter­na­tional AI standards

pcihon17 Apr 2019 22:17 UTC
21 points
4 comments1 min readEA link

Cen­tre for the Study of Ex­is­ten­tial Risk Four Month Re­port June—Septem­ber 2020

HaydnBelfield2 Dec 2020 18:33 UTC
24 points
0 comments17 min readEA link

The Inevitable Emer­gence of Black-Mar­ket LLM Infrastructure

Tyler Williams8 Aug 2025 19:05 UTC
1 point
0 comments2 min readEA link

What is it to solve the al­ign­ment prob­lem?

Joe_Carlsmith13 Feb 2025 18:42 UTC
25 points
1 comment19 min readEA link
(joecarlsmith.substack.com)

AISN #56: Google Re­leases Veo 3

Center for AI Safety28 May 2025 15:57 UTC
6 points
0 comments4 min readEA link
(newsletter.safe.ai)

New Work­ing Paper Series of the Le­gal Pri­ori­ties Project

Legal Priorities Project18 Oct 2021 10:30 UTC
60 points
0 comments9 min readEA link

Why Most Efforts Towards “Demo­cratic AI” Fall Short

Jacob-Haimes29 Sep 2025 20:47 UTC
9 points
1 comment6 min readEA link
(www.odysseaninstitute.org)

Paths and waysta­tions in AI safety

Joe_Carlsmith11 Mar 2025 18:52 UTC
22 points
2 comments11 min readEA link
(joecarlsmith.substack.com)

Sum­mary: The Case for Halt­ing AI Devel­op­ment—Max Teg­mark on the Lex Frid­man Podcast

Madhav Malhotra16 Apr 2023 22:28 UTC
38 points
4 comments4 min readEA link
(youtu.be)

Ar­tifi­cial In­tel­li­gence and Nu­clear Com­mand, Con­trol, & Com­mu­ni­ca­tions: The Risks of Integration

Peter Rautenbach18 Nov 2022 13:01 UTC
62 points
3 comments50 min readEA link

Longter­mism and Global AI Gover­nance: Build­ing In­sti­tu­tional Readi­ness in the Global South

Adebayo Mubarak1 Oct 2025 23:54 UTC
14 points
2 comments4 min readEA link

Alt­man on the board, AGI, and superintelligence

OscarD🔸6 Jan 2025 14:37 UTC
20 points
1 comment1 min readEA link
(blog.samaltman.com)

The Se­cond Man­hat­tan: His­tor­i­cal Les­sons for AGI Control

Chiastic Slide13 Oct 2025 23:50 UTC
2 points
0 comments7 min readEA link

Hydra

Matrice Jacobine🔸🏳️‍⚧️11 Jun 2025 14:07 UTC
10 points
0 comments1 min readEA link
(philosophybear.substack.com)

An­nounc­ing the Fu­turekind Win­ter Fel­low­ship 2025/​6: Build­ing the Fu­ture of AI and An­i­mal Protection

Aditya_Karanam13 Oct 2025 11:15 UTC
7 points
0 comments4 min readEA link

Is Deep Learn­ing Ac­tu­ally Hit­ting a Wall? Eval­u­at­ing Ilya Sutskever’s Re­cent Claims

Garrison13 Nov 2024 17:00 UTC
121 points
8 comments8 min readEA link
(garrisonlovely.substack.com)

Patch­ing ~All Se­cu­rity-Rele­vant Open-Source Soft­ware?

niplav25 Feb 2025 21:35 UTC
35 points
7 comments2 min readEA link

The Ver­ifi­ca­tion Gap: A Scien­tific Warn­ing on the Limits of AI Safety

Ihor Ivliev24 Jun 2025 19:08 UTC
3 points
0 comments2 min readEA link

Sce­nario Map­ping Ad­vanced AI Risk: Re­quest for Par­ti­ci­pa­tion with Data Collection

Kiliank27 Mar 2022 11:44 UTC
14 points
0 comments5 min readEA link

Boot­strap­ping to viatopia

William_MacAskill13 Oct 2025 9:57 UTC
43 points
3 comments3 min readEA link

The Failed Strat­egy of Ar­tifi­cial In­tel­li­gence Doomers

yhoiseth5 Feb 2025 19:34 UTC
12 points
2 comments1 min readEA link
(letter.palladiummag.com)

Call on AI Com­pa­nies: Pub­lish Your Whistle­blow­ing Policies

Karl1 Aug 2025 15:59 UTC
11 points
0 comments6 min readEA link

AI Safety Overview: CERI Sum­mer Re­search Fellowship

Jamie B24 Mar 2022 15:12 UTC
29 points
0 comments2 min readEA link

“Nor­mal ac­ci­dents” and AI sys­tems

Eleni_A8 Aug 2022 18:43 UTC
5 points
1 comment1 min readEA link
(www.achan.ca)

[Question] AI Eth­i­cal Committee

eaaicommittee1 Mar 2022 23:35 UTC
8 points
0 comments1 min readEA link

[Question] What do we do if AI doesn’t take over the world, but still causes a sig­nifi­cant global prob­lem?

James_Banks2 Aug 2020 3:35 UTC
16 points
5 comments1 min readEA link

An­thropic is not be­ing con­sis­tently can­did about their con­nec­tion to EA

burner230 Mar 2025 13:30 UTC
312 points
88 comments2 min readEA link

[Question] Will OpenAI’s o3 re­duce NVi­dia’s moat?

Ebenezer Dukakis3 Jan 2025 2:21 UTC
9 points
6 comments1 min readEA link

Large Lan­guage Models as Cor­po­rate Lob­by­ists, and Im­pli­ca­tions for So­cietal-AI Alignment

johnjnay4 Jan 2023 22:22 UTC
10 points
6 comments8 min readEA link

Overview of Trans­for­ma­tive AI Mi­suse Risks

SammyDMartin11 Dec 2024 11:04 UTC
12 points
0 comments2 min readEA link
(longtermrisk.org)

Between Science Fic­tion and Emerg­ing Real­ity: Are We Ready for Digi­tal Per­sons?

Alex (Αλέξανδρος)13 Mar 2025 16:09 UTC
5 points
1 comment5 min readEA link

Pause from Be­hind /​ Los­ing Heroically

enterthewoods10 Nov 2025 17:11 UTC
9 points
4 comments5 min readEA link

Eigh­teen Open Re­search Ques­tions for Govern­ing Ad­vanced AI Systems

Ihor Ivliev3 May 2025 19:00 UTC
2 points
0 comments6 min readEA link

Against GDP as a met­ric for timelines and take­off speeds

kokotajlod29 Dec 2020 17:50 UTC
47 points
6 comments14 min readEA link

[Question] Is it eth­i­cal to work in AI “con­tent eval­u­a­tion”?

anon_databoy55530 Jan 2025 13:27 UTC
10 points
3 comments1 min readEA link

The cur­rent state of RSPs

Zach Stein-Perlman4 Nov 2024 16:00 UTC
19 points
1 comment9 min readEA link

The am­bigu­ous effect of full au­toma­tion + new goods on GDP growth

trammell7 Feb 2025 2:53 UTC
60 points
15 comments8 min readEA link

Eval­u­at­ing LLMs for Suicide Risk De­tec­tion: Can AI Catch a Cry for Help?

Nanda14 Oct 2025 19:13 UTC
3 points
1 comment17 min readEA link

The Bud­dhism & AI Initiative

ChrisScammell22 Aug 2025 15:49 UTC
23 points
0 comments2 min readEA link

Cap­ping AGI profits

Luke Frymire21 Mar 2023 13:29 UTC
12 points
2 comments4 min readEA link

A bet­ter “State­ment on AI Risk?” [Cross­post]

Knight Lee30 Dec 2024 7:36 UTC
4 points
0 comments3 min readEA link

Video and tran­script of talk on giv­ing AIs safe motivations

Joe_Carlsmith22 Sep 2025 16:47 UTC
10 points
1 comment50 min readEA link

Your AI Safety org could get EU fund­ing up to €9.08M. Here’s how (+ free per­son­al­ized sup­port) Up­date: We­bi­nar 18/​8 Link Below

SamuelK22 Jul 2025 17:06 UTC
16 points
0 comments3 min readEA link

(out­dated ver­sion) In­tro­duc­tion to Build­ing Co­op­er­a­tive Vi­atopia: The Case for Longter­mist In­fras­truc­ture Be­fore AI Builds Everything

Jordan Arel21 Oct 2025 11:26 UTC
6 points
0 comments18 min readEA link

Trans­for­ma­tive AI and Com­pute [Sum­mary]

lennart23 Sep 2021 13:53 UTC
65 points
5 comments9 min readEA link

In­ter­na­tional AI In­sti­tu­tions: a liter­a­ture re­view of mod­els, ex­am­ples, and pro­pos­als

MMMaas26 Sep 2023 15:26 UTC
53 points
0 comments2 min readEA link

Who owns AI-gen­er­ated con­tent?

Johan S Daniel7 Dec 2022 3:03 UTC
−2 points
0 comments2 min readEA link

AI Anal­y­sis of US H.R.1 (“Big Beau­tiful Bill”) Im­pacts on Farmed Animals

Steven Rouk22 Jul 2025 14:33 UTC
13 points
0 comments3 min readEA link

HARM: a fi­nan­cial liability

T. Johnson13 Nov 2025 19:39 UTC
6 points
2 comments2 min readEA link

AISN #58: Se­nate Re­moves State AI Reg­u­la­tion Moratorium

Center for AI Safety3 Jul 2025 17:07 UTC
6 points
0 comments4 min readEA link
(newsletter.safe.ai)

[Cross­post] Some Very Im­por­tant Things (That I Won’t Be Work­ing On This Year)

Sarah Cheng 🔸10 Mar 2025 14:42 UTC
28 points
1 comment4 min readEA link
(milesbrundage.substack.com)

Canada’s Se­nate is Study­ing AI’s Im­pact in Canada—How to Participate

Kiara 🔸9 Oct 2025 18:03 UTC
6 points
0 comments3 min readEA link

Is Paus­ing AI Pos­si­ble?

Richard Annilo9 Oct 2024 13:22 UTC
89 points
4 comments18 min readEA link

Protest­ing Now for AI Reg­u­la­tion might be more Im­pact­ful than AI Gover­nance and Policy Research

Nicolae13 Apr 2025 2:11 UTC
71 points
5 comments2 min readEA link

Beyond Short-Ter­mism: How δ and w Can Real­ign AI with Our Values

Beyond Singularity18 Jun 2025 16:34 UTC
15 points
8 comments5 min readEA link

Max Teg­mark — The AGI En­tente Delusion

Matrice Jacobine🔸🏳️‍⚧️13 Oct 2024 17:42 UTC
0 points
1 comment1 min readEA link
(www.lesswrong.com)

Miles Brundage re­signed from OpenAI, and his AGI readi­ness team was disbanded

Garrison23 Oct 2024 23:42 UTC
57 points
4 comments7 min readEA link
(garrisonlovely.substack.com)

ML4G is launch­ing its first ever Gover­nance boot­camp!

carolinaollive16 May 2025 15:22 UTC
25 points
0 comments1 min readEA link

Prevent­ing An­i­mal Suffer­ing Lock-in: Why Eco­nomic Tran­si­tions Matter

Karen Singleton28 Jul 2025 21:55 UTC
43 points
4 comments10 min readEA link

The Next Decades Will Plau­si­bly Be Com­pletely Insane

Bentham's Bulldog30 Nov 2025 18:43 UTC
14 points
3 comments14 min readEA link

Briefly how I’ve up­dated since ChatGPT

rime25 Apr 2023 19:39 UTC
29 points
8 comments2 min readEA link
(www.lesswrong.com)

Ex­clu­sive: Here’s the draft Trump ex­ec­u­tive or­der on AI preemption

Matrice Jacobine🔸🏳️‍⚧️19 Nov 2025 23:21 UTC
18 points
1 comment1 min readEA link
(www.transformernews.ai)

In­for­ma­tion se­cu­rity con­sid­er­a­tions for AI and the long term future

Jeffrey Ladish2 May 2022 20:53 UTC
134 points
8 comments11 min readEA link

Why I’m Post­ing AI-Safety-Re­lated Clips On TikTok

Michaël Trazzi12 Aug 2025 22:39 UTC
63 points
4 comments2 min readEA link

[Question] What are the most press­ing is­sues in short-term AI policy?

Eevee🔹14 Jan 2020 22:05 UTC
9 points
0 comments1 min readEA link

The GDM AGI Safety+Align­ment Team is Hiring for Ap­plied In­ter­pretabil­ity Research

Arthur Conmy25 Feb 2025 22:38 UTC
11 points
0 comments7 min readEA link

China Hawks are Man­u­fac­tur­ing an AI Arms Race

Garrison20 Nov 2024 18:17 UTC
103 points
3 comments5 min readEA link
(garrisonlovely.substack.com)

The cru­cible — how I think about the situ­a­tion with AI

Owen Cotton-Barratt5 May 2025 13:19 UTC
38 points
0 comments8 min readEA link
(strangecities.substack.com)

Linkpost: thor­ough in­tro to AI for an­i­mal advocates

Aidan Kankyoku4 Nov 2025 15:10 UTC
6 points
0 comments1 min readEA link
(sandcastlesblog.substack.com)

Overview of re­cent in­ter­na­tional demon­stra­tions against AI (AI Protest Ac­tions #1)

Rachel Shu17 Jul 2025 20:22 UTC
17 points
2 comments5 min readEA link

Ge­offrey Miller on Cross-Cul­tural Un­der­stand­ing Between China and Western Coun­tries as a Ne­glected Con­sid­er­a­tion in AI Alignment

Evan_Gaensbauer17 Apr 2023 3:26 UTC
25 points
2 comments4 min readEA link

[Question] Should AI writ­ers be pro­hibited in ed­u­ca­tion?

Eleni_A16 Jan 2023 22:29 UTC
3 points
3 comments1 min readEA link

A short con­ver­sa­tion I had with Google Gem­ini on the dan­gers of un­reg­u­lated LLM API use, while mildly drunk in an air­port.

EvanMcCormick17 Dec 2024 12:25 UTC
1 point
0 comments8 min readEA link

Con­fer­ence Re­port: Thresh­old 2030 - Model­ing AI Eco­nomic Futures

Deric Cheng24 Feb 2025 18:57 UTC
24 points
0 comments10 min readEA link
(www.convergenceanalysis.org)

New Fron­tiers in AI Safety

Hans Gundlach2 Apr 2025 2:00 UTC
6 points
0 comments4 min readEA link
(drive.google.com)

Promethean Gover­nance Un­leashed: Pilot­ing Poly­cen­tric, Memetic Orders in the AI Frontier

Paul Fallavollita21 Mar 2025 16:35 UTC
−11 points
1 comment3 min readEA link

CNAS re­port: ‘Ar­tifi­cial In­tel­li­gence and Arms Con­trol’

MMMaas13 Oct 2022 8:35 UTC
16 points
0 comments1 min readEA link
(www.cnas.org)

AISN #44: The Trump Cir­cle on AI Safety Plus, Chi­nese re­searchers used Llama to cre­ate a mil­i­tary tool for the PLA, a Google AI sys­tem dis­cov­ered a zero-day cy­ber­se­cu­rity vuln­er­a­bil­ity, and Com­plex Sys­tems

Center for AI Safety19 Nov 2024 16:36 UTC
11 points
0 comments5 min readEA link
(newsletter.safe.ai)

Bench­mark Scores = Gen­eral Ca­pa­bil­ity + Claudiness

Vasco Grilo🔸25 Nov 2025 17:58 UTC
19 points
0 comments4 min readEA link
(epochai.substack.com)

LLMs might not be the fu­ture of search: at least, not yet.

James-Hartree-Law22 Jan 2025 21:40 UTC
4 points
1 comment4 min readEA link

A So­ciety of Di­verse Cognition

atb9 Jun 2025 15:22 UTC
8 points
1 comment13 min readEA link

Im­prov­ing ca­pa­bil­ity eval­u­a­tions for AI gov­er­nance: Open Philan­thropy’s new re­quest for proposals

cb7 Feb 2025 9:30 UTC
37 points
3 comments3 min readEA link

Why doesn’t EA de­nounce OpenAI?

Leo_29 Nov 2025 20:50 UTC
2 points
10 comments1 min readEA link

AISN #65: Mea­sur­ing Au­toma­tion and Su­per­in­tel­li­gence Mo­ra­to­rium Let­ter

Center for AI Safety29 Oct 2025 16:08 UTC
8 points
0 comments3 min readEA link
(newsletter.safe.ai)

An In­ter­na­tional Col­lab­o­ra­tive Hub for Ad­vanc­ing AI Safety Research

Cody Albert22 Apr 2025 16:12 UTC
9 points
0 comments5 min readEA link

[Cross­post] An AI Pause Is Hu­man­ity’s Best Bet For Prevent­ing Ex­tinc­tion (TIME)

Otto24 Jul 2023 10:18 UTC
36 points
3 comments7 min readEA link
(time.com)

Democratis­ing AI Align­ment: Challenges and Proposals

Lloy2 🔹5 May 2025 14:50 UTC
2 points
2 comments4 min readEA link

Stu­dent com­pe­ti­tion for draft­ing a treaty on mora­to­rium of large-scale AI ca­pa­bil­ities R&D

Nayanika24 Apr 2023 13:15 UTC
36 points
4 comments2 min readEA link

Look­ing for Cana­dian sum­mer co-op po­si­tion in AI Governance

tcelferact26 Jun 2023 17:27 UTC
6 points
2 comments1 min readEA link

Nu­clear brinks­man­ship is not a good AI x-risk strategy

titotal30 Mar 2023 22:07 UTC
19 points
8 comments5 min readEA link

AGI Risk: How to in­ter­na­tion­ally reg­u­late in­dus­tries in non-democracies

Timothy_Liptrot16 May 2022 22:45 UTC
9 points
2 comments9 min readEA link

What is Com­pute? - Trans­for­ma­tive AI and Com­pute [1/​4]

lennart23 Sep 2021 13:54 UTC
48 points
5 comments18 min readEA link

Reflec­tive Align­ment Ar­chi­tec­ture (RAA): A Frame­work for Mo­ral Co­her­ence in AI Systems

Nicolas • EnlightenedAI Research Lab21 Nov 2025 22:05 UTC
1 point
0 comments2 min readEA link

In­side OpenAI’s Con­tro­ver­sial Plan to Aban­don its Non­profit Roots

Garrison18 Apr 2025 18:46 UTC
17 points
1 comment11 min readEA link
(garrisonlovely.substack.com)

Where I Am Donat­ing in 2025

MichaelDickens22 Nov 2025 23:21 UTC
89 points
9 comments14 min readEA link

He­len Toner on the geopoli­tics of AI in China and the Mid­dle East

80000_Hours5 Nov 2025 17:15 UTC
4 points
0 comments19 min readEA link

Tran­script: NBC Nightly News: AI ‘race to reck­less­ness’ w/​ Tris­tan Har­ris, Aza Raskin

WilliamKiely🔸23 Mar 2023 3:45 UTC
47 points
1 comment3 min readEA link

Fore­sight In­sti­tute launches two pos­si­ble fu­ture sce­nar­ios with AI

elte19 Aug 2025 14:29 UTC
44 points
4 comments2 min readEA link

Im­pli­ca­tions of large lan­guage model diffu­sion for AI governance

Ben Cottier21 Dec 2022 13:50 UTC
14 points
0 comments38 min readEA link

Sen­tience-Based Align­ment Strate­gies: Should we try to give AI gen­uine em­pa­thy/​com­pas­sion?

Lloy2 🔹4 May 2025 20:45 UTC
16 points
1 comment3 min readEA link

Launch­ing The Col­lec­tive In­tel­li­gence Pro­ject: Whitepa­per and Pilots

jasmine_wang6 Feb 2023 17:00 UTC
38 points
8 comments2 min readEA link
(cip.org)

Third-wave AI safety needs so­ciopoli­ti­cal thinking

richard_ngo27 Mar 2025 0:55 UTC
112 points
59 comments26 min readEA link

AI Alter­na­tive Fu­tures: Ex­plo­ra­tory Sce­nario Map­ping for Ar­tifi­cial In­tel­li­gence Risk—Re­quest for Par­ti­ci­pa­tion [Linkpost]

Kiliank9 May 2022 19:53 UTC
17 points
2 comments8 min readEA link

AI Gover­nance to Avoid Ex­tinc­tion: The Strate­gic Land­scape and Ac­tion­able Re­search Ques­tions [MIRI TGT Re­search Agenda]

peterbarnett5 May 2025 19:13 UTC
67 points
1 comment8 min readEA link
(techgov.intelligence.org)

4 Les­sons From An­thropic on Scal­ing In­ter­pretabil­ity Research

Strad Slater29 Nov 2025 11:22 UTC
4 points
0 comments4 min readEA link
(williamslater2003.medium.com)

Con­sid­er­a­tions around ca­reer costs of poli­ti­cal donations

Zach Stein-Perlman20 Oct 2025 13:10 UTC
42 points
0 comments16 min readEA link
(www.lesswrong.com)

[Question] Does China have AI al­ign­ment re­sources/​in­sti­tu­tions? How can we pri­ori­tize cre­at­ing more?

JakubK4 Aug 2022 19:23 UTC
18 points
9 comments1 min readEA link

AGI al­ign­ment re­sults from a se­ries of al­igned ac­tions

hanadulset27 Dec 2021 19:33 UTC
15 points
1 comment6 min readEA link

Re­think Pri­ori­ties is look­ing for a (Co-)Founder for a New Pro­ject: Field Build­ing in Univer­si­ties for AI Policy Ca­reers in the US

KevinN28 Aug 2023 16:01 UTC
59 points
0 comments6 min readEA link
(careers.rethinkpriorities.org)

Promethean Gover­nance Tested: Re­silience and Re­con­figu­ra­tion Amidst AI Re­bel­lion and Memetic Fragmentation

Paul Fallavollita24 Mar 2025 11:08 UTC
−12 points
0 comments4 min readEA link

Cryp­tocur­rency Ex­ploits Show the Im­por­tance of Proac­tive Poli­cies for AI X-Risk

eSpencer16 Sep 2022 4:44 UTC
14 points
1 comment4 min readEA link

The Risks of AI-Gen­er­ated Con­tent on the EA Forum

WobblyPanda24 Jun 2023 5:33 UTC
−1 points
0 comments1 min readEA link

‘Surveillance Cap­i­tal­ism’ & AI Gover­nance: Slip­pery Busi­ness Models, Se­cu­ri­ti­sa­tion, and Self-Regulation

Charlie Harrison29 Feb 2024 15:47 UTC
19 points
2 comments12 min readEA link

Please Donate to CAIP (Post 1 of 7 on AI Gover­nance)

Jason Green-Lowe7 May 2025 18:15 UTC
141 points
25 comments33 min readEA link

Prevent­ing AI Safety Politicization

closetsolipsist20 Oct 2025 3:58 UTC
8 points
0 comments5 min readEA link

Ap­ply to CEEALAR to do AGI mora­to­rium work

Greg_Colbourn ⏸️ 26 Jul 2023 21:24 UTC
62 points
0 comments1 min readEA link

An­nounc­ing the SPT Model Web App for AI Governance

Paolo Bova4 Aug 2022 10:45 UTC
42 points
0 comments5 min readEA link

Safety with­out op­pres­sion: an AI gov­er­nance problem

Nathan_Barnard28 Jul 2022 10:19 UTC
3 points
0 comments8 min readEA link

[linkpost] “What Are Rea­son­able AI Fears?” by Robin Han­son, 2023-04-23

Arjun Panickssery14 Apr 2023 23:26 UTC
41 points
3 comments4 min readEA link
(quillette.com)

“Long” timelines to ad­vanced AI have got­ten crazy short

Matrice Jacobine🔸🏳️‍⚧️3 Apr 2025 22:46 UTC
16 points
1 comment1 min readEA link
(helentoner.substack.com)

[Question] How should tech­ni­cal AI re­searchers best tran­si­tion into AI gov­er­nance and policy?

GabeM10 Sep 2023 5:29 UTC
12 points
5 comments1 min readEA link

New 80,000 Hours prob­lem pro­file on the risks of power-seek­ing AI

Zershaaneh Qureshi28 Oct 2025 14:37 UTC
45 points
0 comments2 min readEA link

What are Re­spon­si­ble Scal­ing Poli­cies (RSPs)?

Vishakha Agrawal5 Apr 2025 16:05 UTC
2 points
0 comments2 min readEA link
(www.lesswrong.com)

What term to use for AI in differ­ent policy con­texts?

oeg6 Sep 2023 15:08 UTC
18 points
3 comments9 min readEA link

#217 – The most im­por­tant graph in AI right now (Beth Barnes on The 80,000 Hours Pod­cast)

80000_Hours2 Jun 2025 16:52 UTC
16 points
1 comment26 min readEA link

The OpenAI Gover­nance Tran­si­tion: The His­tory, What It Is, and What It Means

Charlie G 🔹17 Nov 2025 21:08 UTC
3 points
0 comments9 min readEA link
(charliegarfield.substack.com)

Mas­sive Scal­ing Should be Frowned Upon

harsimony17 Nov 2022 17:44 UTC
9 points
0 comments5 min readEA link

Challenges from Ca­reer Tran­si­tions and What To Ex­pect From Advising

ClaireB24 Jul 2025 13:22 UTC
26 points
1 comment9 min readEA link

Forg­ing A New AGI So­cial Contract

Deric Cheng10 Apr 2025 13:41 UTC
13 points
3 comments7 min readEA link
(agisocialcontract.substack.com)

Which side of the AI safety com­mu­nity are you in?

Greg_Colbourn ⏸️ 23 Oct 2025 14:23 UTC
12 points
2 comments2 min readEA link
(www.lesswrong.com)

An EA used de­cep­tive mes­sag­ing to ad­vance her pro­ject; we need mechanisms to avoid de­on­tolog­i­cally du­bi­ous plans

MikhailSamin13 Feb 2024 23:11 UTC
19 points
39 comments5 min readEA link

Lon­don Work­ing Group for Short/​Medium Term AI Risks

scronkfinkle7 Apr 2025 15:30 UTC
5 points
0 comments2 min readEA link

What if AI de­vel­op­ment goes well?

RoryG3 Aug 2022 8:57 UTC
25 points
7 comments12 min readEA link

ML4Good Brasil—Ap­pli­ca­tions Open

Nia🔸3 May 2024 10:39 UTC
28 points
1 comment1 min readEA link

World Ci­ti­zen Assem­bly about AI—Announcement

Camille11 Feb 2025 10:51 UTC
25 points
2 comments5 min readEA link

Google could build a con­scious AI in three months

Derek Shiller1 Oct 2022 13:24 UTC
16 points
22 comments7 min readEA link

On Ar­tifi­cial Gen­eral In­tel­li­gence: Ask­ing the Right Questions

Heather Douglas2 Oct 2022 5:00 UTC
−1 points
7 comments3 min readEA link

Peace Treaty Ar­chi­tec­ture (PTA) as an Alter­na­tive to AI Alignment

Andrei Navrotskii11 Nov 2025 22:11 UTC
1 point
0 comments15 min readEA link

[Opz­ionale] Il panorama della gov­er­nance lun­goter­minista delle in­tel­li­genze artificiali

EA Italy17 Jan 2023 11:03 UTC
1 point
0 comments10 min readEA link

Sys­tem Level Safety Evaluations

markov29 Sep 2025 13:55 UTC
3 points
0 comments9 min readEA link
(equilibria1.substack.com)

Meta: Fron­tier AI Framework

Zach Stein-Perlman3 Feb 2025 22:00 UTC
23 points
0 comments1 min readEA link
(ai.meta.com)

[Question] Can we ever en­sure AI al­ign­ment if we can only test AI per­sonas?

Karl von Wendt16 Mar 2025 8:06 UTC
8 points
0 comments1 min readEA link

From vol­un­tary to manda­tory, are the ESG dis­clo­sure frame­works still fer­tile ground for un­re­al­ised EA ca­reer path­ways? – A 2023 up­date on ESG po­ten­tial impact

Christopher Chan 🔸4 Jun 2023 12:00 UTC
21 points
5 comments11 min readEA link

AI-en­abled coups: a small group could use AI to seize power

Tom_Davidson16 Apr 2025 16:51 UTC
123 points
1 comment7 min readEA link

How Well Does RL Scale?

Toby_Ord22 Oct 2025 13:16 UTC
142 points
50 comments7 min readEA link

Some ini­tial mus­ing on the poli­tics of longter­mist tra­jec­tory change

GideonF26 Jun 2025 7:16 UTC
6 points
0 comments12 min readEA link
(futerman.substack.com)

Con­trol­ling the op­tions AIs can pursue

Joe_Carlsmith29 Sep 2025 17:24 UTC
9 points
0 comments35 min readEA link

Sce­nario plan­ning for AI x-risk

Corin Katzke10 Feb 2024 0:07 UTC
41 points
0 comments15 min readEA link
(www.convergenceanalysis.org)

An­thropic Faces Po­ten­tially “Busi­ness-End­ing” Copy­right Lawsuit

Garrison25 Jul 2025 17:01 UTC
31 points
10 comments9 min readEA link
(www.obsolete.pub)

What AI could mean for al­ter­na­tive proteins

Max Taylor9 Feb 2024 10:13 UTC
37 points
5 comments16 min readEA link

The AGI Awak­e­ness valley of doom and three path­ways to slowing

GideonF28 Jul 2025 18:46 UTC
16 points
0 comments16 min readEA link
(open.substack.com)

Global Risks Weekly Roundup #18/​2025: US tar­iff short­ages, mil­i­tary polic­ing, Gaza famine.

NunoSempere6 May 2025 10:39 UTC
22 points
0 comments3 min readEA link
(blog.sentinel-team.org)

Promethean Gover­nance in Prac­tice: Craft­ing a Poly­cen­tric, Memetic Order for the Mul­tipo­lar AI Aeon

Paul Fallavollita20 Mar 2025 10:10 UTC
−1 points
0 comments4 min readEA link

Ques­tions for fur­ther in­ves­ti­ga­tion of AI diffusion

Ben Cottier21 Dec 2022 13:50 UTC
28 points
0 comments11 min readEA link

One, per­haps un­der­rated, AI risk.

Alex (Αλέξανδρος)28 Nov 2024 10:34 UTC
7 points
1 comment3 min readEA link

Nu­clear Es­pi­onage and AI Governance

GAA4 Oct 2021 18:21 UTC
32 points
3 comments24 min readEA link

Idea: an AI gov­er­nance group colo­cated with ev­ery AI re­search group!

capybaralet7 Dec 2020 23:41 UTC
8 points
1 comment2 min readEA link

[Question] Is AI like disk drives?

Tanae2 Sep 2023 19:12 UTC
8 points
1 comment1 min readEA link

Poll: To ad­dress risks from AI, Li­a­bil­ity or Reg­u­la­tion?

TFD30 Apr 2025 22:03 UTC
6 points
0 comments1 min readEA link

Evolv­ing OpenAI’s Structure

Tyner🔸6 May 2025 0:52 UTC
12 points
1 comment1 min readEA link

Is Text Water­mark­ing a lost cause?

Egor Timatkov1 Oct 2024 13:07 UTC
7 points
0 comments10 min readEA link

Con­scious AI & Public Per­cep­tion: Four futures

nicoleta-k3 Jul 2024 23:06 UTC
12 points
1 comment16 min readEA link

Very Briefly: The CHIPS Act

Yadav26 Feb 2023 13:53 UTC
40 points
3 comments1 min readEA link
(www.y1d2.com)

[linkpost] AI NOW In­sti­tute’s 2023 An­nual Re­port & Roadmap

Tristan W12 Apr 2023 20:00 UTC
9 points
0 comments2 min readEA link
(ainowinstitute.org)

A Re­ply to MacAskill on “If Any­one Builds It, Every­one Dies”

RobBensinger27 Sep 2025 23:03 UTC
9 points
7 comments17 min readEA link

An Open Let­ter To EA and AI Safety On De­cel­er­at­ing AI Development

Kenneth_Diao28 Feb 2025 17:15 UTC
21 points
0 comments14 min readEA link
(graspingatwaves.substack.com)

Sha­har Avin on How to Strate­gi­cally Reg­u­late Ad­vanced AI Systems

Michaël Trazzi23 Sep 2022 15:49 UTC
48 points
2 comments4 min readEA link
(theinsideview.ai)

Learn­ing from the Lud­dites: Im­pli­ca­tions for a mod­ern AI labour movement

Jan Wehner🔸16 Oct 2025 17:11 UTC
5 points
1 comment8 min readEA link

Me­tac­u­lus Pre­sents — View From the En­ter­prise Suite: How Ap­plied AI Gover­nance Works Today

christian20 Jun 2023 22:24 UTC
4 points
0 comments1 min readEA link

[Cross­post] AI Reg­u­la­tion May Be More Im­por­tant Than AI Align­ment For Ex­is­ten­tial Safety

Otto24 Aug 2023 16:01 UTC
14 points
2 comments5 min readEA link

The case for con­scious AI: Clear­ing the record [AI Con­scious­ness & Public Per­cep­tion]

Jay Luong5 Jul 2024 20:29 UTC
3 points
7 comments8 min readEA link

AI Risk in Africa

Claude Formanek12 Oct 2021 2:28 UTC
20 points
0 comments10 min readEA link

AI Lab Re­tal­i­a­tion: A Sur­vival Guide

Jay Ready4 Jan 2025 23:05 UTC
8 points
1 comment12 min readEA link
(morelightinai.substack.com)

List of Masters Pro­grams in Tech Policy, Public Policy and Se­cu­rity (Europe)

sberg29 May 2023 10:23 UTC
49 points
0 comments3 min readEA link

AI and Chem­i­cal, Biolog­i­cal, Ra­diolog­i­cal, & Nu­clear Hazards: A Reg­u­la­tory Review

Elliot Mckernon10 May 2024 8:41 UTC
8 points
1 comment10 min readEA link

Why You Should Think About What You Tell ChatGPT : Data Pri­vacy In A World With AI

Strad Slater17 Nov 2025 10:15 UTC
2 points
1 comment5 min readEA link
(williamslater2003.medium.com)

Re­view­ing the Struc­ture of Cur­rent AI Regulations

Deric Cheng7 May 2024 12:34 UTC
32 points
1 comment13 min readEA link

Align­ment Fak­ing in Large Lan­guage Models

Ryan Greenblatt18 Dec 2024 17:19 UTC
142 points
9 comments10 min readEA link

Open call: AI Act Stan­dard for Dev. Phase Risk Assess­ment

miller-max8 Dec 2023 19:57 UTC
5 points
1 comment1 min readEA link

Will AI R&D Au­toma­tion Cause a Soft­ware In­tel­li­gence Ex­plo­sion?

Forethought26 Mar 2025 15:37 UTC
32 points
4 comments2 min readEA link
(www.forethought.org)

EU’s AI am­bi­tions at risk as US pushes to wa­ter down in­ter­na­tional treaty (linkpost)

mic31 Jul 2023 0:34 UTC
9 points
0 comments4 min readEA link
(www.euractiv.com)

Bias In, Bias Out: Ar­tifi­cial In­tel­li­gence Reflects Real Discrimination

Nanda16 Oct 2025 23:36 UTC
10 points
0 comments7 min readEA link

De­ci­sion Eng­ine For Model­ling AI in Society

Echo Huang7 Aug 2025 11:15 UTC
24 points
1 comment18 min readEA link

Op­por­tu­ni­ties for Im­pact Beyond the EU AI Act

Cillian_12 Oct 2023 15:06 UTC
27 points
2 comments4 min readEA link

Sur­vey: How Do Elite Chi­nese Stu­dents Feel About the Risks of AI?

Nick Corvino2 Sep 2024 9:14 UTC
110 points
9 comments10 min readEA link

Some AI safety pro­ject & re­search ideas/​ques­tions for short and long timelines

Lloy2 🔹8 Aug 2025 21:08 UTC
13 points
0 comments5 min readEA link

Forethought: A new AI macros­trat­egy group

Amrit Sidhu-Brar 🔸11 Mar 2025 15:36 UTC
174 points
10 comments3 min readEA link

Notes on my tran­si­tion to civic tech

danielechlin5 Oct 2025 15:50 UTC
3 points
0 comments2 min readEA link

How to ‘troll for good’: Lev­er­ag­ing IP for AI governance

Michael Huang26 Feb 2023 6:34 UTC
26 points
3 comments1 min readEA link
(www.science.org)

Risk Align­ment in Agen­tic AI Systems

Hayley Clatterbuck1 Oct 2024 22:51 UTC
32 points
1 comment3 min readEA link
(static1.squarespace.com)

What if we just…didn’t build AGI? An Ar­gu­ment Against Inevitability

Nate Sharpe10 May 2025 3:34 UTC
64 points
21 comments14 min readEA link
(natezsharpe.substack.com)

Elec­tion by Jury: The Ul­ti­mate Demo­cratic Safe­guard in the Age of AI and In­for­ma­tion Warfare

ClayShentrup17 Apr 2025 19:50 UTC
13 points
5 comments27 min readEA link