RSS

Public com­mu­ni­ca­tion on AI safety

TagLast edit: Nov 27, 2023, 10:13 PM by Sarah Cheng

The public communication on AI safety tag covers depictions of AI safety in media as well as meta-discussions regarding the most effective ways to represent the topic when engaging with journalists or publishing content for a broader audience, and posts that discuss ways to convey AI safety to various audiences.

Related entries

AI Safety | Building the field of AI safety | AI governance | Slowing down AI

Anal­ogy Bank for AI Safety

utilistrutilJan 29, 2024, 2:35 AM
14 points
5 comments1 min readEA link

“Near Mid­night in Suicide City”

Greg_ColbournDec 6, 2024, 7:54 PM
5 points
0 comments1 min readEA link
(www.youtube.com)

Care­less talk on US-China AI com­pe­ti­tion? (and crit­i­cism of CAIS cov­er­age)

Oliver SourbutSep 20, 2023, 12:46 PM
52 points
19 comments1 min readEA link
(www.oliversourbut.net)

Talk­ing pub­li­cly about AI risk

Jan_KulveitApr 24, 2023, 9:19 AM
152 points
13 comments1 min readEA link

If try­ing to com­mu­ni­cate about AI risks, make it vivid

Michael Noetel 🔸May 27, 2024, 12:59 AM
19 points
2 comments2 min readEA link

New blog: Planned Obsolescence

AjeyaMar 27, 2023, 7:46 PM
198 points
9 comments1 min readEA link
(www.planned-obsolescence.org)

Why some peo­ple dis­agree with the CAIS state­ment on AI

David_MossAug 15, 2023, 1:39 PM
144 points
15 comments16 min readEA link

AI Safety Ac­tion Plan—A re­port com­mis­sioned by the US State Department

Agustín Covarrubias 🔸Mar 11, 2024, 10:13 PM
25 points
1 comment1 min readEA link
(www.gladstone.ai)

Ar­ti­cles about re­cent OpenAI departures

bruceMay 17, 2024, 5:38 PM
126 points
12 comments1 min readEA link
(www.vox.com)

Wor­ri­some mi­s­un­der­stand­ing of the core is­sues with AI transition

Roman LeventovJan 18, 2024, 10:05 AM
4 points
3 comments1 min readEA link

My Proven AI Safety Ex­pla­na­tion (as a com­put­ing stu­dent)

Mica WhiteFeb 6, 2024, 3:58 AM
8 points
4 comments6 min readEA link

New vol­un­tary com­mit­ments (AI Seoul Sum­mit)

Zach Stein-PerlmanMay 21, 2024, 11:00 AM
12 points
1 comment1 min readEA link
(www.gov.uk)

[Linkpost] State­ment from Scar­lett Jo­hans­son on OpenAI’s use of the “Sky” voice, that was shock­ingly similar to her own voice.

LinchMay 20, 2024, 11:50 PM
46 points
8 comments1 min readEA link
(variety.com)

World’s first ma­jor law for ar­tifi­cial in­tel­li­gence gets fi­nal EU green light

Dane ValerieMay 24, 2024, 2:57 PM
3 points
1 comment2 min readEA link
(www.cnbc.com)

US pub­lic opinion of AI policy and risk

Jamie EMay 12, 2023, 1:22 PM
111 points
7 comments15 min readEA link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

EliezerYudkowskyApr 9, 2023, 3:53 PM
50 points
3 comments1 min readEA link

Sur­vey of 2,778 AI au­thors: six parts in pictures

Katja_GraceJan 6, 2024, 4:43 AM
176 points
10 comments1 min readEA link

Is fear pro­duc­tive when com­mu­ni­cat­ing AI x-risk? [Study re­sults]

Johanna RonigerJan 22, 2024, 5:38 AM
73 points
10 comments5 min readEA link

Keep Mak­ing AI Safety News

GilMar 31, 2023, 8:11 PM
67 points
4 comments1 min readEA link

Spread­ing mes­sages to help with the most im­por­tant century

Holden KarnofskyJan 25, 2023, 8:35 PM
128 points
21 comments18 min readEA link
(www.cold-takes.com)

AI Align­ment in The New Yorker

Eleni_AMay 17, 2023, 9:19 PM
23 points
0 comments1 min readEA link
(www.newyorker.com)

I de­signed an AI safety course (for a philos­o­phy de­part­ment)

Eleni_ASep 23, 2023, 9:56 PM
27 points
3 comments2 min readEA link

Jan Leike: “I’m ex­cited to join @An­throp­icAI to con­tinue the su­per­al­ign­ment mis­sion!”

defun 🔸May 28, 2024, 6:08 PM
35 points
11 comments1 min readEA link
(x.com)

My cover story in Ja­cobin on AI cap­i­tal­ism and the x-risk debates

GarrisonFeb 12, 2024, 11:34 PM
154 points
10 comments6 min readEA link
(jacobin.com)

US pub­lic per­cep­tion of CAIS state­ment and the risk of extinction

Jamie EJun 22, 2023, 4:39 PM
126 points
4 comments9 min readEA link

xAI raises $6B

andzuckJun 5, 2024, 3:26 PM
18 points
1 comment1 min readEA link
(x.ai)

A short con­ver­sa­tion I had with Google Gem­ini on the dan­gers of un­reg­u­lated LLM API use, while mildly drunk in an air­port.

EvanMcCormickDec 17, 2024, 12:25 PM
1 point
0 comments8 min readEA link

AI Risk is like Ter­mi­na­tor; Stop Say­ing it’s Not

skluugMar 8, 2022, 7:17 PM
191 points
43 comments10 min readEA link
(skluug.substack.com)

The Over­ton Win­dow widens: Ex­am­ples of AI risk in the media

AkashMar 23, 2023, 5:10 PM
112 points
11 comments1 min readEA link

Keep Chas­ing AI Safety Press Coverage

GilApr 4, 2023, 8:40 PM
106 points
16 comments5 min readEA link

FLI open let­ter: Pause gi­ant AI experiments

Zach Stein-PerlmanMar 29, 2023, 4:04 AM
220 points
38 comments1 min readEA link

Max Teg­mark’s new Time ar­ti­cle on how we’re in a Don’t Look Up sce­nario [Linkpost]

Jonas HallgrenApr 25, 2023, 3:47 PM
41 points
0 comments1 min readEA link

Claude 3.5 Sonnet

Zach Stein-PerlmanJun 20, 2024, 6:00 PM
31 points
0 comments1 min readEA link
(www.anthropic.com)

FT: We must slow down the race to God-like AI

Angelina LiApr 24, 2023, 11:57 AM
33 points
2 comments2 min readEA link
(www.ft.com)

Let’s think about slow­ing down AI

Katja_GraceDec 23, 2022, 7:56 PM
334 points
9 comments1 min readEA link

[Linkpost] OpenAI lead­ers call for reg­u­la­tion of “su­per­in­tel­li­gence” to re­duce ex­is­ten­tial risk.

Lowe LundinMay 25, 2023, 2:14 PM
5 points
0 comments1 min readEA link

AI Safety Newslet­ter #3: AI policy pro­pos­als and a new challenger approaches

Oliver ZApr 25, 2023, 4:15 PM
35 points
1 comment4 min readEA link
(newsletter.safe.ai)

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibsMar 29, 2023, 11:30 PM
211 points
75 comments3 min readEA link
(time.com)

Or­ga­niz­ing a de­bate with ex­perts and MPs to raise AI xrisk aware­ness: a pos­si­ble blueprint

OttoApr 19, 2023, 10:50 AM
75 points
1 comment4 min readEA link

[US] NTIA: AI Ac­countabil­ity Policy Re­quest for Comment

Kyle J. LuccheseApr 13, 2023, 4:12 PM
47 points
4 comments1 min readEA link
(ntia.gov)

Tar­bell Fel­low­ship 2024 - Ap­pli­ca­tions Open (AI Jour­nal­ism)

Cillian_Sep 28, 2023, 10:38 AM
58 points
1 comment3 min readEA link

How bad a fu­ture do ML re­searchers ex­pect?

Katja_GraceMar 13, 2023, 5:47 AM
165 points
20 comments1 min readEA link

News: Span­ish AI image out­cry + US AI work­force “reg­u­la­tion”

Benevolent_RainSep 26, 2023, 7:43 AM
9 points
0 comments1 min readEA link

INTERVIEW: StakeOut.AI w/​ Dr. Peter Park

Jacob-HaimesMar 5, 2024, 6:04 PM
21 points
7 comments1 min readEA link
(into-ai-safety.github.io)

[Linkpost] 538 Poli­tics Pod­cast on AI risk & politics

jackvaApr 11, 2023, 5:03 PM
64 points
5 comments1 min readEA link
(fivethirtyeight.com)

Short re­view of our Ten­sorTrust-based AI safety uni­ver­sity out­reach event

Milan Weibel🔹Sep 22, 2024, 2:54 PM
15 points
0 comments2 min readEA link

A tran­script of the TED talk by Eliezer Yudkowsky

MikhailSaminJul 12, 2023, 12:12 PM
39 points
0 comments1 min readEA link

Against most, but not all, AI risk analogies

Matthew_BarnettJan 14, 2024, 7:13 PM
43 points
9 comments1 min readEA link

Did Ben­gio and Teg­mark lose a de­bate about AI x-risk against LeCun and Mitchell?

Karl von WendtJun 25, 2023, 4:59 PM
80 points
24 comments1 min readEA link

Ex­cerpts from “Ma­jor­ity Leader Schumer De­liv­ers Re­marks To Launch SAFE In­no­va­tion Frame­work For Ar­tifi­cial In­tel­li­gence At CSIS”

Chris LeongJul 21, 2023, 11:15 PM
19 points
0 comments1 min readEA link
(www.democrats.senate.gov)

An EA used de­cep­tive mes­sag­ing to ad­vance her pro­ject; we need mechanisms to avoid de­on­tolog­i­cally du­bi­ous plans

MikhailSaminFeb 13, 2024, 11:11 PM
22 points
39 comments5 min readEA link

AI policy ideas: Read­ing list

Zach Stein-PerlmanApr 17, 2023, 7:00 PM
60 points
3 comments1 min readEA link

The Best Ar­gu­ment is not a Sim­ple English Yud Essay

Jonathan BostockSep 19, 2024, 3:29 PM
74 points
3 comments5 min readEA link
(www.lesswrong.com)

An­nounc­ing New Begin­ner-friendly Book on AI Safety and Risk

Darren McKeeNov 25, 2023, 3:57 PM
114 points
9 comments1 min readEA link

Brand­ing AI Safety Groups: A Field Guide

Agustín Covarrubias 🔸May 13, 2024, 5:17 PM
44 points
6 comments1 min readEA link

The Cruel Trade-Off Between AI Mi­suse and AI X-risk Concerns

simeon_cApr 22, 2023, 1:49 PM
21 points
17 comments1 min readEA link

Sam Alt­man’s Chip Am­bi­tions Un­der­cut OpenAI’s Safety Strategy

GarrisonFeb 10, 2024, 7:52 PM
286 points
20 comments3 min readEA link
(garrisonlovely.substack.com)

[linkpost] “What Are Rea­son­able AI Fears?” by Robin Han­son, 2023-04-23

Arjun PanicksseryApr 14, 2023, 11:26 PM
41 points
3 comments4 min readEA link
(quillette.com)

[Linkpost] ‘The God­father of A.I.’ Leaves Google and Warns of Danger Ahead

imp4rtial 🔸May 1, 2023, 7:54 PM
43 points
3 comments3 min readEA link
(www.nytimes.com)

AISN #25: White House Ex­ec­u­tive Order on AI, UK AI Safety Sum­mit, and Progress on Vol­un­tary Eval­u­a­tions of AI Risks

Center for AI SafetyOct 31, 2023, 7:24 PM
21 points
0 comments6 min readEA link
(newsletter.safe.ai)

Pres­i­dent Bi­den Is­sues Ex­ec­u­tive Order on Safe, Se­cure, and Trust­wor­thy Ar­tifi­cial Intelligence

Tristan WilliamsOct 30, 2023, 11:15 AM
143 points
8 comments3 min readEA link
(www.whitehouse.gov)

AISN #20: LLM Pro­lifer­a­tion, AI De­cep­tion, and Con­tin­u­ing Drivers of AI Capabilities

Center for AI SafetyAug 29, 2023, 3:03 PM
12 points
0 comments8 min readEA link
(newsletter.safe.ai)

The Bletch­ley Dec­la­ra­tion on AI Safety

Hauke HillebrandtNov 1, 2023, 11:44 AM
60 points
3 comments4 min readEA link
(www.gov.uk)

An­nounc­ing Su­per­in­tel­li­gence Imag­ined: A cre­ative con­test on the risks of superintelligence

TaylorJnsJun 12, 2024, 3:20 PM
17 points
0 comments1 min readEA link

Bi­den-Har­ris Ad­minis­tra­tion An­nounces First-Ever Con­sor­tium Ded­i­cated to AI Safety

ben.smithFeb 9, 2024, 6:40 AM
15 points
1 comment1 min readEA link
(www.nist.gov)

Dis­rupt­ing mal­i­cious uses of AI by state-af­fili­ated threat actors

Agustín Covarrubias 🔸Feb 14, 2024, 9:28 PM
22 points
1 comment1 min readEA link
(openai.com)

In­tro­duc­ing StakeOut.AI

Harry LukFeb 17, 2024, 12:21 AM
52 points
6 comments9 min readEA link

My ar­ti­cle in The Na­tion — Cal­ifor­nia’s AI Safety Bill Is a Mask-Off Mo­ment for the Industry

GarrisonAug 15, 2024, 7:25 PM
134 points
0 comments1 min readEA link
(www.thenation.com)

Propos­ing the Con­di­tional AI Safety Treaty (linkpost TIME)

OttoNov 15, 2024, 1:56 PM
12 points
6 comments3 min readEA link
(time.com)

Demis Hass­abis — Google Deep­Mind: The Podcast

Zach Stein-PerlmanAug 16, 2024, 12:00 AM
22 points
2 comments1 min readEA link
(www.youtube.com)

An­thropic An­nounces new S.O.T.A. Claude 3

Joseph MillerMar 4, 2024, 7:02 PM
10 points
5 comments1 min readEA link
(twitter.com)

Claude Doesn’t Want to Die

GarrisonMar 5, 2024, 6:00 AM
22 points
14 comments10 min readEA link
(garrisonlovely.substack.com)

AISN #32: Mea­sur­ing and Re­duc­ing Hazardous Knowl­edge in LLMs Plus, Fore­cast­ing the Fu­ture with LLMs, and Reg­u­la­tory Markets

Center for AI SafetyMar 7, 2024, 4:37 PM
15 points
2 comments8 min readEA link
(newsletter.safe.ai)

OpenAI o1

Zach Stein-PerlmanSep 12, 2024, 6:54 PM
38 points
0 comments1 min readEA link

OpenAI: Pre­pared­ness framework

Zach Stein-PerlmanDec 18, 2023, 6:30 PM
24 points
0 comments1 min readEA link
(openai.com)

OpenAI an­nounces new mem­bers to board of directors

Will Howard🔹Mar 9, 2024, 11:27 AM
47 points
12 comments2 min readEA link
(openai.com)

Among the A.I. Doom­say­ers—The New Yorker

Agustín Covarrubias 🔸Mar 11, 2024, 9:12 PM
66 points
0 comments1 min readEA link
(www.newyorker.com)

Cy­ber­se­cu­rity and AI: The Evolv­ing Se­cu­rity Landscape

Center for AI SafetyMar 14, 2024, 8:14 PM
9 points
0 comments12 min readEA link
(www.safe.ai)

INTERVIEW: Round 2 - StakeOut.AI w/​ Dr. Peter Park

Jacob-HaimesMar 18, 2024, 9:26 PM
8 points
0 comments1 min readEA link
(into-ai-safety.github.io)

Some thoughts from a Univer­sity AI Debate

Charlie HarrisonMar 20, 2024, 5:03 PM
25 points
2 comments1 min readEA link

Pod­cast: In­ter­view se­ries fea­tur­ing Dr. Peter Park

Jacob-HaimesMar 26, 2024, 12:35 AM
1 point
0 comments2 min readEA link
(into-ai-safety.github.io)

AISN #28: Cen­ter for AI Safety 2023 Year in Review

Center for AI SafetyDec 23, 2023, 9:31 PM
17 points
1 comment5 min readEA link
(newsletter.safe.ai)

AI safety ad­vo­cates should con­sider pro­vid­ing gen­tle push­back fol­low­ing the events at OpenAI

I_machinegun_KellyDec 22, 2023, 9:05 PM
86 points
5 comments3 min readEA link
(www.lesswrong.com)

NYT is su­ing OpenAI&Microsoft for alleged copy­right in­fringe­ment; some quick thoughts

MikhailSaminDec 28, 2023, 6:37 PM
29 points
0 comments1 min readEA link

AISN #29: Progress on the EU AI Act Plus, the NY Times sues OpenAI for Copy­right In­fringe­ment, and Con­gres­sional Ques­tions about Re­search Stan­dards in AI Safety

Center for AI SafetyJan 4, 2024, 4:03 PM
5 points
0 comments6 min readEA link
(newsletter.safe.ai)

#176 – The fi­nal push for AGI, un­der­stand­ing OpenAI’s lead­er­ship drama, and red-team­ing fron­tier mod­els (Nathan Labenz on the 80,000 Hours Pod­cast)

80000_HoursJan 4, 2024, 4:00 PM
15 points
0 comments22 min readEA link

U.S. Com­merce Sec­re­tary Gina Raimondo An­nounces Ex­pan­sion of U.S. AI Safety In­sti­tute Lead­er­ship Team [and Paul Chris­ti­ano up­date]

PhibApr 16, 2024, 5:10 PM
116 points
8 comments1 min readEA link
(www.commerce.gov)

AI Safety Newslet­ter #42: New­som Ve­toes SB 1047 Plus, OpenAI’s o1, and AI Gover­nance Summary

Center for AI SafetyOct 1, 2024, 8:33 PM
10 points
0 comments6 min readEA link
(newsletter.safe.ai)

£1 mil­lion prize for the most cut­ting-edge AI solu­tion for pub­lic good [link post]

rileyharrisJan 17, 2024, 2:36 PM
8 points
0 comments2 min readEA link
(manchesterprize.org)

I read ev­ery ma­jor AI lab’s safety plan so you don’t have to

sarahhwDec 16, 2024, 2:12 PM
65 points
2 comments11 min readEA link
(longerramblings.substack.com)

AISN #35: Lob­by­ing on AI Reg­u­la­tion Plus, New Models from OpenAI and Google, and Le­gal Regimes for Train­ing on Copy­righted Data

Center for AI SafetyMay 16, 2024, 2:26 PM
14 points
0 comments6 min readEA link
(newsletter.safe.ai)

The Failed Strat­egy of Ar­tifi­cial In­tel­li­gence Doomers

yhoisethFeb 5, 2025, 7:34 PM
12 points
2 comments1 min readEA link
(letter.palladiummag.com)

Miti­gat­ing ex­treme AI risks amid rapid progress [Linkpost]

AkashMay 21, 2024, 8:04 PM
36 points
1 comment1 min readEA link

Publi­ca­tion of the In­ter­na­tional Scien­tific Re­port on the Safety of Ad­vanced AI (In­term Re­port)

James HerbertMay 21, 2024, 9:58 PM
11 points
2 comments2 min readEA link
(www.gov.uk)

He­len Toner (ex-OpenAI board mem­ber): “We learned about ChatGPT on Twit­ter.”

defun 🔸May 29, 2024, 7:40 AM
123 points
13 comments1 min readEA link
(x.com)

The U.S. and China Need an AI In­ci­dents Hotline

christian.rJun 3, 2024, 6:46 PM
25 points
0 comments1 min readEA link
(www.lawfaremedia.org)

An­thropic rewrote its RSP

Zach Stein-PerlmanOct 15, 2024, 2:30 PM
32 points
1 comment1 min readEA link

Is prin­ci­pled mass-out­reach pos­si­ble, for AGI X-risk?

Nicholas / Heather KrossJan 21, 2024, 5:45 PM
12 points
2 comments1 min readEA link

AISN #30: In­vest­ments in Com­pute and Mili­tary AI Plus, Ja­pan and Sin­ga­pore’s Na­tional AI Safety Institutes

Center for AI SafetyJan 24, 2024, 7:38 PM
7 points
1 comment6 min readEA link
(newsletter.safe.ai)

AI Safety: Why We Need to Keep Our Smart Machines in Check

adityaraj@eanitaDec 17, 2024, 12:29 PM
1 point
0 comments2 min readEA link
(medium.com)

Ex­ec­u­tive Direc­tor for AIS France—Ex­pres­sion of interest

gergoDec 19, 2024, 8:11 AM
33 points
0 comments4 min readEA link

Fron­tier AI sys­tems have sur­passed the self-repli­cat­ing red line

Greg_ColbournDec 10, 2024, 4:33 PM
25 points
14 comments1 min readEA link
(github.com)

It is time to start war gam­ing for AGI

yanni kyriacosOct 17, 2024, 5:14 AM
14 points
4 comments1 min readEA link

OpenAI defected, but we can take hon­est actions

RemmeltOct 21, 2024, 8:41 AM
19 points
1 comment2 min readEA link

Miles Brundage re­signed from OpenAI, and his AGI readi­ness team was disbanded

GarrisonOct 23, 2024, 11:42 PM
57 points
4 comments7 min readEA link
(garrisonlovely.substack.com)

Finish­ing The SB-1047 Doc­u­men­tary In 6 Weeks

Michaël TrazziOct 28, 2024, 8:26 PM
67 points
0 comments4 min readEA link

Why Did Elon Musk Just Offer to Buy Con­trol of OpenAI for $100 Billion?

GarrisonFeb 11, 2025, 12:20 AM
137 points
2 comments6 min readEA link
(garrisonlovely.substack.com)

o3 is not be­ing re­leased to the pub­lic. First they are only giv­ing ac­cess to ex­ter­nal safety testers. You can ap­ply to get early ac­cess to do safety testing

Kat WoodsDec 20, 2024, 6:30 PM
13 points
0 comments1 min readEA link
(openai.com)

We are in a New Paradigm of AI Progress—OpenAI’s o3 model makes huge gains on the tough­est AI bench­marks in the world

GarrisonDec 22, 2024, 9:45 PM
26 points
0 comments4 min readEA link
(garrisonlovely.substack.com)

A bet­ter “State­ment on AI Risk?” [Cross­post]

Knight LeeDec 30, 2024, 7:36 AM
4 points
0 comments3 min readEA link

Chi­nese Re­searchers Crack ChatGPT: Repli­cat­ing OpenAI’s Ad­vanced AI Model

Evan_GaensbauerJan 5, 2025, 3:50 AM
1 point
0 comments1 min readEA link
(www.geeky-gadgets.com)

AI Lab Re­tal­i­a­tion: A Sur­vival Guide

Jay ReadyJan 4, 2025, 11:05 PM
6 points
1 comment12 min readEA link
(morelightinai.substack.com)

Alt­man on the board, AGI, and superintelligence

OscarD🔸Jan 6, 2025, 2:37 PM
20 points
1 comment1 min readEA link
(blog.samaltman.com)

Tar­bell Fel­low­ship 2025 - Ap­pli­ca­tions Open (AI Jour­nal­ism)

Tarbell Center for AI JournalismJan 8, 2025, 3:25 PM
62 points
0 comments1 min readEA link

Are AI safe­ty­ists cry­ing wolf?

sarahhwJan 8, 2025, 8:54 PM
60 points
21 comments16 min readEA link
(longerramblings.substack.com)

Is AI Hit­ting a Wall or Mov­ing Faster Than Ever?

GarrisonJan 9, 2025, 10:18 PM
35 points
3 comments5 min readEA link
(garrisonlovely.substack.com)

The Com­pendium, A full ar­gu­ment about ex­tinc­tion risk from AGI

adamShimiOct 31, 2024, 12:02 PM
9 points
1 comment2 min readEA link
(www.thecompendium.ai)

Ex­plor­ing AI Safety through “Es­cape Ex­per­i­ment”: A Short Film on Su­per­in­tel­li­gence Risks

Gaetan_SelleNov 10, 2024, 4:42 AM
4 points
0 comments2 min readEA link

The Game Board has been Flipped: Now is a good time to re­think what you’re doing

LintzAJan 28, 2025, 9:20 PM
351 points
60 comments13 min readEA link

PSA: Say­ing “1 in 5” Is Bet­ter Than “20%” When In­form­ing about risks publicly

BlankaJan 30, 2025, 7:03 PM
17 points
1 comment1 min readEA link

China Hawks are Man­u­fac­tur­ing an AI Arms Race

GarrisonNov 20, 2024, 6:17 PM
95 points
3 comments5 min readEA link
(garrisonlovely.substack.com)

OpenAI’s CBRN tests seem unclear

Luca Righetti 🔸Nov 21, 2024, 5:26 PM
82 points
3 comments7 min readEA link

[Question] Seek­ing Tan­gible Ex­am­ples of AI Catastrophes

clifford.banesNov 25, 2024, 7:55 AM
9 points
2 comments1 min readEA link

OpenAI’s o1 tried to avoid be­ing shut down, and lied about it, in evals

Greg_ColbournDec 6, 2024, 3:25 PM
23 points
9 comments1 min readEA link
(www.transformernews.ai)

Ex­ec­u­tive Direc­tor for AIS Brus­sels—Ex­pres­sion of interest

gergoDec 19, 2024, 9:15 AM
28 points
0 comments4 min readEA link

Ter­minol­ogy sug­ges­tion: stan­dard­ize terms for prob­a­bil­ity ranges

Egg SyntaxAug 30, 2024, 4:05 PM
2 points
0 comments1 min readEA link

AI Safety Newslet­ter #41: The Next Gen­er­a­tion of Com­pute Scale Plus, Rank­ing Models by Sus­cep­ti­bil­ity to Jailbreak­ing, and Ma­chine Ethics

Center for AI SafetySep 11, 2024, 7:11 PM
12 points
0 comments5 min readEA link
(newsletter.safe.ai)

Meta: Fron­tier AI Framework

Zach Stein-PerlmanFeb 3, 2025, 10:00 PM
23 points
0 comments1 min readEA link
(ai.meta.com)

An­thropic is be­ing sued for copy­ing books to train Claude

RemmeltAug 31, 2024, 2:57 AM
3 points
0 comments1 min readEA link
(fingfx.thomsonreuters.com)

Unions for AI safety?

dEAsignSep 24, 2023, 12:13 AM
7 points
12 comments2 min readEA link

[Con­gres­sional Hear­ing] Over­sight of A.I.: Leg­is­lat­ing on Ar­tifi­cial Intelligence

Tristan WilliamsNov 1, 2023, 6:15 PM
5 points
1 comment7 min readEA link
(www.judiciary.senate.gov)

Ama­zon to in­vest up to $4 billion in Anthropic

Davis_KingsleySep 25, 2023, 2:55 PM
38 points
34 comments1 min readEA link
(twitter.com)

An­nounc­ing #AISum­mitTalks fea­tur­ing Pro­fes­sor Stu­art Rus­sell and many others

OttoOct 24, 2023, 10:16 AM
9 points
1 comment1 min readEA link

Go Mo­bi­lize? Les­sons from GM Protests for Paus­ing AI

Charlie HarrisonOct 24, 2023, 3:01 PM
48 points
11 comments31 min readEA link

The Dis­solu­tion of AI Safety

RokoDec 12, 2024, 10:46 AM
−7 points
0 comments1 min readEA link
(www.transhumanaxiology.com)

[Linkpost] NY Times Fea­ture on Anthropic

GarrisonJul 12, 2023, 7:30 PM
34 points
3 comments5 min readEA link
(www.nytimes.com)

Sam Alt­man fired from OpenAI

LarksNov 17, 2023, 9:07 PM
133 points
90 comments1 min readEA link
(openai.com)

Thoughts on yes­ter­day’s UN Se­cu­rity Coun­cil meet­ing on AI

Greg_ColbournJul 19, 2023, 4:46 PM
31 points
2 comments1 min readEA link

AI Im­pacts Quar­terly Newslet­ter, Apr-Jun 2023

HarlanJul 18, 2023, 6:01 PM
4 points
0 comments3 min readEA link
(blog.aiimpacts.org)

AISN #16: White House Se­cures Vol­un­tary Com­mit­ments from Lead­ing AI Labs and Les­sons from Oppenheimer

Center for AI SafetyJul 25, 2023, 4:45 PM
7 points
0 comments6 min readEA link
(newsletter.safe.ai)

[Cross­post] An AI Pause Is Hu­man­ity’s Best Bet For Prevent­ing Ex­tinc­tion (TIME)

OttoJul 24, 2023, 10:18 AM
36 points
3 comments7 min readEA link
(time.com)

Linkpost: 7 A.I. Com­pa­nies Agree to Safe­guards After Pres­sure From the White House

MHR🔸Jul 21, 2023, 1:23 PM
61 points
4 comments1 min readEA link
(www.nytimes.com)

[link post] AI Should Be Ter­rified of Humans

BrianKJul 24, 2023, 11:13 AM
28 points
0 comments1 min readEA link
(time.com)

[Linkpost] Eric Sch­witzgebel: AI sys­tems must not con­fuse users about their sen­tience or moral status

🔸Zachary BrownAug 18, 2023, 5:21 PM
6 points
0 comments2 min readEA link
(www.sciencedirect.com)

AISN #17: Au­to­mat­i­cally Cir­cum­vent­ing LLM Guardrails, the Fron­tier Model Fo­rum, and Se­nate Hear­ing on AI Oversight

Center for AI SafetyAug 1, 2023, 3:24 PM
15 points
0 comments8 min readEA link

Elic­it­ing re­sponses to Marc An­dreessen’s “Why AI Will Save the World”

ColemanJul 17, 2023, 7:58 PM
2 points
2 comments1 min readEA link
(a16z.com)

Fron­tier Model Forum

Zach Stein-PerlmanJul 26, 2023, 2:30 PM
40 points
7 comments1 min readEA link
(blog.google)

As­ter­isk Magaz­ine Is­sue 03: AI

Alejandro OrtegaJul 24, 2023, 3:53 PM
34 points
3 comments1 min readEA link
(asteriskmag.com)

AISN #27: Defen­sive Ac­cel­er­a­tionism, A Ret­ro­spec­tive On The OpenAI Board Saga, And A New AI Bill From Se­na­tors Thune And Klobuchar

Center for AI SafetyDec 7, 2023, 3:57 PM
10 points
0 comments6 min readEA link
(newsletter.safe.ai)

Gavin New­som ve­toes SB 1047

LarksSep 30, 2024, 12:06 AM
39 points
14 comments1 min readEA link
(www.wsj.com)

The costs of caution

Kelsey PiperMay 1, 2023, 8:04 PM
112 points
17 comments4 min readEA link

AI Safety Newslet­ter #4: AI and Cy­ber­se­cu­rity, Per­sua­sive AIs, Weaponiza­tion, and Ge­offrey Hin­ton talks AI risks

Center for AI SafetyMay 2, 2023, 4:51 PM
35 points
2 comments5 min readEA link
(newsletter.safe.ai)

AI Safety Newslet­ter #1 [CAIS Linkpost]

AkashApr 10, 2023, 8:18 PM
38 points
0 comments1 min readEA link

AI Safety Newslet­ter #2: ChaosGPT, Nat­u­ral Selec­tion, and AI Safety in the Media

Oliver ZApr 18, 2023, 6:36 PM
56 points
1 comment4 min readEA link
(newsletter.safe.ai)

My choice of AI mis­al­ign­ment in­tro­duc­tion for a gen­eral audience

BillMay 3, 2023, 12:15 AM
7 points
2 comments1 min readEA link
(youtu.be)

AI X-risk in the News: How Effec­tive are Re­cent Me­dia Items and How is Aware­ness Chang­ing? Our New Sur­vey Re­sults.

OttoMay 4, 2023, 2:04 PM
49 points
1 comment9 min readEA link

[Link Post: New York Times] White House Un­veils Ini­ti­a­tives to Re­duce Risks of A.I.

RockwellMay 4, 2023, 2:04 PM
50 points
1 comment2 min readEA link

An Up­date On The Cam­paign For AI Safety Dot Org

yanni kyriacosMay 5, 2023, 12:19 AM
26 points
4 comments1 min readEA link

AI Safety Newslet­ter #5: Ge­offrey Hin­ton speaks out on AI risk, the White House meets with AI labs, and Tro­jan at­tacks on lan­guage models

Center for AI SafetyMay 9, 2023, 3:26 PM
60 points
0 comments4 min readEA link
(newsletter.safe.ai)

AI-Risk in the State of the Euro­pean Union Address

Sam BogerdSep 13, 2023, 1:27 PM
25 points
0 comments3 min readEA link
(state-of-the-union.ec.europa.eu)

The In­ter­na­tional PauseAI Protest: Ac­tivism un­der uncertainty

Joseph MillerOct 12, 2023, 5:36 PM
129 points
3 comments4 min readEA link

AI Safety Newslet­ter #6: Ex­am­ples of AI safety progress, Yoshua Ben­gio pro­poses a ban on AI agents, and les­sons from nu­clear arms control

Center for AI SafetyMay 16, 2023, 3:14 PM
32 points
1 comment6 min readEA link
(newsletter.safe.ai)

Effi­cacy of AI Ac­tivism: Have We Ever Said No?

Charlie HarrisonOct 27, 2023, 4:52 PM
78 points
25 comments20 min readEA link

Sam Alt­man /​ Open AI Dis­cus­sion Thread

John SalterNov 20, 2023, 9:21 AM
40 points
36 comments1 min readEA link

Ilya: The AI sci­en­tist shap­ing the world

David VargaNov 20, 2023, 12:43 PM
6 points
1 comment4 min readEA link

Former Is­raeli Prime Minister Speaks About AI X-Risk

Yonatan CaleMay 20, 2023, 12:09 PM
73 points
6 comments1 min readEA link

Pos­si­ble OpenAI’s Q* break­through and Deep­Mind’s AlphaGo-type sys­tems plus LLMs

BurnydelicNov 23, 2023, 7:02 AM
13 points
4 comments2 min readEA link

[Linkpost] “Gover­nance of su­per­in­tel­li­gence” by OpenAI

Daniel_EthMay 22, 2023, 8:15 PM
51 points
6 comments2 min readEA link
(openai.com)

OpenAI board re­ceived let­ter warn­ing of pow­er­ful AI

JordanStoneNov 23, 2023, 12:16 AM
26 points
2 comments1 min readEA link
(www.reuters.com)

[Question] Would an An­thropic/​OpenAI merger be good for AI safety?

MNov 22, 2023, 8:21 PM
6 points
1 comment1 min readEA link

Rishi Su­nak men­tions “ex­is­ten­tial threats” in talk with OpenAI, Deep­Mind, An­thropic CEOs

Arjun PanicksseryMay 24, 2023, 9:06 PM
44 points
2 comments1 min readEA link

Tim Cook was asked about ex­tinc­tion risks from AI

Saul MunnJun 6, 2023, 6:46 PM
8 points
1 comment1 min readEA link

Could AI ac­cel­er­ate eco­nomic growth?

Tom_DavidsonJun 7, 2023, 7:07 PM
28 points
0 comments6 min readEA link

On Deep­Mind and Try­ing to Fairly Hear Out Both AI Doomers and Doubters (Ro­hin Shah on The 80,000 Hours Pod­cast)

80000_HoursJun 12, 2023, 12:53 PM
28 points
1 comment15 min readEA link

UK gov­ern­ment to host first global sum­mit on AI Safety

DavidNashJun 8, 2023, 1:24 PM
78 points
1 comment5 min readEA link
(www.gov.uk)

Linkpost: Dwarkesh Pa­tel in­ter­view­ing Carl Shulman

Stefan_SchubertJun 14, 2023, 3:30 PM
110 points
5 comments1 min readEA link
(podcastaddict.com)

Google Deep­Mind re­leases Gemini

YarrowDec 6, 2023, 5:39 PM
21 points
7 comments1 min readEA link
(deepmind.google)

Com­mu­ni­ca­tion by ex­is­ten­tial risk or­ga­ni­za­tions: State of the field and sug­ges­tions for improvement

Existential Risk Communication ProjectAug 13, 2024, 7:06 AM
10 points
3 comments13 min readEA link

The UK AI Safety Sum­mit tomorrow

SebastianSchmidtOct 31, 2023, 7:09 PM
17 points
2 comments2 min readEA link
No comments.