RSS

Holden Karnofsky

Karma: 8,099

Good job op­por­tu­ni­ties for helping with the most im­por­tant century

Holden Karnofsky18 Jan 2024 19:21 UTC
49 points
1 comment4 min readEA link
(www.cold-takes.com)

We’re Not Ready: thoughts on “paus­ing” and re­spon­si­ble scal­ing policies

Holden Karnofsky27 Oct 2023 15:19 UTC
143 points
23 comments1 min readEA link

3 lev­els of threat obfuscation

Holden Karnofsky2 Aug 2023 17:09 UTC
31 points
0 comments6 min readEA link
(www.alignmentforum.org)

A Play­book for AI Risk Re­duc­tion (fo­cused on mis­al­igned AI)

Holden Karnofsky6 Jun 2023 18:05 UTC
81 points
17 comments1 min readEA link

Seek­ing (Paid) Case Stud­ies on Standards

Holden Karnofsky26 May 2023 17:58 UTC
92 points
14 comments1 min readEA link

Suc­cess with­out dig­nity: a nearcast­ing story of avoid­ing catas­tro­phe by luck

Holden Karnofsky15 Mar 2023 20:17 UTC
99 points
3 comments1 min readEA link

What does Bing Chat tell us about AI risk?

Holden Karnofsky28 Feb 2023 18:47 UTC
100 points
8 comments2 min readEA link
(www.cold-takes.com)

How ma­jor gov­ern­ments can help with the most im­por­tant century

Holden Karnofsky24 Feb 2023 19:37 UTC
55 points
4 comments4 min readEA link
(www.cold-takes.com)

Tak­ing a leave of ab­sence from Open Philan­thropy to work on AI safety

Holden Karnofsky23 Feb 2023 19:05 UTC
424 points
31 comments2 min readEA link

What AI com­pa­nies can do to­day to help with the most im­por­tant century

Holden Karnofsky20 Feb 2023 17:40 UTC
104 points
8 comments11 min readEA link
(www.cold-takes.com)

Jobs that can help with the most im­por­tant century

Holden Karnofsky12 Feb 2023 18:19 UTC
52 points
2 comments32 min readEA link
(www.cold-takes.com)

We’re no longer “paus­ing most new longter­mist fund­ing com­mit­ments”

Holden Karnofsky30 Jan 2023 19:29 UTC
199 points
40 comments6 min readEA link

Spread­ing mes­sages to help with the most im­por­tant century

Holden Karnofsky25 Jan 2023 20:35 UTC
123 points
21 comments18 min readEA link
(www.cold-takes.com)

How we could stum­ble into AI catastrophe

Holden Karnofsky16 Jan 2023 14:52 UTC
78 points
0 comments31 min readEA link
(www.cold-takes.com)

Trans­for­ma­tive AI is­sues (not just mis­al­ign­ment): an overview

Holden Karnofsky6 Jan 2023 2:19 UTC
31 points
0 comments22 min readEA link
(www.cold-takes.com)

Rac­ing through a minefield: the AI de­ploy­ment problem

Holden Karnofsky31 Dec 2022 21:44 UTC
79 points
1 comment13 min readEA link
(www.cold-takes.com)

High-level hopes for AI alignment

Holden Karnofsky20 Dec 2022 2:11 UTC
118 points
14 comments19 min readEA link
(www.cold-takes.com)

AI Safety Seems Hard to Measure

Holden Karnofsky11 Dec 2022 1:31 UTC
90 points
3 comments14 min readEA link
(www.cold-takes.com)

Why Would AI “Aim” To Defeat Hu­man­ity?

Holden Karnofsky29 Nov 2022 18:59 UTC
24 points
0 comments32 min readEA link
(www.cold-takes.com)

My takes on the FTX situ­a­tion will (mostly) be cold, not hot

Holden Karnofsky18 Nov 2022 23:57 UTC
398 points
33 comments5 min readEA link