RSS

Holden Karnofsky

Karma: 8,640

Case stud­ies on so­cial-welfare-based stan­dards in var­i­ous industries

Holden Karnofsky20 Jun 2024 13:33 UTC
73 points
2 comments1 min readEA link

Join­ing the Carnegie En­dow­ment for In­ter­na­tional Peace

Holden Karnofsky29 Apr 2024 15:45 UTC
228 points
14 comments2 min readEA link

Good job op­por­tu­ni­ties for helping with the most im­por­tant century

Holden Karnofsky18 Jan 2024 19:21 UTC
46 points
1 comment4 min readEA link
(www.cold-takes.com)

We’re Not Ready: thoughts on “paus­ing” and re­spon­si­ble scal­ing policies

Holden Karnofsky27 Oct 2023 15:19 UTC
150 points
23 comments1 min readEA link

3 lev­els of threat obfuscation

Holden Karnofsky2 Aug 2023 17:09 UTC
31 points
0 comments6 min readEA link
(www.alignmentforum.org)

A Play­book for AI Risk Re­duc­tion (fo­cused on mis­al­igned AI)

Holden Karnofsky6 Jun 2023 18:05 UTC
81 points
17 comments1 min readEA link

Seek­ing (Paid) Case Stud­ies on Standards

Holden Karnofsky26 May 2023 17:58 UTC
99 points
14 comments1 min readEA link

Suc­cess with­out dig­nity: a nearcast­ing story of avoid­ing catas­tro­phe by luck

Holden Karnofsky15 Mar 2023 20:17 UTC
113 points
3 comments1 min readEA link

What does Bing Chat tell us about AI risk?

Holden Karnofsky28 Feb 2023 18:47 UTC
99 points
8 comments2 min readEA link
(www.cold-takes.com)

How ma­jor gov­ern­ments can help with the most im­por­tant century

Holden Karnofsky24 Feb 2023 19:37 UTC
56 points
4 comments4 min readEA link
(www.cold-takes.com)

Tak­ing a leave of ab­sence from Open Philan­thropy to work on AI safety

Holden Karnofsky23 Feb 2023 19:05 UTC
420 points
31 comments2 min readEA link

What AI com­pa­nies can do to­day to help with the most im­por­tant century

Holden Karnofsky20 Feb 2023 17:40 UTC
104 points
8 comments11 min readEA link
(www.cold-takes.com)

Jobs that can help with the most im­por­tant century

Holden Karnofsky12 Feb 2023 18:19 UTC
57 points
2 comments32 min readEA link
(www.cold-takes.com)

We’re no longer “paus­ing most new longter­mist fund­ing com­mit­ments”

Holden Karnofsky30 Jan 2023 19:29 UTC
201 points
39 comments6 min readEA link

Spread­ing mes­sages to help with the most im­por­tant century

Holden Karnofsky25 Jan 2023 20:35 UTC
128 points
21 comments18 min readEA link
(www.cold-takes.com)

How we could stum­ble into AI catastrophe

Holden Karnofsky16 Jan 2023 14:52 UTC
83 points
0 comments31 min readEA link
(www.cold-takes.com)

Trans­for­ma­tive AI is­sues (not just mis­al­ign­ment): an overview

Holden Karnofsky6 Jan 2023 2:19 UTC
36 points
0 comments22 min readEA link
(www.cold-takes.com)

Rac­ing through a minefield: the AI de­ploy­ment problem

Holden Karnofsky31 Dec 2022 21:44 UTC
79 points
1 comment13 min readEA link
(www.cold-takes.com)

High-level hopes for AI alignment

Holden Karnofsky20 Dec 2022 2:11 UTC
123 points
14 comments19 min readEA link
(www.cold-takes.com)

AI Safety Seems Hard to Measure

Holden Karnofsky11 Dec 2022 1:31 UTC
90 points
4 comments14 min readEA link
(www.cold-takes.com)