RSS

ukc10014

Karma: 56

PhD candidate at Goldsmiths College, University of London. Title: ‘Reasons for Persons, or the Good Successor Problem’. Abstract: AI alignment aims for advanced machine intelligences that preserves and enhances human welfare (appropriately defined). In a narrow sense, this includes not reflecting existing biases in society or destabilising political systems; more broadly, it could also mean not creating conditions that result in the extinction or disempowerment of humanity. This project tries to define a alternative, speculative vision of alignment: one that relaxes the assumption (tacit in some alignment discourse) that humans must indefinitely retain control over the future. My exploration instead aims for some ambitious and catholic notion of value (that is, avoiding maximisers of the squiggle/​paperclip/​hedonium varieties and fleshing out sources of value that don’t hinge on a biological human subject). I draw upon philosophy (moral realism, population ethics, and decision theory), aesthetic theory (can the human tendency to make and appreciate aesthetic products, something broadly shared across cultures, be generalised to AIs or is it some contingent, and evolutionarily-useful, practice that arose amongst a particular set of primates?) This project has an empirical aspect: to the extent possible, I want to enrich speculations with experiments using LLMs in multi-agent setups.

Linkpost: Epis­tle to the Successors

ukc1001414 Jul 2024 20:07 UTC
4 points
0 comments1 min readEA link
(ukc10014.github.io)

Reza Ne­garestani’s In­tel­li­gence & Spirit

ukc1001427 Jun 2024 18:17 UTC
7 points
1 comment4 min readEA link

Analysing a 2036 Takeover Scenario

ukc100146 Oct 2022 20:48 UTC
4 points
1 comment1 min readEA link

Defec­tive Altru­ism ar­ti­cle in Cur­rent Af­fairs Magazine

ukc1001422 Sep 2022 13:27 UTC
13 points
7 comments2 min readEA link

(Re)con­sid­er­ing the Aes­thet­ics of EA

ukc1001420 May 2022 15:01 UTC
24 points
4 comments6 min readEA link