RSS

Remmelt

Karma: 971

See explainer on why AGI could not be controlled enough to stay safe:
https://​​www.lesswrong.com/​​posts/​​xp6n2MG5vQkPpFEBH/​​the-control-problem-unsolved-or-unsolvable

Note: I am no longer part of EA because of the community’s/​philosophy’s overreaches. I still post here about AI safety.

Test­ing an EA net­work-build­ing strat­egy in the Netherlands

Remmelt3 Jul 2017 11:28 UTC
16 points
9 comments5 min readEA link

Effec­tive Altru­ism as a Mar­ket in Mo­ral Goods – Introduction

Remmelt6 Aug 2017 2:29 UTC
2 points
1 comment1 min readEA link

Reflec­tions on com­mu­nity build­ing in the Netherlands

Remmelt2 Nov 2017 22:01 UTC
10 points
4 comments10 min readEA link

Pro­posal for the AI Safety Re­search Camp

Remmelt2 Feb 2018 8:07 UTC
16 points
0 comments9 min readEA link

The Values-to-Ac­tions De­ci­sion Chain: a rough model

Remmelt2 Mar 2018 14:54 UTC
1 point
5 comments2 min readEA link

The first AI Safety Camp & onwards

Remmelt7 Jun 2018 18:49 UTC
25 points
2 comments8 min readEA link

The Values-to-Ac­tions De­ci­sion Chain: a lens for im­prov­ing coordination

Remmelt30 Jun 2018 9:26 UTC
33 points
33 comments10 min readEA link

Con­sider pay­ing me (or an­other en­trepreneur) to cre­ate ser­vices for effec­tive altruism

Remmelt3 Nov 2020 20:50 UTC
44 points
17 comments6 min readEA link

Del­e­gated agents in prac­tice: How com­pa­nies might end up sel­l­ing AI ser­vices that act on be­half of con­sumers and coal­i­tions, and what this im­plies for safety research

Remmelt26 Nov 2020 16:39 UTC
11 points
0 comments4 min readEA link

Are we ac­tu­ally im­prov­ing de­ci­sion-mak­ing?

Remmelt4 Feb 2021 23:57 UTC
22 points
5 comments3 min readEA link

A parable of brightspots and blindspots

Remmelt21 Mar 2021 17:31 UTC
10 points
2 comments3 min readEA link

Some blindspots in ra­tio­nal­ity and effec­tive altruism

Remmelt21 Mar 2021 18:01 UTC
53 points
6 comments13 min readEA link

How teams went about their re­search at AI Safety Camp edi­tion 5

Remmelt28 Jun 2021 15:18 UTC
24 points
0 comments6 min readEA link

Two ten­ta­tive con­cerns about OpenPhil’s Macroe­co­nomic Sta­bi­liza­tion Policy work

Remmelt3 Jan 2022 12:37 UTC
42 points
26 comments7 min readEA link

Why mechanis­tic in­ter­pretabil­ity does not and can­not con­tribute to long-term AGI safety (from mes­sages with a friend)

Remmelt19 Dec 2022 12:02 UTC
17 points
3 comments1 min readEA link

List #1: Why stop­ping the de­vel­op­ment of AGI is hard but doable

Remmelt24 Dec 2022 9:52 UTC
30 points
2 comments1 min readEA link

List #2: Why co­or­di­nat­ing to al­ign as hu­mans to not de­velop AGI is a lot eas­ier than, well… co­or­di­nat­ing as hu­mans with AGI co­or­di­nat­ing to be al­igned with humans

Remmelt24 Dec 2022 9:53 UTC
3 points
0 comments1 min readEA link

List #3: Why not to as­sume on prior that AGI-al­ign­ment workarounds are available

Remmelt24 Dec 2022 9:54 UTC
6 points
0 comments1 min readEA link

Nine Points of Col­lec­tive Insanity

Remmelt27 Dec 2022 3:14 UTC
1 point
0 comments1 min readEA link

How ‘Hu­man-Hu­man’ dy­nam­ics give way to ‘Hu­man-AI’ and then ‘AI-AI’ dynamics

Remmelt27 Dec 2022 3:16 UTC
4 points
0 comments1 min readEA link