See also the following posts, published a few months after this one, which discuss AGI race dynamics (in the context of a fictional AI lab named Magma):
‘AI strategy nearcasting’ (Karnofsky)
‘How might we align transformative AI if it’s developed very soon?’ (Karnofsky)
‘Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover’ (Cotra)
Current theme: default
Less Wrong (text)
Less Wrong (link)
Arrow keys: Next/previous image
Escape or click: Hide zoomed image
Space bar: Reset image size & position
Scroll to zoom in/out
(When zoomed in, drag to pan; double-click to close)
Keys shown in yellow (e.g., ]) are accesskeys, and require a browser-specific modifier key (or keys).
]
Keys shown in grey (e.g., ?) do not require any modifier keys.
?
Esc
h
f
a
m
v
c
r
q
t
u
o
,
.
/
s
n
e
;
Enter
[
\
k
i
l
=
-
0
′
1
2
3
4
5
6
7
8
9
→
↓
←
↑
Space
x
z
`
g
See also the following posts, published a few months after this one, which discuss AGI race dynamics (in the context of a fictional AI lab named Magma):
‘AI strategy nearcasting’ (Karnofsky)
‘How might we align transformative AI if it’s developed very soon?’ (Karnofsky)
‘Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover’ (Cotra)