Archive
About
Search
Log In
Home
All
Wiki
Shortform
Recent
Comments
RSS
Apart Research
Karma:
285
All
Posts
Comments
New
Top
Old
Join the AI governance and interpretability hackathons!
Esben Kran
23 Mar 2023 14:39 UTC
33
points
1
comment
5
min read
EA
link
(alignmentjam.com)
Results from the AI testing hackathon
Esben Kran
2 Jan 2023 15:46 UTC
35
points
4
comments
5
min read
EA
link
(alignmentjam.com)
Join the AI Testing Hackathon this Friday
Esben Kran
12 Dec 2022 14:24 UTC
33
points
0
comments
8
min read
EA
link
(alignmentjam.com)
Join the interpretability research hackathon
Esben Kran
28 Oct 2022 16:26 UTC
48
points
0
comments
5
min read
EA
link
Newsletter for Alignment Research: The ML Safety Updates
Esben Kran
22 Oct 2022 16:17 UTC
30
points
0
comments
7
min read
EA
link
Apart Research
18 Oct 2022 19:50 UTC
4
points
0 ∶ 0
in reply to:
FlorentBerthet
’s
comment
on:
AI Safety Ideas: An Open AI Safety Research Platform
Very true! We have graciously adopted your formulation. Thank you.
AI Safety Ideas: A collaborative AI safety research platform
Apart Research
17 Oct 2022 17:01 UTC
67
points
13
comments
4
min read
EA
link
Safety timelines: How long will it take to solve alignment?
Esben Kran
19 Sep 2022 12:51 UTC
45
points
9
comments
6
min read
EA
link
Black Box Investigations Research Hackathon
Esben Kran
15 Sep 2022 10:09 UTC
23
points
0
comments
2
min read
EA
link
Back to top
Very true! We have graciously adopted your formulation. Thank you.