RSS

Rea­son­ing transparency

TagLast edit: 14 Jul 2022 15:42 UTC by Pablo

Reasoning transparency is a form of transparency that prioritizes the sharing of information about underlying general thinking processes and philosophy and the communication of this information in ways that make it easier for the recipient to determine what updates to make in response to it.

Further reading

Muehlhauser, Luke (2017) Reasoning transparency, Open Philanthropy, December.

Related entries

transparency

Rea­son­ing Transparency

Lizka28 Sep 2022 12:22 UTC
117 points
8 comments17 min readEA link
(www.openphilanthropy.org)

AI X-Risk: In­te­grat­ing on the Shoulders of Giants

TD_Pilditch1 Nov 2022 16:07 UTC
34 points
0 comments47 min readEA link

[Question] Should 80,000 Hours be more trans­par­ent about how they rank prob­lems and ca­reers?

Vasco Grilo🔸20 Dec 2023 8:14 UTC
91 points
17 comments3 min readEA link

Silly idea to en­hance List rep­re­sen­ta­tion accuracy

Phib24 Apr 2023 0:30 UTC
7 points
4 comments2 min readEA link

Epistemic Legibility

Elizabeth21 Mar 2022 19:18 UTC
79 points
1 comment20 min readEA link

Say how much, not more or less ver­sus some­one else

Gregory Lewis🔸28 Dec 2023 22:24 UTC
100 points
10 comments5 min readEA link

Challenges of Transparency

Holden Karnofsky11 Sep 2014 12:46 UTC
10 points
0 comments7 min readEA link
(www.openphilanthropy.org)

Feed­back I’ve been giv­ing to ju­nior x-risk researchers

Will Aldred15 Aug 2022 20:46 UTC
146 points
2 comments5 min readEA link

[Question] Should the main text of the write-ups of Open Philan­thropy’s large grants be longer than 1 para­graph?

Vasco Grilo🔸31 Mar 2024 9:04 UTC
55 points
40 comments1 min readEA link

You Don’t Need To Jus­tify Everything

TW12312 Jun 2022 18:36 UTC
146 points
11 comments3 min readEA link

Ab­solute Zero: AlphaZero for LLM

alapmi12 May 2025 14:54 UTC
2 points
0 comments1 min readEA link

[Question] On what ba­sis did Founder’s Pledge dis­perse $1.6 mil. to Qvist Con­sult­ing from its Cli­mate Change Fund?

Kieran.M27 Mar 2023 3:26 UTC
39 points
13 comments2 min readEA link

Even af­ter GPT-4, AI re­searchers fore­casted a 50% chance of AGI by 2047 or 2116, de­pend­ing how you define AGI

Yarrow Bouchard 🔸28 Oct 2025 16:55 UTC
18 points
17 comments3 min readEA link

Should I Fol­low This Ad­vice? A Frame­work and Workbook

Lin BL25 Apr 2025 17:03 UTC
10 points
1 comment3 min readEA link

Re­mem­ber to trans­late your thoughts back again

brook1 Nov 2022 8:51 UTC
5 points
0 comments3 min readEA link
(www.lesswrong.com)

Les­sons and in­sights from an EA marketer

Igor Scaldini28 Oct 2025 17:19 UTC
15 points
3 comments5 min readEA link

CORVUS 2.0 First Tests: Found Crit­i­cal Limi­ta­tions in My Con­sti­tu­tional AI System

Frankle Fry21 Oct 2025 15:14 UTC
−5 points
0 comments3 min readEA link

LW4EA: Epistemic Legibility

Jeremy16 Aug 2022 15:55 UTC
5 points
2 comments3 min readEA link
(www.lesswrong.com)

Ra­tion­al­ist Epistemics and the Se­quences (Effec­tive Altru­ism Defi­ni­tions Se­quence)

ozymandias8 Aug 2025 18:58 UTC
30 points
1 comment6 min readEA link

Limits to Legibility

Jan_Kulveit29 Jun 2022 17:45 UTC
104 points
3 comments5 min readEA link
(www.lesswrong.com)

Fea­tures that make a re­port es­pe­cially helpful to me

lukeprog12 Apr 2022 13:57 UTC
107 points
1 comment2 min readEA link

Why Mo­ral Con­flict Re­s­olu­tion Still Breaks Our Best Safety Tools

JBug18 Nov 2025 7:49 UTC
6 points
0 comments2 min readEA link

Au­ton­omy Consequentialism

Richard Y Chappell🔸16 Jul 2025 21:06 UTC
7 points
0 comments3 min readEA link
(www.goodthoughts.blog)

5 ways to im­prove CoT faithfulness

CBiddulph8 Oct 2024 4:17 UTC
8 points
0 comments6 min readEA link

‘GiveWell for AI Safety’: Les­sons learned in a week

Lydia Nottingham30 May 2025 16:10 UTC
45 points
1 comment6 min readEA link

Limit­ing Reason

Richard Y Chappell🔸2 Jun 2025 16:57 UTC
25 points
0 comments4 min readEA link
(www.goodthoughts.blog)
No comments.