Thanks for the input!
On Scheming: I actually donât think scheming risk is the most important factor. Even removing it completely doesnât change my final conclusion. I agree that a bimodal distribution with scheming/ânon-scheming would be appropriate for a more sophisticated model. I just ended up lowering the weight I assign to the scheming factor (by half) to take into account that I am not sure whether scheming will/âwonât be an issue.
In my analysis, the ability to get good feedback signals/âsuccess criteria is the factor that moves me the most to thinking that capabilities get sped up before safety.
On Task length: You have more visibility into this, so Iâm happy to defer. But Iâd love to hear more about why you think tasks in capabilities research have longer task lengths. Is it because you have to run large evals or do pre-training runs? Do you think this argument applies to all areas of capabilities research?
I quite like the âImportance of quantity over qualityâ factor and hadnât thought of this before! Itâs slightly entangled with verifiability (if you can cheaply assess whether something is good, then itâs fine to have a large quantity of attempts even if many are bad). But I think quantity vs quality also matters beyond that. Iâll add it as a possible additional factor in the post.
I agree that this factor advantages Data Generation and AI Control. I think Dangerous Capability Evals also benefits from quantity a lot.