It also appears that the link to ELK in this section is incorrect
Making use of an AI’s internal state,2 not just its outputs. For example, giving positive reinforcement to an AI when it seems likely to be “honest” based on an examination of its internal state (and negative reinforcement when it seems likely not to be). Eliciting Latent Knowledge provides some sketches of how this might look.
It also appears that the link to ELK in this section is incorrect
Very belatedly fixed—thanks!