I don’t know about “no way,” but the consensus is that simulation isn’t obviously very helpful because an AI could infer that it is simulated and behave differently in simulation, not to mention that sufficiently capable systems could escape simulation for the same reasons that ‘keep the AI in a box’ is an inadequate control strategy.
Simulation probably isn’t useless for safety, but it’s not obviously a top priority, and “the creation of an adequate AGI Sandbox” is prima facie intractable.
I have never been satisfied by the “AI infers that it is simulated and changes its behavior” argument because it seems like the root issue is always that some information has leaked into the simulation. The problem goes from, “how do we prevent AI from escaping a box?” to “How do we prevent information from entering a box?” The components of this problem being:
What information is communicated via the nature of the box itself?
What information is built into an AI.
What information is otherwise entering the box?
These questions seem relatively approachable compared to other avenues of AI safety research.
I don’t know about “no way,” but the consensus is that simulation isn’t obviously very helpful because an AI could infer that it is simulated and behave differently in simulation, not to mention that sufficiently capable systems could escape simulation for the same reasons that ‘keep the AI in a box’ is an inadequate control strategy.
Simulation probably isn’t useless for safety, but it’s not obviously a top priority, and “the creation of an adequate AGI Sandbox” is prima facie intractable.
I have never been satisfied by the “AI infers that it is simulated and changes its behavior” argument because it seems like the root issue is always that some information has leaked into the simulation. The problem goes from, “how do we prevent AI from escaping a box?” to “How do we prevent information from entering a box?” The components of this problem being:
What information is communicated via the nature of the box itself?
What information is built into an AI.
What information is otherwise entering the box?
These questions seem relatively approachable compared to other avenues of AI safety research.