My proposal is to engineer powerful and reliable AI immediately, as fast as feasible. If this is true endgame—whoever wins the race owns the planet if not the accessible universe—then spending and effort should be proportional. It’s the only way.
You deal with the dangerous out of control AI by tasking your reliable models with destroying them.
The core of your approach is to subdivide and validate all the subtasks. No model is manufacturing the drones used to do this by itself, it’s thousands of temporary instances. You filter the information used to reach the combat solvers that decide how to task each drone to destroy the enemy so any begging from the enemy is never processed. You design the killer drones with lots of low level interlocks to prevent the obvious misuse and they would use controllers maybe using conventional software so they cannot be convinced not to carry out the mission as they can’t understand language.
The general concept is if 99 percent of the drones are “safe” like this then even if escaped models are smart they just can’t win.
Or in more concrete terms, I am saying say a simple reliable combat solver is not going to be a lot worse than a more complex one. That superintelligence saturates. Simple and reliable hypersonic stealth drones are still almost as good as whatever a superintelligence cooks up etc. It’s an assumption on available utility relative to compute.
So my thought on this is I think of flamethrowers and gas shells and the worst ww1 battlefields. I am not sure what taboo humans won’t violate in order to win.