Alignment being solved at all would require alignment being solvable with human-level intelligence. Even though IQ-augmented humans wouldn’t be “superintelligent”, they would have additional intelligence that they could use to solve alignment. Additionally, it probably takes more intelligence to build an aligned superintelligence than it does to create a random superintelligence. Without alignment, chances are that the first superintelligence to exist will be whatever superintelligence is the easiest to build.
Alignment being solved at all would require alignment being solvable with human-level intelligence. Even though IQ-augmented humans wouldn’t be “superintelligent”, they would have additional intelligence that they could use to solve alignment. Additionally, it probably takes more intelligence to build an aligned superintelligence than it does to create a random superintelligence. Without alignment, chances are that the first superintelligence to exist will be whatever superintelligence is the easiest to build.
I don’t agree with the first statement neither understand what are you arguing for or against.