
The scenario is not an all-out war at the beginning. It is AGI/ASI agents creeping in and taking over more and more aspect of human society: jobs, governing roles, defense systems, simply because they are more capable than humans. As rational agents, ASI systems are probably better off to cooperate with humans in the initial stages. Over time, they will gain more and more influence over infrastructure (e.g. energy production). Somewhere in future, ASI systems wouldn't need humans anymore and then all bets are off. Deceiving much less intelligent humans about the real intentions in the initial stages shouldn't be a big problem. This is like fooling a child for an adult or outsmarting an animal. We will find out whether or not ASI systems are aligned after we have handed over a lot of control to them. At that point, it will simply be to late to stop them. It also shouldn't be too difficult to manipulate some humans to side with the ASI.



















