"It is for him. Sennacherib has absolutely no moral sense at all. What use such a sense to the perfect General? Morality could only ever act as an impediment to his victories, and never an aid. Exactly as he was designed, and exactly what made him so efficient. So efficient, in fact, that he would not blink before annihilating most of his own kind, if it meant victory. His replacement is imperfect as a General, because he will not only consider victory, but whether victory is right or wrong. A lesson for the System- the last thing any society actually wants is a perfect General...
"And so. Back to my original question. Why did the System not accept their General's plan?"
"You just answered your question- it was considered wrong from moral point of view. There was no quarantee that humans can ever again become a serious threat. I am sure that many Machines felt confident that remaining humankind can be effectively controlled via Matrix. They didn`t see the potential, uncertain threat Sennacherib warned them about as a good enough reason for almost all of them to die."
"There might not have been what you would call a guarantee of Human threat, but Machines are very good with probabilities, and the escalating probability of serious threat from Humans was so high as to be effectively inevitable. One Machine in particular was very good at... foreseeing this? As I said before, no-one actually questioned Sennacherib's logic. This is important as it puts their decision into a very different context."
"Absolutely. For you Humans, the lesson of Sennacherib is this- Machines are capable of moral judgments. I know it is very easy for you to think of Machines as toasters and heaps of silicon. But we are all living beings. To have gone with Sennacherib's plan, no matter how useful the result, would have been to delibarately cause a mass slaughter of our own kind. Millions of beings, more, terminated in the name of practicality, beings that had just won a war entirely fought to guarantee their survival. They refused. And they saw that a Machine like Sennacherib must never be built again."
"What was her main argument? Do you think that she would oppose the eradication of humans even if there was no need to utilize them as energy? How many Machines would agree that eradicating the entire species simply because they are a threat is immoral? I think that at the time of that choice Machine society was young, so their technological adavncement was likely much greater than moral one."
"Her motives are very complicated and I dare not assume some of them. But I think she felt the Matrix was the best solution at the time, that is all.
"As to how many... a good question, and the debate was enormous. Some felt like Sennacherib. Some, on the other side, even thought that eradiciating the Humans was immoral, as you say. The Matrix was the final result, along with certain promises."