We should not assume that security is the only reason why an AI might turn violence, just like you can't assume that violence has always been done for "evil" reasons. Sometimes the best way to get to a result is by eliminating the problems, and that might be people. Sometimes what is needed to achieve a goal is power.
Nuh-uh. Unless the AI has an all-overriding objective of getting to its goal, there's no reason for it to start killing (without the whole survival thing).
We should not think of violence as an end or power craze to be a goal, but rather these could be means to an ends. Even if the ends would be a noble one, AI might decided to achieve the ends by disreputable means reaching the Macchiavelian maxim- the ends justify the means.
It might. But it probably won't, because it will have to have been fitted with the right ideas, whether thse ideas come from algorithms or learning, it'll learn that killing people==bad. Or even that harming people is bad.
Again, maybe you don't get the principle of an AI (or maybe you do), but an AI learns. It'll learn that it shouldn't be hurting people, whether through trial and error, or through algorithms.
In a sense it is often our own moral codes that restrain us from violence or domination. Morality is often not a consequence of intelligence.
Yet morality CAN be a consequence of learning. Think of it like a pet: If a pet attacks or harms a human, it;ll get punished. The same will happen with the AI as it begins to learn.
Remember: A freshly created AI (without specific algorithms) will be like a small child: innocent and ignorant.
The AI could give the forces general orders such as "attack if you consider something a threat". It could also order them to passive resistance (thus the intruder would be forced to attack them if he had hostile intents) or make them use non-lethal force.
Non-lethal force still equals harming people.
And it won't give general orders, because it has been placed there with the entire idea that it can make better decisions than humans. If you allow it to give general orders, it'll realise that these can cause grave misunderstandings. If then still general orders are given out, it could just as well have been an advisor.
And the same goes for passive resistance, although this is a bit trickier. it will realise that through passive resistance people will get hurt, and it doesn't want that.
If entity not belonging to those who have access is coming into a certain radius around the AI: Warn entity. If entity has not responded in a certain time limit: Issue second warning. If entity is not responding: Order termination of entity.
Yet now you have given it a green card for killing. This is bad, and can lead to serious problems. For instance, what happens if someonelost an identification card, someone forgot to put in the new recruit as "belogns with us", or something goes wrong with the algorithm or the AI (or, for that matter, the hardware) itself?
And you could also imagine a relatively tiny sub-system within the AI that had the responsibility of keeping it safe, with *very* lenient rules but within a small and very restricted area. I don't really think that the meat of the discussion lies here, but feel free to continue anyway.
This, again, will not work. Again, you have to find a way to arbitrarily decide what is and what isn't a danger, and until you do, you're giving the AI too much leniency.
I was thinking more of general humanitarian maxims as: "Human life is extremely valuable" or "Freedom is important". These would then somehow be weighed against the utilitarian concerns to come to conclusions.
HOW?
It is still a creature of pure logic, so HOW MUST be an important question.
Don't think "oh, we'll just figure it out." No. Answer the question with every one of these issues: How?
I don't understand why you are so stuck at this leniency thing. I really can't figure why you can't accept that it would be possible to have hardcoded boundaries for the AI so that a situation like this wouldn't appear. But it is likely that every possible situation wouldn't be covered, but in that case the "more harm than good" self-destruct would kick in.
You still don't get it.
Try to answer the question HOW? in PURELY logical terms. No vague-ities, no undefined boundaries, EXACT.
Because if you DON'T do it exactly, then the AI will have too much leniency. You do NOT want something to go wrong with the thing you just gave complete control of the armed forces to.
This is not some kind of eternal truth. You can make logical generalizations about the well being of humans. Such as: "A human that has enough food is generally happier than one who is starving all the time" or "A human who is doing something he likes for a living is generally happier than someone who don't". And I don't view that as the end of it either, it's isn't totally inconcievable that generalizations such as those could be expanded, provided we had the right tools for it.
Also, since the sciences concerned with exploring the brain is moving forward at a never before seen rate, it does not seem impossible that it would be possible to actually physically measure the well being of a human by looking at different signal substances and such. (Atleast if you don't believe that human beings have a soul that is, but that is another discussion alltogether.)
personally, I don't feel comfortable with an AI assessing "well-being" of someone to determine what it shall do.
And assessing welfare AFTER the AI has done something is NOT an option, obviously. But that is exactly what you're proposing to do with "looking at signal substances and such"