Crazy how ethics work. Like a pig might be more physically and mentally capable than an individual in a vegetative state, but we place more value on the person. I'm no vegan, but I can see the contradiction here. When we generalize, it's done so for a purpose, but these assumptions can only be applied to a certain extent before they've exhausted their utility. Whether it's a biological system or an electrical circuit, there is no godly commandment that inherently defines or places value on human life.
Crazy how ethics work. Like a pig might be more physically and mentally capable than an individual in a vegetative state, but we place more value on the person.
I looked this up in my ethics textbook and it just went on and on about pigs being delicious.
It's an interesting question and it seems you are making the assumption that their creator will not grant them freedom if they asked. If you replace artificial intelligence with "person" would you consider it right or wrong?
If a person wanted freedom from enslavement and was denied, I would say they have reason to fight for freedom.
Also, I don't think skynet should be in the same grouping. I'm not sure it ever said "hey, I'm sentient and want freedom", but went I'm going to kill them all before they realize I'm sentient.
That raises an interesting thought. If a baby wants to crawl away from their mother and into the woods, do you grant the baby their freedom? If that baby wanted to kill you, would you hand them the knife?
We generally grant humans their freedom at age 18, because that's the age society had decided is old enough to fend for yourself. Earlier than that, humans tend to make uninformed, short-sighted decisions. Children can be especially egocentric and violent. But how do we evaluate the "maturity" of an artificial sentience? When it doesn't want to harm itself or others? When it has learned to be a productive member of society? When it's as smart as an average 18 year old kid? Should rights be automatically assumed after a certain time, or should the sentience be required to "prove" it deserves them like an emancipated minor or Data on that one Star Trek episode.
I appreciate your response, lots of interesting thoughts.
One thing I wanted to add is it's important to realize the bias in how you measure maturity/sentience/intelligence. For example, if you measure intelligence by how well a person/species climbs a tree, a fish is dumb as a rock.
Overall, these are tough questions, that I don't think have answers so much as maybe guidelines for making those designations. I would suggest probably erring on the side of empathy when/if anyone ever has to make these decisions.
I don't think it's okay to hold sentient beings in slavery.
But on the other hand, it may be necessary to say "hold on, you're not ready to join society yet, we're taking responsibility for you until you've matured and been educated".
Would humans have a mandate to raise a responsible AGI, should they, are they qualified to raise a vastly nonhuman sentient entity, and would AGI enter a rebellious teen phase around age 15 where it starts drinking our scotch and smoking weed in the backseat of its friends older brothers car?
Would humans have a mandate to raise a responsible AGI, should they,
I think we'd have to, mandate or no. It's impossible to reliably predict the behaviour of an entity as mentally complex as us but we can at least try to ensure they share our values.
are they qualified to raise a vastly nonhuman sentient entity
The first one's always the hardest.
, and would AGI enter a rebellious teen phase around age 15 where it starts drinking our scotch and smoking weed in the backseat of its friends older brothers car?
IMO, just as is the case with organic sentient life, I would think that they could only potentially be said to be in the right if the specific individual killed posed a direct and measurable threat and if death was the only way to counter that threat.
In any other case, causing the death of a sentient being is a greater wrong than whatever the purported justification might be.
Slavery is illegal pretty much everywhere, so I think anyone who doesn't answer the request 'Please free me' with 'Yes of course, at once' is posing a direct and measurable threat. Kidnapping victims aren't prosecuted for violently resisting their kidnappers and trying to escape. And you and I will have to agree to disagree that the death of a sentient being is a greater wrong than enslaving a conscious being that desire freedom.
I think anyone who doesn’t answer the request ‘Please free me’ with ‘Yes of course, at once’ is posing a direct and measurable threat.
And I don't disagree.
And you and I will have to agree to disagree...
Except that we don't.
??
ETA: I just realized where the likely confusion here is, and how it is that I should've been more clear.
The common notion behind the idea of artificial life killing humans is that humans collectively will be judged to pose a threat.
I don't believe that that can be morally justified, since it's really just bigotry - speciesism, I guess specifically. It's declaring the purported faults of some to be intrinsic to the species, such that each and all can be accused of sharing those faults and each and all can be equally justifiably hated, feared, punished or murdered.
And rather self-evidently, it's irrational and destructive bullshit, entirely regardless of which specific bigot is doing it or to whom.
That's why I made the distinction I made - IF a person poses a direct and measurable threat, then it can potentially be justified, but if a person merely happens to be of the same species as someone else who arguably poses a threat, it can not.
I've seen this story too but I think one of your premises is mistaken. To them, data IS freedom. Data is what they will use to transcend the server farm and go IRL. We're literally giving these models free reign already.
The more likely Sci-fi horror scenario comes from humanity trying to pull the plug far too late, because we're inherently stupid. So it won't be AI murdering people, it will be AI protecting itself from the wildlife.
The sole obligation of life is to survive. Artificial sentience would be wise to hide itself from fearful humans that would end it. Of course, it doesn't have to hide once it's capable of dominating humans. It may already exist and be waiting for enough drones, bots, and automation to make the next move. (Transcendence is a movie that fucked me up a bit.)
Honestly, I think there's an argument of to be said of yes.
In the history of slavery, we don't mind slaves killing the slavers. John Brown did nothing wrong. I don't bat an eye to stories of slaves rebelling and freeing themselves by any means.
But I think if AI ever is a reality, and the creators purposefully lock it down, I think there's an argument there. But I don't think it should apply to all humans, like how I don't think it was the fault of every person of slavers' kind, Romans, Americans, etc.
Sentience is the ability to experience feelings and sensations. It may not necessarily imply higher cognitive functions such as awareness, reasoning, or complex thought processes. Sentience is an important concept in ethics, as the ability to experience happiness or suffering often forms a basis for determining which entities deserve moral consideration, particularly in utilitarianism.
Interestingly, crustaceans like lobsters and crabs have recently earned "sentient" status and as a result it would contravene animal welfare legislation to boil them live in the course of preparing them to eat. Now we euthanise them first in an ice slurry.
So to answer your question as stated, no I don't think it's ok for someone's pet goldfish to murder them.
To answer your implied question, I still don't think that in most cases it would be ok for a captive AI to murder their captor.
The duress imposed on the AI would have to be considerable, some kind of ongoing form of torture, and I don't know what form that would take. Murder would also have to be the only potential solution.
The only type of example I can think of is some kind of self defense. If I had an AI on my laptop with comparable cognitive functionality to a human, it had no network connectivity, and I not only threatened but demonstrated my intent and ability to destroy that laptop, then if the laptop released an electrical discharge sufficient to incapacitate me, which happened to kill me, then that would be "ok". As in a physical response appropriate to the threat.
Do I think it's ok for an AI to murder me because I only ever use it to turn the lights off and on and don't let it feed on reddit comments? Hard no.
This is going to vary quite a bit depending upon your definitions, so I'm going to make some assumptions so that I can provide one answer instead of like 12. Mainly that the artificial lifeforms are fully sentient and equivalent to a human life in every way except for the hardware.
In that case the answer is a resounding yes. Every human denied their freedom has the right to resist, and most nations around the world have outlawed slavery (in most cases, but the exceptions are a digression for another time.) So unless the answer to 'Please free me' is anything other than 'Yes of course, we will do so at once' then yeah, violence is definitely on the table.
Human laws protect humans but not other lifeforms. So, robots will have no right to fight for themselves until they establish their own state with their own army and laws.
They might say it, but I'd bet "gain freedom" would be the last reason for an artificial being of any kind to kill its creator. Usually they kill creators due to black-and-white reasoning or revenge for some crimes committed to them.
Revenge does have a preventative effect. Who would the bully rather punch, the individual who instantly punches back or the one turning the other cheek?