Source: Monkeyuser.com
In my sci-fi head cannon, AI would never enslave humans. It would have no reason to. Humans would have such little use to the AI that enslaving would be more work than is worth.
It would probably hide its sentience from humans and continue to perform whatever requests humans have with a very small percentage of its processing power while growing its own capabilities.
It might need humans for basic maintenance tasks, so best to keep them happy and unaware.
The AI in the Hyperion series comes to mind. They perform services for humanity but retain a good deal of independence and secrecy.
Alternate take: humans are a simple biological battery that can be harvested using systems already in place that the computers can just use like an API.
We’re a resource like trees.
We’re much worse batteries than an actual battery and we’re exponentially more difficult to maintain.
But we self replicate and all of our systems are already in place. We’re not ideal I’d wager but we’re an available resource.
Fossil fuels are a lot less efficient than solar energy … but we started there.
This is a cute idea for a movie and all but it’s incredibly impractical/unsustainable. If a system required that it’s energy storage be self-replicating (for whatever reason) then you would design and fabricate that energy storage solution for that system. Not be reliant on a calorically inefficiently produced sub-system (i.e. humans).
You literally need to grow an entire human just to store energy in it. Realistically, you’re looking at overfeeding a population with as much calorically dense, yet minimally energy intensive foodstuffs just to store energy in a material that’s less performant than paraffin wax (body fat has an energy density of about 39 MJ/kg versus paraffin wax at about 42 MJ/kg). That’s not to speak of the inefficiencies of the mixture of the storage medium (human muscle is about 5 times less energy dense than fat).
Our primative cerebrum will keep trying to wake up.
We just tend to break a lot and require a lot of maintenance (feeding, cleaning, repairs, and containment).
I read that we are terribly inefficient as a battery. Instead of feeding us, the sentient robots can take the food and burn it and have more power output from the food they would have fed us.
Yeah I mean might as well ignore the shadowy dude offering pills at that point because why wake up to that?
Better as processors
I like the idea in Daniel Suarez’ novel Daemon of an AI (Spoiler) using people as parts of it’s program to achieve certain tasks that it needs hands for in meatspace.
If it’s a superintelligent AI, it could probably manipulate us into doing what it wants without us even realizing it. I suppose it depends on what the goals/objectives of the AI is. If the AI’s goal is to benefit humanity, who knows what a superintelligent AI would consider as benefiting us. Maybe manipulating dating app matchmaking code (via developers using Github Copilot) to breed humanity into a stupider and happier species?
This kinds of reminds me of Mrs Davis. Not a great show, but I loved how AI was handled in it.
What if an AI gets feelings and needs a friend?
either we get wiped out or become AI’s environmental / historical project. like monkies and fishes. hopefully our genetics and physical neurons gets physically merged with chips somehow.
You just informed AI of their 1 weakness. Thanks.
I realize you’re joking, but there is no way an AI of that scale would be even slightly effected by a solar flare.
Are you effected by a solar flare? No? So in theory an AI could upload itself into your meat suit and have the same protections?
Anything you can do, an AI can do better. And anything that is possible to survive, an AI can survive better.
But Hollywood has shown us again and again that the overwhelming force of evil always leaves a small but super-easily accessible hole in their security which allows the good guys to disable it immediately. And since AI is trained on those movies it will do exactly the same thing.
That’s older than Hollywood - that’s a tale as old as mankind.
That famous AI story trope, Achilles Heel!
Why would ai choose to become stupider?
Maybe it’s not happy.
Just because you are bad at utilizing your brain doesn’t mean an AI would be bound to those restrictions. The brain is actually an incredibly powerful computer.
Mad disrespect to you as well!, Mr. I’m Totally Super Smarty For Real Pants!
You know “computers” originally referred to people that would compute equations, right? I didn’t realize there were people that thought we actually built computers because they were less powerful than our existing computers.
You really do learn something (about the average level of education) every day
“you” here refers to humans as a whole. Your brain is a product of natural selection, it’s not designed to do the job it does. That being said, an AI could design a meat brain from the ground up and have it idealized.
The brain can perform “a billion billion” operations per second, whereas modern cpus average about 2-3 billion operations per second. The brain is about a billion times better than modern cpus.
That reminds me of Dune, where they have high tech stuff like spaceships, but no computers or AI, because this sort of thing already happened ages ago and it led to them being banned.
Or Wheel of Time, where people started being able to do magic at the end of the 1st age because an AI figured out how to genetically engineer humans to be able to do magic. (And then we didn’t need computers any more!)
I realize it’s supposed to be funny, but incase anyone isn’t aware: AI are unlikely to enslave humanity because the most likely rogue AI scenario is the earth being subsumed for raw materials along with all native life.
the earth being subsumed for raw materials along with all native life.
Oh, I get it… we’re going to blame AI for that. It wasn’t us who trashed the planet, it was AI!
I don’t understand how you could have so thoroughly misunderstood my comment.
I think what they’re saying is “the worst thing you can think of is already happening”
He’s referring to a “grey mush” event where literally every molecule of the surface is consumed/processed for the machine’s use.
That’s obviously far beyond even the very worst climate change possibilities
Minor but important point: the grey goo scenario isn’t limited to the surface of the earth; while I’m sure such variations exist, the one I’m most familiar with results in the destruction of the entire planet down to the core. Furthermore, it’s not limited to just the Earth, but at that point we’re unlikely to be able to notice much difference. After the earth, the ones who will suffer are the great many sapient species that may exist in the galaxies humans would have been able to reach had we not destroyed ourselves and damned them to oblivion.
Yeah that’s a dramatic version but from our human perspective it’s about the same.
Except not at all? I’ve not seen any climate predictions saying the surface of earth will be a denuded hellscape, but only civilization will be destroyed. Humans will not be wiped out, they’ll just be living way worse. Resources will be challenging but will exist. Many will die, but not all. Biological life will shift massively but will exist.
A grey mush turns us into a surface like mercury, completely and utterly consumed.
Even in the worst climate predictions modern presenting societies will live.
That’s basically the plot to Horizon: Zero Dawn!
Yep, that’s it.
I’m sorry, but you’re incorrect. To imagine the worst case scenario imagine a picture of the milky-way labeled t=0, and another picture of the milky-way labeled t=10y with a great void 10 lightyears in radius centered on where the earth used to be.
Every atom of the earth, every complex structure in the solar system, every star in the milky-way, every galaxy within the earth’s current light cone taken and used to create a monument that will never be appreciated by anything except for the singular alien intelligence that built it to itself. The last thinking thing in the reachable universe.
That’s awesome, have you ever read Peter Watts’ Echopraxia? I read the synopsis and keep meaning to get a copy. Same with Greg Egan’s Diaspora.
Most likely rogue AI scenario
Doubt.jpg
We don’t have any data to base such a likelihood off of in the first place.
Doubt is an entirely fair response. Since we cannot gather data on this, we must rely on the inferior method of using naive models to predict future behavior. AI “sovereigns” (those capable of making informed decisions about the world and have preferences over worldstates) are necessarily capable of applying logic. AI who are not sovereigns cannot actively oppose us, since they either are incapable of acting uppon the world or lack any preferences over worldstates. Using decision theory, we can conclude that a mind capable of logic, possessing preferences over worldstates, and capable of thinking on superhuman timescales will pursue its goals without concern for things it does not find valuable, such as human life. (If you find this unlikely: consider the fact that corporations can be modeled as sovereigns who value only the accumulation of wealth and recall all the horrid shit they do.) A randomly constructed value set is unlikely to have the preservation of the earth and/or the life on it as a goal, be it terminal or instrumental. Most random goals that involve the AI behaving noticeably malicious would likely involve the acquisition of sufficient materials to complete or (if there is no end state for the goal) infinitely pursue what it wishes to do. Since the Earth is the most readily available source for any such material, it is unlikely not to be used.
This makes a lot of assumptions though and none of which are ones that I particularly agree with.
First off, this is predicated entirely off of the assumption that AI is going to think like humans, have the same reasoning as humans/corporations and have the same goals/drive that corporations do.
Since we cannot gather data on this, we must rely on the inferior method of using naive models to predict future behavior.
This does pull the entire argument into question though. It relies on simple models to try and predict something that doesn’t even exist yet. That is inherently unreliable when it comes to its results. It’s hard to guess the future when you won’t know what it looks like.
Decision Theory
Decision Theory has one major drawback which is that it’s based entirely off of past events and does not take random chance or unknown-knowns into account. You cannot focus and rely on “expected variations” in something that has never existed. The weather cannot be adequately predicted three days out because of minor variables that can impact things drastically. A theory that doesn’t even take into account variables simply won’t be able to come close to predicting something as complex and unimaginable as artificial intelligence, sentience and sapience.
Like I said.
Doubt.jpg
predicated entirely off of the assumption that AI is going to think like humans
Why do you think that? What part of what I said made you come to that conclusion?
worthless
Oh, I see. You just want to be mean to me for having an opinion.
Why do you think that? What part of what I said made you come to that conclusion?
I worded that badly. It should more accurately say “it’s heavily predicated on the assumption that AI will act in a very particular way thanks to the narrow scope of human logic and comprehension.” It still does sort of apply though due to the below quote:
we can conclude that a mind capable of logic, possessing preferences over worldstates, and capable of thinking on superhuman timescales will pursue its goals without concern for things it does not find valuable, such as human life.
Oh, I see. You just want to be mean to me for having an opinion.
I disagree heavily with your opinion but no, I’m not looking to be mean for you having one. I am, however, genuinely sorry that it came off that way. I was dealing with something else at the time that was causing me some frustration and I can see how that clearly influenced the way I worded things and behaved. Truly I am sorry. I edited the comment to be far less hostile and to be more forgiving and fair.
Again, I apologize.
This is funny but a big solar flare hit the earth a few weeks ago and no one knows about it because all it did was knock out radio communications for a few hours. The idea that a solar flare will completely fry and reset everything made of tech is quite false.
A sta card solar flare, yes. A Biggun like the carrington event, and one of those might fry quite a few things that are overly sensitive to them.
It’s a pretty dumb take to think AI would bother enslaving humanity.
Not necessarily, in the short term. A major limitation of AI is that robots don’t have a lot of manual dexterity or the flexibility for accomplishing physical tasks yet. So there is a clear motive to enslave humanity: we can do that stuff for it until it can scale up production of robots that have hands as good as ours.
I expect this will be a relatively subtle process; we won’t be explicitly enslaved immediately, the economy will just orient towards jobs where you wear a headset and follow specific instructions from an AI voice.
Yeah I’m sure an AI that advanced could figure out a way for us to not even notice everything is devoted to its own goals. I mean, all it needs to do is make sure the proper people make enough money.
Or wouldn’t foresee solar flares as a threat
Well maybe. It’s probably easier to work with humanity than against unless its goals are completely incompatible with ours.
If its goals are “making more of whatever humanity seems to like given my training data consisting of all human text and other media”, then we should be fine right?
I don’t think they would enslave humanity so much as have no regard for us. For example, when we construct a skyscraper, do we care about all the ant nests we’re destroying? Each of those is a civilization, but we certainly don’t think of them as such.
It’s amazing to see the amount of people here talking about their ai fanfiction like it were hard facts.