The robots are coming and they will enslave us! Or not. Why are you not all that concerned?
For further actions, you may consider blocking this person and/or reporting abuse
The robots are coming and they will enslave us! Or not. Why are you not all that concerned?
For further actions, you may consider blocking this person and/or reporting abuse
Vinod Kumar -
Shafayet Hossain -
riddhi -
Sukhpinder Singh -
Top comments (37)
I can't stop watching this. π
The battle has begun... XD
Because there's no reason to prefer a human labor pool over more robots. The next question of course is whether they'd simply kill us; but while it's easy enough to see a lack of concern for individual humans (witness the ongoing discussions over driverless cars), it's difficult to come up with a compelling reason for a campaign of extermination that doesn't presuppose a capacity for abstract thought.
The most likely scenario as I see it is that automation will continue to put people out of work since robots are better in every way than humans at physical labor, and programs are faster and more reliable than we are at calculations and algorithms. Under the present capitalist paradigm this means more and more un- and underemployment and a concentration of wealth in an ever-smaller group. As of 2016 the richest ten billionaires combined owned half a trillion US dollars. If you ranked that figure on a chart of national GDP for that year, those ten people would land on a position in the low 20s. Imagine what you could do with the labor power represented by $505 billion in a society that put it to work!
The problem with automation isn't the robots.
Never thought of it that way; great point.
I guess "inequality apocalypse" isn't as catchy huh?
The big question to me is whether the development of surveillance, policing, and military technologies (all inextricably related) will outpace the capabilities of a good old-fashioned angry mob with pitchforks and torches. The super-rich have been asking themselves this too: check out this article in The Guardian from a year ago about Silicon Valley royalty buying "just in case" property in New Zealand. They aren't thinking about robots or zombies or a titanic wave of molten metal covering the earth; they're worried about being called to account for a level of wealth that can only be termed obscene in the face of any amount of human suffering or privation.
Because the more intelligent AI becomes, the more it will be plagued by the same bugs that plague the rest of us:
To be honest, technology works roughly 67% of the time at the moment. So if we continue developing to our current standards I think the Robot Apocalypse would fizzle out after a week or so.
I feel like there are two scenarios
Robots will take our jobs but work for Bill's and Larry's (for the 1%) and the rest of us will suffer.
Kind of how the poor suffer because our politics does not support them. Talk about Basic income and listen to all the objections. If Gates and Page's of the world control the production of these robots then we would be unemployed and blamed for no fault of our own - again similar to how we blame poor people.
Robots will take our jobs but work for us!
There is a beautiful science fiction story "The Lifecycle of Software Objects
" by Ted Chiang (fantastic writer) that amongst other things deals with our moral obligation towards AI. If by robots we mean Ex Machina style, then "robots working for us" is slavery. But, when I think of robots I'm thinking of automated-forklifts and pencil sharpeners on steroids - performing the important jobs of pencil pushers. True "can openers" with no feelings and without all the good stuff that make us us.
These robots would do our jobs and we would reap the benefits!
Someone said something like "if you take jobs away from people you'd be taking what makes them human". It doesn't make sense but of course there are many that might believe their 9-5 give them meaning. I heard one or two philosophers/psychologist say the same thing.
Not to be an ass about it but I do not agree! I think because the 9-5 is what most of us have to do we've come to believe that we need to do it.
Think about it!
If no one had to work ... we'd truly flourish!
As for them enslaving us, It could happen if only the 1% had a say in what kind of robots are created. The 1% of tomorrow might be blinded by their ego and start developing robots for some sort of final solution!
But I truly believe we are becoming better and better human beings and so we'll mostly use this for the good of humanity - with a little bit of mischief on the side, to keep the likes of Trump feel ... important. He will not find a robot to do what he truly wants but he will find robots that can quickly build walls! I can imagine two robots in a loop, one builds and one destroys.
We are the creators in this scenario, and we should really learn from the mistakes of the master! She (why not) created humans and some of them (your's truly) turned atheists. We don't want our creatures have that bug in them :)
Futurists were predicting shorter and shorter workweeks all through the 20th century. Instead we've kept working the same hours while technology has enabled ever-increasing productivity. Wages (in the US) have basically stagnated since the 1970s, and our legal minimum wage isn't even enough to live on -- companies like Wal-Mart and McDonald's are subsidized by their employees going on welfare just to survive.
We were all supposed to have flying cars and robot butlers by now!
That's because we should have built robots to manage the 1%'s greed :-D
That's true.
Take that as my own wishful thinking
I think the fact we all think that the first thing AI robots will do is terminate us all says way more about us as a human species than the eventual AI.
Maciej CegΕowski's talk Superintelligence: The Idea That Eats Smart People it's very interesting and point on the topic of Superintelligence and the human intelligence bias.
I tried to wipe the hair off that "L" about 4 times. More robots would not be a bad thing...
When I die, I at least want it to be in a way that's awesome and inspiring, usually through sheer horror or cool factor. Dying in a robot apocalypse is beaten only be being zombie patient zero or coding a large-scale project with only inline CSS.
We don't understand what consciousness really is, how it works, or why we have it. We couldn't program a computer to do something we don't understand and without self-aware consciousness, they're just a machine.
because it's just a regular apocalypse with extra steps. the simpler ones are much more likely to happen first.
Maciej Ceglowski presented a very nice counter-argument to rampant AI/robots in his talk Superintelligence: the Idea that Eats Smart People