“Self-Driving Cars Are Merely The Vanguard Of An Approaching Fleet Of Equally Autonomous Devices”

In Jerry Kaplan’s excellent WSJ essay about ethical robots, which is adapted from his forthcoming book, Humans Need Not Apply, the author demonstrates it will be difficult to come up with consistent standards for our silicon sisters, and even if we do, machines following rules 100% of the time will not make for a perfect world. The opening:

As you try to imagine yourself cruising along in the self-driving car of the future, you may think first of the technical challenges: how an automated vehicle could deal with construction, bad weather or a deer in the headlights. But the more difficult challenges may have to do with ethics. Should your car swerve to save the life of the child who just chased his ball into the street at the risk of killing the elderly couple driving the other way? Should this calculus be different when it’s your own life that’s at risk or the lives of your loved ones?

Recent advances in artificial intelligence are enabling the creation of systems capable of independently pursuing goals in complex, real-world settings—often among and around people. Self-driving cars are merely the vanguard of an approaching fleet of equally autonomous devices. As these systems increasingly invade human domains, the need to control what they are permitted to do, and on whose behalf, will become more acute.

How will you feel the first time a driverless car zips ahead of you to take the parking spot you have been patiently waiting for? Or when a robot buys the last dozen muffins at Starbucks while a crowd of hungry patrons looks on? Should your mechanical valet be allowed to stand in line for you, or vote for you?•

 

Tags: