Jenny Shank interviews me about The Glass Cage over at MediaShift. The conversation gets into some topics that haven’t been covered much elsewhere, including my suggestion that Roomba, the automated vacuum cleaner, provides an early and ever so slightly ominous example of robot morality (or lack thereof). “Roomba makes no distinction between a dust bunny and an insect,” I write in the book. “It gobbles both, indiscriminately. If a cricket crosses its path, the cricket gets sucked to its death. A lot of people, when vacuuming, will also run over the cricket. They place no value on a bug’s life, at least not when the bug is an intruder in their home. But other people will stop what they’re doing, pick up the cricket, carry it to the door, and set it loose. … When we set Roomba loose on a carpet, we cede to it the power to make moral choices on our behalf.”
Here’s the relevant bit from the interview:
Shank: “The Glass Cage” made explicit for me a number of problems with automation that I had been vaguely worried about. But one thing that I had never worried about until reading “The Glass Cage” was the morality of the Roomba. You write, “Roomba makes no distinction between a dust bunny and an insect.” Why is it so easy to overlook the fact, as I did, that when a Roomba vacuums indiscriminately, it’s following a moral code?
Carr: It’s easier not to think about it, frankly. The workings of automated machines often raise tricky moral questions. We tend to ignore those gray areas in order to enjoy the conveniences the machines provide without suffering any guilt. But I don’t think we’re going to be able to remain blind to the moral complexities raised by robots and other autonomous machines much longer. As soon as you allow robots, or software programs, to act freely in the world, they’re going to run up against ethically fraught situations and face hard choices that can’t be resolved through statistical models. That will be true of self-driving cars, self-flying drones, and battlefield robots, just as it’s already true, on a lesser scale, with automated vacuum cleaners and lawnmowers. We’re going to have to figure out how to give machines moral codes even if it’s not something we want to think about.
Image: Juliette Culver.