Book Read Free

Machines of Loving Grace

Page 24

by John Markoff


  At a TED talk in 2013, Horvitz showed the reaction of a Microsoft intern to her first encounter with his robotic greeter. He played a clip of the interaction from the point of view of the system, which tracked her face. The young woman approached the system and, when it told her that Eric was speaking with someone in his office and offered to put her on his calendar, she balked and declined the computer’s offer. “Wow, this is amazing,” she said under her breath, and then, anxious to end the conversation added, “Nice meeting you!” This was a good sign, Horvitz concluded, and he suggested that this type of interaction presages a world in which humans and machines are partners.

  Conversational systems are gradually slipping into our daily interactions. Inevitably, the partnerships won’t always develop in the way we have anticipated. In December of 2013, the movie Her, a love story starring Joaquin Phoenix and the voice of Scarlett Johansson, became a sensation. Her was a science-fiction film set in some unspecified not-far-in-the-future Southern California, and it told the story of a lonely man falling in love with his operating system. This premise seemed entirely plausible to many people who saw it. By the end of 2013 millions of people around the globe already had several years of experience with Apple’s Siri, and there is a growing sense that “virtual agents” are making the transition from novelties to the mainstream.

  Part of Her is also about the singularity, the idea that machine intelligence is accelerating at such a pace that it will eventually surpass human intelligence and become independent, rendering humans “left behind.” Both Her and Transcendence, another singularity-obsessed science-fiction movie introduced the following spring, are most intriguing for the way they portray human-machine relationships. In Transcendence the human-computer interaction moves from pleasant to dark, and eventually a superintelligent machine destroys human civilization. In Her, ironically, the relationship between the man and his operating system disintegrates as the computer’s intelligence develops so quickly that, not satisfied even with thousands of simultaneous relationships, it transcends humanity and . . . departs.

  This may be science fiction, but in the real world, this territory had become familiar to Liesl Capper almost a decade earlier. Capper, then the CEO of the Australian chatbot company My Cybertwin, was reviewing logs from a service she had created called My Perfect Girlfriend with growing horror. My Perfect Girlfriend was intended to be a familiar chatbot conversationalist that would show off the natural language technologies offered by Capper’s company. However, the experiment ran amok. As she read the transcripts from the website, Capper discovered that she had, in effect, become an operator of a digital brothel.

  Chatbot technology, of course, dates back to Weizenbaum’s early experiments with his Eliza program. The rapid growth of computing technology threw into relief the question of the relationship between humans and machines. In Alone Together: Why We Expect More from Technology and Less from Each Other, MIT social scientist Sherry Turkle expresses discomfort with technologies that increase human interactions with machines at the expense of human-to-human contact. “I believe that sociable technology will always disappoint because it promises what it can’t deliver,” Turkle writes. “It promises friendship but can only deliver ‘performances.’ Do we really want to be in the business of manufacturing friends that will never be friends?”20 Social scientists have long described this phenomenon as the false sense of community—“pseudo-gemeinschaft”—and it is not limited to human-machine interactions. For example, a banking customer might value a relationship with a bank teller, even though it exists only in the context of a commercial transaction and such a relationship might be only a courteous, shallow acquaintanceship. Turkle also felt that the relationships she saw emerging between humans and robots in MIT research laboratories were not genuine. The machines were designed to express synthetic emotions only to provoke or elucidate specific human emotional responses.

  Capper would eventually see these kinds of emotional—if not overtly sexual—exchanges in the interactions customers were having with her Perfect Girlfriend chatbots. A young businesswoman who had grown up in Zimbabwe, she had previously obtained a psychology degree and launched a business franchising early childhood development centers. Capper moved to Australia just in time to face the collapse of the dot-com bubble. In Australia, she first tried her hand at search engines and developed Mooter, which personalized search results. Mooter, however, couldn’t hold its own against Google’s global dominance. Although her company would later go public in Australia, she left in 2005, along with her business partner, John Zakos, a bright Australian AI researcher enamored since his teenage years with the idea of building chatbots. Together they built My Cybertwin into a business selling FAQbot technology to companies like banks and insurance companies. These bots would give website users relevant answers to their frequently asked questions about products and services. It proved to be a great way for companies to inexpensively offer personalized information to their customers, saving money by avoiding customer call center staffing and telephony costs. At the time, however, the technology was not yet mature. Though the company had some initial business success, My Cybertwin also had competitors, so Capper looked for ways to expand into new markets. They tried to turn My Cybertwin into a program that created a software avatar that would interact with other people over the Internet, even while its owner was offline. It was a powerful science-fiction-laced idea that yielded only moderately positive results.

  Capper has been equivocal and remains uncommitted about whether virtual assistants will take away human jobs. In interviews, she would note that virtual assistants don’t directly displace workers and would focus instead on mundane work her Cybertwins do for many companies, which she argued freed up humans to do more complex and ultimately more satisfying work. At the same time, Zakos attended conferences, making assertions that when companies ran A-B testing that compared the way the Cybertwins responded to text-based questions to the way humans in call centers responded to text-based questions, the Cybertwins outperformed the humans in customer satisfaction. They boasted that when they deployed a commercial system on the website of National Australia Bank, the country’s largest bank, more than 90 percent of visitors to the site believed that they were interacting with a human rather than a software program. In order to be convincing, conversational software on a bank website might need to answer about 150,000 different questions—a capability that is now easily within the range of computing and storage systems.

  Despite their unwillingness to confront the human job-displacement question, the consequences of Capper and Zakos’s work are likely to be dramatic. Much of the growth of the U.S. white-collar workforce after World War II was driven by the rapid spread of communications networks: telemarketers, telephone operators, and technical and sales support jobs all involved giving companies the infrastructure to connect customers with employees. Computerization transformed these occupations: call centers moved overseas and the first generation of automated switchboards replaced a good number of switchboard and telephone operators. Software companies like Nuance, the SRI spin-off that offers speaker-independent voice recognition, have begun to radically transform customer call centers and airline reservation systems. Despite consumers’ rejection of “voicemail hell,” system technology like My Cybertwin and Nuance will soon put at risk jobs that involve interacting with customers via the telephone. The My Cybertwin conversational technology might not be good enough to pass a full-on Turing test, but it was a step ahead of most of the chatbots that were available via the Internet at the time.

  Capper believes deeply that we will soon live in a world in which virtual robots are routine human companions. She holds none of the philosophical reservations that plagued researchers like Weizenbaum and Turkle. She also had no problem conceptualizing the relationship between a human and a Cybertwin as a master-slave relationship.21 In 2007 she began to experiment with programs called My Perfect Boyfriend and My Perfect Girlfriend. Not surprisingly, there
was substantially more traffic on the Girlfriend site, so she set up a paywall for premium parts of the service. Sure enough, 4 percent of the people—presumably mostly men—who had previously visited the site were willing to pay for the privilege of creating an online relationship. These people were told that there was nothing remotely human on the other end of the connection and that they were interacting with an algorithm that could only mimic a human partner. Indeed, they were willing to pay for this service, even though already at the time there was no shortage of “sex chat” websites with actual humans on the other end of the conversation.

  Maybe that was the explanation. Early in the personal computer era, there was a successful text-adventure game publisher called Infocom whose marketing slogan was: “The best graphics are in your head.” Perhaps the freedom of interacting with a robot relaxed the mind precisely because there was no messy human at the other end of the line. Maybe it wasn’t about a human relationship at all, but more about having control and being the master. Or, perhaps, the slave.

  Whatever the psychology underpinning the interactions, it freaked Capper out. She was seeing more of the human psyche than she had bargained for. And so, despite the fact that she had stumbled onto a nascent business, she backed away and shut down My Perfect Girlfriend in 2014. There must be a better way of building a business, she decided. It would turn out that Capper’s business sense was well timed. Apple’s embrace of Siri had transformed the market for virtual agents. The computing world no longer understood conversational systems as quirky novelties, but rather as a legitimate mainstream form of computer interaction. Before My Perfect Girlfriend, Capper had realized that her business must expand to the United States if it was to succeed. She raised enough money, changed the company’s name from My Cybertwin to Cognea, and set up shop in both Silicon Valley and New York. In the spring of 2014, she sold her company to IBM. The giant computer firm followed its 1997 victory in chess over Garry Kasparov with a comparable publicity stunt in which one of its robots competed against two of the best human players of the TV quiz show Jeopardy! In 2011, the IBM Watson system triumphed over Brad Rutter and Ken Jennings. Many thought the win was evidence that AI technologies had exceeded human capabilities. The reality, however, was more nuanced. The human contestants could occasionally anticipate the brief window of time in which they could press the button and buzz in before Watson. In practice, Watson had an overwhelming mechanical advantage that had little to do with artificial intelligence. When it had a certain statistical confidence that it had the correct answer, Watson was able to press the button with unerring precision, timing its button press with much greater accuracy than its human competitors, literally giving the machine a winning hand.

  The irony with regards to Watson’s ascendance is that IBM has historically portrayed itself as an augmentation company rather than a company that sought to replace humans. Going all the way back to the 1950s, when it terminated its first formal foray into AI research, IBM has been unwilling to advertise that the computers it sells often displace human workers.22 In the wake of its Watson victory, the company portrayed its achievement as a step toward augmenting human workers and stated that it planned to integrate Watson’s technology into the health-care field as an intellectual aid to doctors and nurses.

  However, Watson was slow to take off as a physicians’ advisor, and the company has broadened its goal for the system. Today the Watson business group is developing applications that will inevitably displace human workers. Watson had originally been designed as a “question-answering” system, making progress toward the fundamental goals in artificial intelligence. With Cognea, Watson gained the ability to carry on a conversation. How will Watson be used? The choice faced by IBM and its engineers is remarkable. Watson can serve as an intelligent assistant to any number of professionals, or it can replace them. At the dawn of the field of artificial intelligence IBM backed away from the field. What will the company do in the future?

  Ken Jennings, the human Jeopardy! champion, saw the writing on the wall: “Just as factory jobs were eliminated in the 20th century by new assembly-line robots, Brad and I were the first knowledge-industry workers put out of work by the new generation of ‘thinking’ machines. ‘Quiz show contestant’ may be the first job made redundant by Watson, but I’m sure it won’t be the last.”23

  7|TO THE RESCUE

  The robot laboratory was ghostly quiet on a weekend afternoon in the fall of 2013. The design studio itself could pass for any small New England machine shop, crammed with metalworking and industrial machines. Marc Raibert, a bearded roboticist and one of the world’s leading designers of walking robots, stood in front of a smallish interior room, affectionately called the “meat locker,” and paused for effect. The room was a jumble of equipment, but at the far end seven imposing humanoid robots were suspended from the ceiling, as if on meat hooks. Headless and motionless, the robots were undeniably spooky. Without skin, they were cybernetic skeleton-men assembled from an admixture of steel, titanium, and aluminum. Each was illuminated by an eerie blue LED glow that revealed a computer embedded in the chest that monitored its motor control. Each of the presently removed “heads” housed another computer that monitored the body’s sensor control and data acquisition. When they were fully equipped, the robots stood six feet high and weighed 330 pounds. When moving, they were not as lithe in real life as they were in videos, but they had an undeniable presence.

  It was the week before DARPA would announce that it had contracted Boston Dynamics, the company that Raibert had founded two decades earlier, to build “Atlas” robots as the common platform for a new category of Grand Challenge competitions. This Challenge aimed to create a generation of mobile robots capable of operating in environments that were too risky or unsafe for humans. The company, which would be acquired by Google later that year, had already developed a global reputation for walking and running robots that were built mostly for the Pentagon.

  Despite taking research dollars from the military, Raibert did not believe that his firm was doing anything like “weapons work.” For much of his career, he had maintained an intense focus on one of the hardest problems in the world of artificial intelligence and robotics: building machines that moved with the ease of animals through an unstructured landscape. While artificial intelligence researchers have tried for decades to simulate human intelligence, Raibert is a master at replicating the agility and grace of human movement. He had long believed that creating dexterous machines was more difficult than many other artificial intelligence challenges. “It is as difficult to reproduce the agility of a squirrel jumping from branch to branch or a bird taking off and landing,” Raibert argued, “as it is to program intelligence.”

  The Boston Dynamics research robots, with names like LittleDog, BigDog, and Cheetah, had sparked lively and occasionally hysterical Internet discussion about the Terminator-like quality of modern robots. In 2003 the company had received its first DARPA research contract for a biologically inspired quadruped robot. Five years later, a remarkable video on YouTube showed BigDog walking over uneven terrain, skittering on ice, and withstanding a determined kick from a human without falling. With the engine giving off a banshee-like wail, it did not take much to imagine being chased through the woods by such a contraption. More than sixteen million people viewed the video, and the reactions were visceral. For many, BigDog exemplified generations of sinister sci-fi and Hollywood robots.

  Raibert, who usually wears jeans and Hawaiian shirts, was unfazed by, and even enjoyed, his Dr. Evil image. As a rule, he would shy away from engaging directly with the media, and communicated instead through a frequent stream of ever more impressive “killer” videos. Yet he monitored the comments and felt that many of them ignored the bigger picture: mobile robots were on the cusp of becoming a routine part of the way humans interact with the world. When speaking on the record, he simply said that he believed his critics were missing the point. “Obviously, people do find it creepy,” he told a Britis
h technical journal. “About a third of the 10,000 or so responses we have to the BigDog videos on YouTube are from people who are scared, who think that the robots are coming for them. But the ingredient that affects us most strongly is a sense of pride that we’ve been able to come so close to what makes people and animals animate, to make something so lifelike.”1 Another category of comments, he pointed out, was from viewers who feigned shock while enjoying a sci-fi-style thrill.

  The DARPA Robotics Challenge (DRC) underscored the desired spectrum of possibilities for the relationship between humans and robots even more clearly than the previous Grand Challenge for driverless cars. It foreshadowed a world in which robots would partner with humans, dance with them, be their slaves, or potentially replace them entirely. In the initial DRC competition in 2013, the robots were almost completely teleoperated by a human reliant on the robot’s sensor data, which was sent over a wired network connection. Boston Dynamics built Atlas robots with rudimentary motor control capabilities like walking and arm movements and made them available to competing teams, but the higher-level functions that the robots would need to complete specific tasks were to be programmed independently by the original sixteen teams. Later that fall, when Boston Dynamics delivered the robots to the DRC, and also when they actually competed in a preliminary competition held in Florida at the end of the year, the robots proved to be relatively slow and clumsy.

  Hanging in the meat locker waiting to be deployed to the respective teams, however, they looked poised to spring into action with human nimbleness. On a quiet afternoon it evoked a scene from the 2004 movie I, Robot, where a police detective played by actor Will Smith walks, gun drawn, through a vast robot warehouse containing endless columns of frozen humanoid robots awaiting deployment. In a close-up shot, the eyes of one sinister automaton focus on the moving detective before it springs into action.

 

‹ Prev