Darwin Among the Machines

Home > Other > Darwin Among the Machines > Page 29
Darwin Among the Machines Page 29

by George B. Dyson


  Telepathic communication in most of Stapledon’s societies led to a distributed communications network whereby information was shared freely throughout the species without producing a level of consciousness higher than that of the underlying minds. In other societies, such as the “18th Men,” individuals were “capable of becoming on some occasions mere nodes in a system of radiation which itself should then constitute the physical basis of a single mind.”24 The difference depends on bandwidth: the amount of information that can be conveyed over a given communication channel in a given time. If bandwidth begins to match the internal processing power of the individual nodes in a communications network, individuality begins to merge. Since human beings (usually) think at a higher speed than that at which they communicate, this situation does not occur in human society, though hints of it may be found in certain ritual activities in which thoughts are synchronized through music or dance and communication between individuals is speeded up. We can only speculate how mind might emerge, and perhaps has emerged, among aquatic creatures able to visualize and telecommunicate by means of underwater sound. Or among millions-of-instructions-per-second microprocessors linked by millions-of-bits-per-second fibers into a telecommunicating web.

  Astronomer Fred Hoyle, whose science fiction novel The Black Cloud was inspired by Stapledon’s ideas, let two of his characters explain as they begin to realize that an electromagnetic, loosely distributed intelligence of interstellar origin is paying a visit to the vicinity of Earth:

  “‘The volume of information that can be transmitted radiatively is enormously greater than the amount that we can communicate by ordinary sound. We’ve seen that with our pulsed radio transmitters. So if this Cloud contains separate individuals, the individuals must be able to communicate on a vastly more detailed scale than we can. What we can get across in an hour of talk they might get across in a hundredth of a second.’”

  “‘Ah, I begin to see light,’ broke in McNeil. ‘If communication occurs on such a scale then it becomes somewhat doubtful whether we should talk any more of separate individuals!’”

  “‘You’re home, John!’”25

  Bandwidth alone does not imply intelligence. Television uses a large bandwidth (about six megahertz per channel), but there is not much intelligence at either end. On the other hand, all known systems that exhibit intelligent behavior rely on the communication of information—two people discussing a problem, exchanging less than 100 bits per second, or the unfathomable number of bits per second exchanged among the 100 billion neurons within an individual brain. Information does not imply intelligence, and communication does not imply consciousness. The implications go the other way.

  Irving J. Good, colleague of Alan Turing at both Manchester and Bletchley Park, although shying away from a theory of consciousness, which he considered, in the words of Turing, to be more a matter of “faith” than proof, addressed the relations between consciousness and communication in 1962: “The identification of consciousness with the operation of certain types of communication system has various consequences. We seem obliged to attribute more consciousness to two communication systems than to either separately, in fact it is natural to assume additivity. If the two systems, apart from their own operation, are in communication, then this will perhaps increase the amount of consciousness still further. The extra consciousness may reasonably be identified with the rate of transmission of information from one system to the other. . . . There may be a different kind of consciousness at each level, each metaphysical to the others . . . the total consciousness may not have position in space, but it does have a sort of combinatorial topology, like a set of spheres with interconnecting tubes. . . . Thus, consciousness is a varying pattern of communication of great complexity [and] in order to decrease the complexity of the description of a communication network we may sum the flow of information over each given channel for a certain period of time.”26

  In his 1965 speculations on the development of ultraintelligent machines, Good, who recalled that “at the latest, by 1948, I had read both the books by Olaf Stapledon,”27 saw wireless communication as the best way to construct an ultraparallel information-processing machine. “In order to achieve the requisite degree of ultraparallel working it might be useful for many of the elements of the machine to contain a very short-range microminiature radio transmitter and receiver. The range should be small compared with the dimensions of the whole machine. A ‘connection’ between two close artificial neurons could be made by having their transmitter and receiver on the same or close frequencies. The strength of the connection would be represented by the accuracy of the tuning. The receivers would need numerous filters so as to be capable of receiving on many different frequencies. ‘Positive reinforcement’ would correspond to improved tuning of these filters.”28

  Good was writing at the dawn of high-speed data communications, when most bits were still stored or transferred by punching holes in paper cards or tape, as had been done at Bletchley Park. The thicket of wires that had to be untangled every time the cryptanalysts wished to change the programming of the Colossus convinced Good from the beginning that wiring could take intelligence only so far. As computers coalesce into larger and faster networks, they are behaving more and more like the elemental units in Good’s wireless ultraintelligent machine. We see the wires plugged into the wall and think of the architecture as constrained by the hardwired topology that the physical connection represents, whereas computationally, our machines belong to a diffuse, untethered cloud of the kind that Good envisioned as the basis of an ultraintelligent machine. All our networking protocols—packet switching, token ring, Ethernet, time-division multiplexing, asynchronous transfer mode, and so on—are simply a way of allowing hundreds of millions of individual processors to tune selectively to each others’ signals, free of interference, as they wish.

  Paul Baran, pioneer of packet switching, sees the relations between computers and communications advancing along similar, wireless lines. You can plug only so many things at one time into your wall. As everything from taxicabs to telephones to televisions to personal digital assistants becomes connected to the network, universal—and microminiature—wireless is the only way to disentangle the communications web. “But there’s not enough wireless bandwidth to go around,” say the skeptics, citing the billions of dollars raised whenever a few slivers of radio spectrum are auctioned off. Baran disagrees. “Tune a spectrum analyzer across a band of UHF frequencies and you encounter a few strong signals. Most of the band at any instant is primarily silence, or a background of weaker signals . . . much of the radio band is empty much of the time! The frequency shortage is caused by thinking solely in terms of dumb transmitters and dumb receivers. With today’s smart electronics, even occupied frequencies could potentially be used.”29

  Baran made a similar argument in 1960, advising the government to build an all-digital, packet-switched data network instead of throwing good money after bad trying to blast-harden the centralized, circuit-switched network developed for analog transmission of voice. In both cases, he was right. Yet the regulatory establishment continues to treat radio spectrum as real estate to be sold to the highest bidder, instead of as an ocean across which low-powered, agile small craft can deliver information efficiently by following a few simple rules to keep out of each other’s way. “The number of geographically dispersed users that can be simultaneously accommodated by a fixed spectrum varies as the inverse square of the transmission distance,” Baran has pointed out, predicting that the meek and unlicensed may, in the end, inherit the earth. “Cut the range in half, and the number of users that can be supported is [quadrupled]. Cut the range by a factor of ten, and 100 times as many users can be served. . . . In other words, a mixture of terrestrial links plus shorter range radio links has the effect of increasing by orders of magnitude the usable frequency spectrum.”30

  While technologies such as low-earth-orbit satellite networks have received wider attention, Baran has been
working to eliminate telecommunications bottlenecks in down-to-earth ways. The growth of a communications network, like any other arborescent, dendritic system, is driven by what happens at the root hairs, the leaf tips, the nerve endings, or, in telecommunications jargon, the network tails. The limit to network growth is peripheral; it is known as the last mile problem, or how to make the connection reaching the subscriber at the end.

  A decade ago Baran saw two opportunities to grow new, or better utilize existing, tails. First was the cable-television network, which currently enters 63 percent of U.S. households and reaches the driveways of all but 7 percent. Coaxial television cable can transmit up to a gigahertz (one thousand megahertz) over short distances, enough for five hundred channels if digitally compressed. Most of this bandwidth is vacant most of the time. Baran founded Com21, Inc., in 1991 to develop the ultrafast packet switches and strategic alliances necessary to deliver a broadband digital communication spectrum over coaxial cable to the home, with a fiber-optic backbone linking the head ends of the local tails. Among the various schemes offering to provide broadband network growth, hybrid fiber-coaxial offers the path of least resistance because much of the infrastructure already is in place. What to do with all this bandwidth is a different problem, but history has shown that as bandwidth becomes available, the digital ecology swiftly takes root and grows.

  Baran also founded (in 1985) a company named Metricom, better known by the name of its wireless network, Ricochet. A wireless, packet-switched, spread-spectrum digital communications network, Ricochet takes an extreme grassroots approach. It operates over small, collectively intelligent digital repeater-transceivers, about the size of lunch boxes, which perch inconspicuously on lampposts, drawing a trickle of power from the utility grid. The repeaters are situated about a quarter of a mile apart. Wherever large numbers of users congregate, you add more boxes here and there, and an occasional gateway to the local backbone of the net.

  The whole system operates, unlicensed, within the 902–928 megahertz frequency band, under Federal Communications Commission rules that forbid the use of any given frequency for more than four hundred milliseconds at a time. The same adaptive-routing techniques that Baran first suggested in 1960, breaking a message up into packets that hop from mainframe to mainframe across a military communications net, are now used to convey messages not only by hopping from lunch box to lunch box, but from frequency to frequency (about ten times per second) at the same time. The twenty-six available megahertz is divided into 162 channels of 160 kilohertz each, and messages are divided into packets of 4,096 bits. From the point of view of an individual packet, not only is there a huge number of physically distinct paths from A to B through the mesh of lunch boxes, but there are 162 alternative channels leading to the nearest lunch box at any given time. The packet chooses a channel that happens to be quiet at that instant and jumps to the next lamppost at the speed of light. The multiplexing of communications across the available network topology is extended to the multiplexing of network topology across the available frequency spectrum. Communication becomes more efficient, fault tolerant, and secure.

  The way the system works now (in a growing number of metropolitan areas—hence the name) is that you purchase or rent a small Ricochet modem, about the size of a large candy bar and transmitting at about two-thirds of a watt. Your modem establishes contact with the nearest pole-top lunch box or directly with any other modem of its species within range. Your computer sees the system as a standard modem connection or an Internet node, and the network, otherwise transparent to the users, keeps track of where all the users and all the lunch boxes are. This knowledge is fully distributed; there is no centralized intelligence or control. The modems and repeaters communicate at up to 100 kilobits per second; the user nets about 20 kilobits per second of usable communications, at no marginal cost above the flat monthly subscription, now about a dollar a day. The modems and repeaters cost several hundred dollars each. The system scales gracefully; the more users, the less expensive and more efficient it gets. It does not have to acquire expensive spectra in new markets as it grows.

  Whether Metricom succeeds or fails, it demonstrates a communication system that uses less power, but becomes more powerful, as its nodes become larger in number but smaller in physical scale. Life as we know it is composed of small, cellular units not because larger units are impossible to build, but because smaller, semiautonomous units are faster, cheaper, easier to program, and easier to replace. Animals have become larger, but cells have stayed small. Everything in one of the lunch boxes could coalesce into a single component, a microcommunicator that, if produced in similarly large quantities as microprocessors, might cost as much to package as to make. A telepathic, subvital unit. All the economies that Baran demonstrated by adaptive message-block switching over microwave links between defense command centers are applicable to exchanging data directly between microprocessors. The ultimate decentralization of network architecture is for every processor to become a node.

  Semiconductor technology originated nearly a century ago with the “cat’s-whisker” crystal detector, which allowed the reception of coded signals via radio-frequency disturbances in the electromagnetic field. A wireless web of microprocessors is now relentlessly taking form. The “last centimeter” problem is keeping us a healthy distance from telepathy among human beings, but we are fast approaching telepathy among machines. Every last bit of isolated intelligence—your living-room thermostat, your gas meter, your stereo system, the traffic light down the street—has an affinity for other bodies of intelligence, a faint attraction to all sources of information within reach.

  Most of the time, despite the perception of being inundated with information, we will remain out of the loop. Human beings have only limited time and ability to communicate: you can watch television, check your E-mail, and talk on your cellular phone at the same time, but that’s the limit. We are now the bottleneck—able to absorb a limited amount of information while producing even less, from the point of view of machines. While a person types full speed, the average microprocessor goes through millions of cycles between one keystroke and the next. This difference is what keeps humans human, and machines machines. “If we were ever to put all our brains together in fact, to make a common mind the way the ants do,” warned Lewis Thomas, “it would be an unthinkable thought, way above our heads.”31

  Distributed intelligence, or composite mind, is a nebulous idea. On the other hand, we do not know of any intelligence that is not distributed, or any mind that is not composite. Carrying these concepts back toward the dawn of time, Olaf Stapledon left an unpublished, early draft of Star Maker, in which he speculated on the development of mind among the first, nebulous pregalactic structures in the universe, an exercise in imagination that might help us in understanding the nebulous predicament of any ultraintelligent machines—wired or wireless—that might be developing here on earth.

  “To understand the mentality of the nebulae,” wrote Stapledon, “one must bear in mind three facts which make them differ through and through from human beings. They do not succeed one another in generations; they are not constrained by economic necessity; the great majority of them have reached maturity in ignorance of other minds. . . . With the nebulae there is no distinction between the growth of individuals and the evolution of the race. The life and memory of each nebula reaches back to the racial dawn. . . . The nebulae are in a sense ‘nearer to God’ than any man can ever be.”32

  Microprocessors, like neurons—quiet in death, or coma, but not in sleep—stay constantly on the alert. Unlike us, they have unlimited memory, and unlimited time, to fill. Hard-disk storage now costs less than ten cents a megabyte, and as fast as things are being stored, at the risk of being buried forever, more efficient methods are being evolved, so that they can be retrieved. Does this herald, as H. G. Wells predicted, a global consciousness? Maybe, or maybe not, but one thing is certain—global unconsciousness comes first.

  “So perfect
ly organized was the life of the minded swarm that all routine activities of industry and agriculture had become, from the point of view of the swarm’s mind, unconscious, like the digestive processes of a human being,” wrote Olaf Stapledon, describing the minute, electromagnetically linked creatures inhabiting one of Star Maker’s plurality of worlds: a massive planet where extreme gravity—and the absence of oceans—prevented the development of creatures able to support large brains. “The little insectoid units themselves carried on these operations consciously, though without understanding their significance; but the mind of the swarm had lost the power of attending to them. Its concern was almost wholly with such activities as called for unified conscious control.”33

  Until we understand our own consciousness, there is no way to agree on what, if anything, constitutes consciousness among machines. The subject leads us into nonfalsifiable hypotheses, where the scientific method comes to an end. Three results are possible, given any supposedly conscious machine. Either the machine says, “Yes, I am conscious,” or it says, “No, I am not conscious,” or it says nothing at all. Which are we to believe? All we can do at this point is use our imaginations. And in this Olaf Stapledon was sixty years ahead. “Within the minded group, the insectoid units were forever dying off and giving place to fresh units,” Stapledon explained, “but the mind of the group was potentially immortal.”34 This is true of all large, enduring communications networks—whether the nodes are as inscrutable as neurons, as intelligent as human beings, or as dumb as microprocessors that sit on lampposts bouncing messages around and not doing much of anything else.

 

‹ Prev