Darwin Among the Machines

Home > Other > Darwin Among the Machines > Page 26
Darwin Among the Machines Page 26

by George B. Dyson


  This mesh was to be inoculated with a population of artificial agents corresponding to elements of the reality the model was intended to learn to represent: “The notion of direct representation of real social facts by means of activations in a computing machine can be made clearer if we view the computer employed in such an enterprise as a system of switches, comparable to a large telephone system. . . . Our program, then, begins with a design for an automaton that will be made to operate in a computer. Circuits will be so imposed on the computer that their performance will be analogous to the operations of human agents in real social groups. Next, we shall cut the computer free from its ordinarily employed ability to accomplish mathematical computations. . . . The patterns and entailments of the specific activations will be controlled so that they will be isomorphic with significant aspects of individual human actions. Thus, in a very precise sense, we shall be using a digital computer in an analogue mode. . . . The micro-processes in the computer resemble the micro-functioning of real agents under social constraints.”23 In their 1962 memorandum, the Romes were more specific about the workings of these “artificial agents,” even coining a new unit, the “taylor” (after F. W. Taylor, founder of time-and-motion studies), to measure the relative values of four different kinds of “social energy” with which individual agents were endowed.

  Then comes the crucial assumption—that without having to pay attention to the details, a behavioral gradient rewarding a closer match between model and reality will encourage the internal structure of the model to organize itself. “As we ‘tune’ our model, so to speak, that is, make the whole and the components resemble patterns of performance observable in real life, then the adequacy of the specific parametric values imposed on Leviathan for specific experiments should improve; and Leviathan will come to behave more like a real society both in its components and as a whole. In this way system properties can be imposed downward into the hierarchical structures.”24 As the Romes found out, this is easier said than done. They were on the right track, but with Leviathan running on an IBM 7090 computer, even with what was then an enormous amount of memory—over 1 million bits—the track wasn’t wide enough. Their model was defeated by the central paradox of artificial intelligence: systems simple enough to be understandable are not complicated enough to behave intelligently; and systems complicated enough to behave intelligently are not simple enough to understand.

  Citing von Neumann, the Romes concluded “that neither human thinking nor human social organization resembles a combinational logical or mathematical system.”25 Von Neumann believed the foundations of natural intelligence were distinct from formal logic, but through repeated association of the von Neumann architecture with attempts to formalize intelligence this distinction has been obscured. The Romes, following von Neumann’s lead, believed a more promising approach to be the cultivation of these behavioral processes within a random, disorganized computational system, but the available matrix delivered disappointing results. Nonetheless, they foresaw that given a more fertile computational substrate, humans would not only instruct the system but would begin following instructions that were reflected back. “Once we incorporate live agents into our dynamic computer model, the result can be a machine that will be capable of teaching humans how to function better in large man-machine system contexts in which information is being processed and decisions are being made.”26

  The Leviathan project grew into an extensive experiment in communication and information handling in composite human-computer organizations (or, in the Romes’ words, organisms), occupying a portion of the twenty-thousand-square-foot Systems Simulation Research Laboratory established at the System Development Corporation’s Colorado Avenue headquarters in 1961. Twenty-one experimental human subjects were installed in separate cubicles, equipped with keyboards and video displays connected to SDC’s Philco 2000 computer, the first fully transistorized system to be commercially announced. All activity was recorded electronically, and human behavior could be observed by researchers overlooking the complex through one-way glass. The Leviathan Technological System incorporated “16 live group heads (level III) reporting to four live branch leaders . . . (level IV) [who] in turn report to a single commanding officer (level V). . . . Underneath the live officers . . . 64 squads of robots are distributed . . . (level II) and report to the group heads directly over them. Each squad of robots consists of artificial enlisted men (level I) who exist only in the computer.”27 The human-machine system was embedded in an artificial environment and fed a flow of artificial work. Under various communication architectures, the Romes observed how organization, adaptive behavior, and knowledge evolved. The Romes concluded that “social hierarchies are no mere aggregations of component individuals but coherent, organic systems in which organic subsystems are integrated. . . . Social development is a telic advance, not just a concatenation of correlated events.”28

  With a tightening of air force purse strings, Leviathan was quietly abandoned, but in the SAGE system and its successors (which include most computer systems and data networks in existence today) these principles were given the room to freely grow. They developed not as an isolated experiment, but as an electronic representation coupled as closely to the workings of human culture, industry, and commerce as the SAGE system was coupled to the network of radar stations scanning the continental shelf. SAGE’s knowledge of scheduled airline flight paths, for example, gave rise to the SABRE airline-reservation system, refining the grain of the model down to an aircraft-by-aircraft accounting of how many seats are empty and how many are occupied in real time. Ashby’s law of requisite variety demands this level of detail in a system that can learn to control the routing, frequency, and occupancy of passenger aircraft, rather than simply identifying which flights are passing overhead. The tendency of representative models—whether a bank account that represents the exchange of goods, a nervous system that represents an organism’s relation to the world, or a reservation system that represents the number of passengers on a plane—is to translate increasingly detailed knowledge into decision making and control. In less than forty years, adding one subsystem at a time, we have constructed a widely distributed electronic model that is instructing much of the operation of human society, rather than the other way around. The smallest transactions count. Centrally planned economies, violating Ashby’s law of requisite variety, have failed not by way of ideology but by lack of attention to detail.

  Large information-processing systems have evolved through competition over the limited amount of reality—whether airline customers, market volume, or cold warfare—available to go around. This Darwinian struggle can be implemented within a single computer as well as at levels above; all available evidence indicates that nature’s computers are so designed. An influential step in this direction was an elder and much less ambitious cousin of Leviathan named Pandemonium, developed by Oliver Selfridge at the Lincoln Laboratory using an IBM 704. Instead of attempting to comprehend something as diffuse and complex as the SAGE air-defense system, Pandemonium was aimed at comprehending Morse code sent by human operators—a simple but nontrivial problem in pattern recognition that had confounded all machines to date.

  Selfridge’s program was designed to learn from its mistakes as it went along. Pandemonium—“the uproar of all demons”—sought to embody the Darwinian process whereby information is selectively evolved into perceptions, concepts, and ideas. The prototype operated on four distinct levels, a first approximation to the manifold levels by which a cognitive system makes sense of the data it receives. “At the bottom the data demons serve merely to store and pass on the data. At the next level the computational demons or sub-demons perform certain more or less complicated computations on the data and pass the results of these up to the next level, the cognitive demons who weigh the evidence, as it were. Each cognitive demon computes a shriek, and from all the shrieks the highest level demon of all, the decision demon, merely selects the loudest.”29

  “
The scheme sketched is really a natural selection on the processing demons,” Selfridge explained. “If they serve a useful function they survive, and perhaps are even the source for other subdemons who are themselves judged on their merits. It is perfectly reasonable to conceive of this taking place on a broad scale—and in fact it takes place almost inevitably. Therefore, instead of having but one Pandemonium we might have some crowd of them. . . . Eliminate the relatively poor and encourage the rest to generate new machines in their own images.”30 In the 1950s, when machine cycles were a rationed commodity and memory cost fifty cents per bit, Pandemonium was far too wasteful to compete with programs in which every instruction, and every memory address, counted. But when parts are cheap and plentiful (whether neurons or microprocessors or object-oriented programming modules), Pandemonium becomes a viable approach. In the 1950s, while computer engineers were preoccupied with architecture and hardware, Selfridge played the outfield in anticipation of semiautonomous processes spawned by machines within machines. Forty years later architecture and hardware are taken for granted and software engineers spend most of their time trying to cultivate incremental adaptations among the host of unseen demons within. “In fact, the ecology of programming is such that overall, programmers spend over 80 percent of their time modifying code, not writing it,” said Selfridge in 1995.31

  It makes no difference whether the demons are strings of bits, sequences of nucleotides, patterns of connection in a random genetic or electronic net, living organisms, cultural institutions, languages, or machines. It also makes no difference whether the decision demon is nature as a whole, an external predator, an internal parasite, a debugging program, or all of the above making a decisive racket at the same time. As Nils Barricelli discovered through his study of numerical symbioorganisms and John von Neumann discovered through his development of game theory, the tendency to form coalitions makes it impossible to keep the levels of decision making from shifting from one moment to the next.

  In this ambiguity lies the persistence of the nineteenth century’s argument from design—a proprietary dispute over whether the power of selection and the intelligence that it represents belongs to an all-knowing God or to impartial nature alone; whether it descends from above, arises from below, or is shared universally at every level of the scale. “Suppose there were a being who did not judge by mere external appearances,” wrote Charles Darwin to Asa Gray in 1857, “but who could study the whole internal organization, who was never capricious, and should go on selecting for one object during millions of generations; who will say what he might not effect?”32

  The nature of this selective being determines the scale of the effects—and organisms—that can be evolved. Charles Darwin, knowing it best to attempt one revolution at a time, replaced one higher intelligent being with another of a different kind. Intelligence, by any measure, is based on the ability to be selective—to recognize signal amidst noise, to discriminate right from wrong, to select the strategy leading to reward. The process is additive. Darwin was able to replace the supreme intelligence of an all-knowing God, who selected the whole of creation all at once, with the lesser intelligence of a universe that selected nature’s creatures step-by-step. But Darwin still dispensed this intelligence downward from the top.

  Darwin’s success at explaining the origin of species by natural selection may have obscured the workings of evolution in different forms. The Darwinian scenario of a population of individuals competing for finite resources, with natural selection guiding the improvement of a species one increment at a time, tempts us to conclude that where circumstances do not fit this scenario—despite the lengths to which Darwinism has been stretched—evolutionary processes are not involved. Large, self-organizing systems challenge these assumptions—perhaps even the assumption that a system must be self-reproducing, competing against similar systems and facing certain death or possible extinction, to be classified as evolving or alive. It is possible to construct self-preserving systems that grow, evolve, and learn but do not reproduce, compete, or face death in any given amount of time. It is also possible to view large, complex systems, such as species, gene pools, and ecosystems, as information-processing organizations providing a degree of guiding intelligence to component organisms whose evolution is otherwise characterized as blind. A blind watchmaker who can build an eye can evidently assemble structures that no longer stumble around.

  Samuel Butler argued against Darwin in 1887 that “we must also have mind and design. The attempt to eliminate intelligence from among the main agencies of the universe has broken down. . . . There is design, or cunning, but it is a cunning not despotically fashioning us from without as a potter fashions his clay, but inhering democratically within the body which is its highest outcome, as life inheres within an animal or plant.”33 Butler did not doubt the power of descent with modification, but he believed Darwin’s interpretation of the evidence to be upside down. “Bodily form may be almost regarded as idea and memory in a solidified state—as an accumulation of things each one of them so tenuous as to be practically without material substance. It is as a million pounds formed by accumulated millionths of farthings. . . . The theory that luck is the main means of organic modification is the most absolute denial of God which it is possible for the human mind to conceive—while the view that God is in all His creatures, He in them and they in Him, is only expressed in other words by declaring that the main means of organic modification is, not luck, but cunning.”34 Butler saw each species—indeed, the entire organic kingdom—as a store of knowledge and intelligence transcending the life of its individual members as surely as we transcend the life and intelligence of our component cells.

  The dispute between luck and cunning extends beyond evolutionary biology, promising a decisive influence over the future of technology as well. “The notion that no intelligence is involved in biological evolution may prove to be as far from reality as any interpretation could be,” argued Nils Barricelli in 1963. “When we submit a human or any other animal for that matter to an intelligence test, it would be rather unusual to claim that the subject is unintelligent on the grounds that no intelligence is required to do the job any single neuron or synapse in its brain is doing. We are all agreed upon the fact that no intelligence is required in order to die when an individual is unable to survive or in order not to reproduce when an individual is unfit to reproduce. But to hold this as an argument against the existence of an intelligence behind the achievements in biological evolution may prove to be one of the most spectacular examples of the kind of misunderstandings which may arise before two alien forms of intelligence become aware of one another.”35 Likewise, to conclude from the failure of individual machines to act intelligently that machines are not intelligent may represent a spectacular misunderstanding of the nature of intelligence among machines.

  The step-by-step expression of evolutionary intelligence, compared to the human attention span, is immeasurably slow. The evidence may become inescapable when speeded up. The invisible web of connections that bind an ecology—biological, computational, or both—into a living whole begins to move at a visible pace when the machines evolve from year to year, new generations of software are exchanged in minutes, and control is exercised from one microsecond to the next. Samuel Butler saw that the intelligence of evolution and the intelligence of human beings were heading on a collision course. Butler knew, as did Barricelli, that our definition of intelligence is so anthropocentric as to be next to useless for anything else. “Nothing, we say to ourselves, can have intelligence unless we understand all about it—as though intelligence in all except ourselves meant the power of being understood rather than of understanding,” he wrote. “We are intelligent, and no intelligence so different from our own as to baffle our powers of comprehension deserves to be called intelligence at all. The more a thing resembles ourselves, the more it thinks as we do—and thus by implication tells us that we are right, the more intelligent we think it; and the less it thinks as
we do, the greater fool it must be; if a substance does not succeed in making it clear that it understands our business, we conclude that it cannot have any business of its own.”36

  As Darwinism was perceived as iconoclasm by the church, so any attempt to attribute intelligence to evolutionary processes is viewed with suspicion by those of orthodox Darwinian faith. Butler and his followers have not stopped at seeking recognition for species-level intelligence, but keep attempting to invoke unorthodox levels of life and mind: “The only thing of which I am sure is, that the distinction between the organic and inorganic is arbitrary; that it is more coherent with our other ideas, and therefore more acceptable, to start with every molecule as a living thing, and then deduce death as the breaking up of an association or corporation, than to start with inanimate molecules and smuggle life into them.”37 Contraband passes across the boundaries between life and nonlife as freely as between intelligence and nonintelligence. The traffic goes both ways, attributing life and intelligence to processes below our own level as well as to those above. Everywhere we look things are turning out to be more intelligent and more alive than they once seemed. The more we come to understand the information-processing systems epitomized by the human brain, the more we find them to be functioning as evolutionary systems, and the more we come to understand evolutionary systems, the more we discover them to be operating as information-processing machines.

  Our faith in the intelligence of self-organizing, complex adaptive systems is reminiscent of William Paley’s faith in the argument from design. Where we see the emergence of order, Paley saw the hand of design. His protagonist, stumbling upon a watch, finds it utterly impossible “that there existed in things a principle of order, which had disposed the parts of the watch into their present form and situation. He never knew a watch made by the principle of order; nor can he even form to himself an idea of what is meant by a principle of order, distinct from the intelligence of the watchmaker.”38 Comparing watches, we find that Butler’s watch was designed from within; Paley’s watch was designed from without; Darwin’s watch was designed by the accumulation of sheer coincidence over time.

 

‹ Prev