Geek Sublime
Page 5
That second computer was the Apple II, the machine that defined personal computing, that is on every list of the greatest computers ever made. Woz designed all the hardware and all the circuit boards and all the software that went into the Apple II, while the other Steve spewed marketing talk at potential investors and customers on the phone. Every piece and bit and byte of that computer was done by Woz, and not one bug has ever been found, “not one bug in the hardware, not one bug in the software.”15 The circuit design of the Apple II is widely considered to be astonishingly beautiful, as close to perfection as one can get in engineering.
Woz did both hardware and software. Woz created a programming language in machine code. Woz is hardcore.
Working in machine code is very hard, so assembly code was created by adding some mnemonics to machine code. Working in assembly code is still hard; doing anything complex—like making games—in it is insanely hard. On programmers.stackexchange.com, a user going by the nom de guerre “DFectuoso” asked, “Are there any famous one-man-army programmers?”16 During the ensuing discussion about lone coders, one of the participants mentioned Chris Sawyer, who wrote the hugely successful 1999 game, RollerCoaster Tycoon: “He had a little help with music and graphics, but otherwise RollerCoaster Tycoon was all him. Amazing, especially given the physics engine. Last but not least, the entire game was written in assembly language.” And another commenter responded, “He wrote that in assembly?! Jesus Christ. I think I need to go boil my brain now.”
Sawyer’s achievements are indeed brain-boilingly immense, but why and how does a lone Scottish geek toiling obsessively over his virtual roller coasters become a “one-man-army”? Why do the Einsteins of programming affect—in their online personas and sometimes in person—a blustery True Grit swagger? Therein lies one of the tales that Nathan Ensmenger tells in his illuminating social history of computing, The Computer Boys Take Over: Computers, Programmers, and the Politics of Technical Expertise. Ensmenger does for computing in North America what historians of science have done for other disciplines: locate the development of knowledge and technology firmly within a messy matrix of human agency and politics; there is no orderly teleological progress from triumph to triumph, only competing interests that struggle over authority, access, and power. The computer boys of Ensmenger’s title are the early software-builders, the pioneers; the irony, as Ensmenger shows, is that many of them were women. In fact, the earliest programmers were all women: the “ENIAC girls” were women recruited by the (male) engineers and managers of the Electronic Numerical Integrator and Computer. The creators of the ENIAC had a clear division of labor in mind: the male scientist or “planner” would do the hard intellectual work of creating the mathematical algorithms and structures necessary to solve a problem; the female “coder” would then carry out the “static” manual labor of entering this plan into the machine by manipulating plugboard wiring and thousands of switches. Numerous wires had to be plugged and switches flipped for each machine instruction; each problem required thousands of instructions, so wiring—which was the programming—would take several days, and checking the wiring would take another few.17
This division of tasks of course echoed the hierarchies already present; men did the thinking and inventing, women were clerks. “The telephone switchboard-like appearance of the ENIAC programming cable-and-plug panels,” Ensmenger writes, “reinforced the notion that programmers were mere machine operators, that programming was more handicraft than science, more feminine than masculine, more mechanical than intellectual.”18 The planners considered the coding process so transparently simple that they couldn’t imagine that once in the machines, their algorithms might fault and hang, might need to be stopped. One of the ENIAC programmers, Betty Holberton, had to work very hard to convince John von Neumann that programs were complex and therefore fragile:
But to my astonishment, [Dr von Neumann] never mentioned a stop instruction. So I did coyly say, “Don’t we need a stop instruction in this machine?” He said, “No we don’t need a stop instruction. We have all these empty sockets here that just let it go to bed.” And I went back home and I was really alarmed. After all, we had debugged the machine day and night for months just trying to get jobs on it.
So the next week when I came up with some alterations in the code, I approached him again with the same question. He gave me the same answer. Well I really got red in the face. I was so excited and I really wanted to tell him off. And I said, “But Dr. von Neumann, we are programmers and we sometimes make mistakes.” He nodded his head and the stop order went in.19
Once von Neumann and everyone else involved with computers understood this hitherto unimaginable fact—that programming, translating algorithms into the language of machines, was very difficult—programmers became valuable commodities. A 1959 Price Waterhouse report warned that “high quality individuals are the key to top grade programming. Why? Purely and simply because much of the work involved is exacting and difficult enough to require real intellectual ability and above average personal characteristics.”20 Such individuals weren’t easy to find, and as corporations looked for competitive advantage by computerizing their business processes, a shortage resulted. Corporations set up training programmes, fly-by-night vocational schools sprang up guaranteeing jobs: “There’s room for everyone. The industry needs people. You’ve got what it takes.”21
In 1967, Cosmopolitan magazine carried an article titled “The Computer Girls” that emphasized that programming was a field in which there was “no sex discrimination in hiring”—“every company that makes or uses computers hires women to program them … If a girl is qualified, she’s got the job.” Admiral Grace Hopper, programming pioneer, assured the Cosmo readers that programming was “just like planning a dinner … You have to plan ahead and schedule everything so it’s ready when you need it. Programming requires patience and the ability to handle detail. Women are ‘naturals’ at computer programming.”22
Already, though, the “masculinization process” of the computing industry was underway. The severe limitations of memory and processing power in the machines of the day demanded Mel the Real Programmer’s wizardry; John Backus described programming in the fifties as “a black art, a private arcane matter … [in which] the success of a program depended primarily on the programmer’s private techniques and inventions.”23 The aptitude tests used by the industry to identify potential Mels consisted primarily of mathematical and logical puzzles, which often required a formal education in these disciplines; even Cosmopolitan magazine, despite its breezy confidence about the absence of sexism in computing, recognized this as a barrier to entry for its readers. An industry analyst argued that the “Darwinian selection” of personnel profiling resulted in an influx of programmers who were “often egocentric, slightly neurotic, and [bordering] upon a limited schizophrenia. The incidence of beards, sandals, and other symptoms of rugged individualism or nonconformity are notably greater among this demographic group.”24 The Real Programmers that the industry found through these aptitude tests were weird male geeks wielding keyboards.
As Ensmenger puts it:
It is almost certainly the case that these [personality] profiles represented, at best, deeply flawed scientific methodology. But they almost equally certainly created a gender-biased feedback cycle that ultimately selected for programmers with stereotypically masculine characteristics. The primary selection mechanism used by the industry selected for antisocial, mathematically inclined males, and therefore antisocial, mathematically inclined males were over-represented in the programmer population; this in turn reinforced the popular perception that programmers ought to be antisocial and mathematically inclined (and therefore male), and so on ad infinitum.25
The surly male genius, though, was perceived as a threat by corporate managers, especially as the initial enthusiasm over computerization gave way to doubts about actual value being produced for the companies spending the money. A programmer-turned-management-consultant de
scribed the long-haired computer wonks as a “Cosa Nostra” holding management to ransom, and warned that computer geeks were “at once the most unmanageable and the most poorly managed specialism in our society. Actors and artists pale by comparison.”26 Managers should “refuse to embark on grandiose or unworthy schemes, and refuse to let their recalcitrant charges waste skill, money and time on the fashionable idiocies of our [computer] racket.”27
The solution that everyone agreed upon was professionalization. Managers liked the idea of standardization, testing, and certification; it would reduce their dependence on arty individuals practising arcana: “The concept of professionalism affords a business-like answer to the existing and future computer skills market.”28 Programmers wanted to be recognized as something other than mere technicians—the rewards would be “status, greater autonomy, improved opportunities for advancement, and better pay.”29 Within academia, researchers were struggling to establish computer science as a distinct discipline and establish a theoretical foundation for their work. So, “an activity originally intended to be performed by low-status, clerical—and more often than not, female—[workers],” Ensmenger tells us:
was gradually and deliberately transformed into a high-status, scientific, and masculine discipline.
As Margaret Rossiter and others have suggested, professionalization nearly always requires the exclusion of women …
As computer programmers constructed a professional identity for themselves during the crucial decades of the 1950s and 1960s … they also constructed a gender identity. Masculinity was just one of many resources that they drew on to distance their profession from its low-status origins in clerical data processing. The question of “who made for a good programmer” increasingly involved in its answer the qualifier “male.”30
In 1892, a British colonial official named Sir Lepel Griffin wrote:
The characteristics of women which disqualify them for public life and its responsibilities are inherent in their sex and are worthy of honour, for to be womanly is the highest praise for a woman, as to be masculine is her worst reproach. But when men, as the Bengalis, are disqualified for political enfranchisement by the possession of essentially feminine characteristics, they must expect to be held in such contempt by stronger and braver races, who have fought for such liberties as they have won or retained.31
Sir Griffin was certain that Bengalis were unfit for political power because they were effeminate, weak, and it was unthinkable that they might “represent … precede … and govern the martial races of India”—that is, certain other ethnic groups within India who were deemed sufficiently warlike by the English. If a Bengali were to demonstrate such aspirations to equality, “then the English, as the common conqueror and master of all, may justly laugh at his pretensions and order him to take the humbler place which better suits a servile race which has never struck a blow against an enemy.”32
The gender politics of the Raj were of course much on my mind while I wrote my first novel in the late eighties and early nineties, as I traveled back and forth between India and America. The British “Cult of Manliness” had been an essential component of the creed of Empire, which—as above—conflated masculinity, violence, civic virtue, and morality. Even intelligence and intellectual capability were inextricably intertwined with masculinity; women and all others who exhibited symptoms of femininity were fuzzy headed, illogical, and easily overcome by emotion; they were incapable especially of scientific reasoning and therefore self-knowledge and progress. The state of the world—women without power, Englishmen ruling Indians—bore out the truth of these propositions.
By now, I’d read my Edward Said, and I prided myself on being aware of the ideological mechanisms that transformed local contingencies of history and culture into Nature itself. The attractions of Nick Carter, Killmaster, seemed altogether more sinister now that I had listened to many scholarly deconstructions of imperial American masculinity. But at the time, I didn’t question much the demographics of programming. The meetings of the special interest groups of HAL-PC devoted to programming were all-male; I think in all my years of consulting work I met one female programmer. This was just the way things were. The male programmers I met were often astonishingly generous with knowledge and technical advice, and yet, the very same men were also abrupt and outright rude. Indians are frequently taken aback by the American virtues of quick intimacy and bluntness, which come across as shockingly bad manners; I knew to discount for this, and understood that our own predilection for face-saving, izzat-preserving niceties made us maddeningly opaque and slippery to the average American. Still, these coders were deliberately obnoxious by anyone’s standards, especially online. They ad-hominemed, flamed, name-called, dismissed, despised. Not to put too fine a point on it: these guys were assholes. Preeminence among programmers was often decided by competitions of assholery, a kind of ritual jousting.
This unfortunate condition has only intensified over the decades. The “masculinization process” that Ensmenger describes has resulted in a contemporary American culture of programming that is overwhelmingly male, as one can see at conferences, on websites and blogs. The metaphors used within this world of one-man armies are very often martial. Teams working against impossible deadlines go on “death marches.” Finding and fixing defects in software is a pains-taking, detail-oriented task, one which Grace Hopper might have compared to housekeeping; but in the parlance of many programming shops, the most proficient bug sweepers are “bug slayers.”
In March 2011, David Barrett, CEO of Expensify (“Expense Reports That Don’t Suck”), blogged about how his start-up wouldn’t hire programmers who used Microsoft’s very large and elaborate .NET framework, which—according to him—provided ready-made, assembly-line tools that turned these programmers into drudges capable of only mass-producing pre-designed code, the programming equivalent of fast-food burgers. No, he wanted passionate programmers who could write “everything from assembly to jQuery, on PCs to mobile phones, [and code] hard core computer graphics to high level social networking.”33 Barrett wanted Einsteins, not Morts—fair enough. But this is how he described his Einsteins:
As you might know, we’re hiring the best programmers in the world. Sure, everyone says that. But my coders will beat up your coders, any day of the week. For example, Mich is barely 5 foot tall, but is a competitive fencer. Witold is a 6’3” former professional hockey player. Nate practices knife fighting for fun.34
Over a few days, I read hundreds of comments and blog posts debating the merits of Barrett’s case against .NET programmers; some argued that many great programmers used .NET, and that other frameworks had as many bad or lazy programmers. The discussions were long and nuanced. But nobody seemed to notice his very literal conflation of omnivorous intellectual curiosity with manly combat skills. He extends his fast-food riff—“Programming with .NET is like cooking in a McDonalds kitchen. It is full of amazing tools that automate absolutely everything”—but then turns the metaphor into a paean to programmer-as-blood-soaked-pioneer:
The sort of person [we are looking for] grew up cooking squirrels over a campfire with sharpened sticks—squirrels they caught and skinned while scavenging in the deep forests for survival. We don’t want a short order chef, we want a Lord of the Flies, carried by wolves into civilization and raised in a French kitchen full of copper-bottomed pots and fresh-picked herbs.35
“A Lord of the Flies in a French kitchen” neatly catches the geek machismo and extraordinary privilege that are essential ingredients in the cultural paradox that is Silicon Valley. Wages are so high here, Rebecca Solnit reports, that “you hear tech workers complaining about not having time to spend their money.”36 Depending on which San Francisco neighborhood you live in, your rent rose by anywhere from 10 percent to 135 percent over 2012, driven up by young techies outbidding each other.37 In the booming restaurants and cafés, there’s a general disdain for government, which is often described as fatally broken, in desperate need of “d
isruption,” that condition beloved of programmers and venture capitalists. Workers’ unions are regarded as anachronisms that hold back progress. Company founders chafe at any restrictions imposed by local or federal government as leftover mechanisms from a failed system which prevent the markets from working properly.38
Given these attitudes, it’s easy to conclude that Silicon Valley is a haven for Libertarians. Doing so would be simplistic. President Obama won his second presidential election by 49 percentage points in the Bay Area, as compared to his 22-point lead in California as a whole. Employees at Google gave 97 percent of their campaign contributions to Obama, and Apple employees gave 91 percent.39 But these denizens of the tech campuses aren’t, as we’ve seen, leftists or progressives of the Berkeley-Oakland ilk either. Rather, this new “virtual class” of digital overlords combine the social and sexual attitudes of San Francisco bohemianism with a neoliberal passion for idealized free markets and unchecked profit-making, thus producing a caste orthodoxy for people who might be best described as “hippie capitalists.”