It is of course undeniable that television is an example of Low Art, the sort of art that has to please people in order to get their money. Because of the economics of nationally broadcast, advertiser-subsidized entertainment, television’s one goal—never denied by anybody in or around TV since RCA first authorized field tests in 1936—is to ensure as much watching as possible. TV is the epitome of Low Art in its desire to appeal to and enjoy the attention of unprecedented numbers of people. But it is not Low because it is vulgar or prurient or dumb. Television is often all these things, but this is a logical function of its need to attract and please Audience. And I’m not saying that television is vulgar and dumb because the people who compose Audience are vulgar and dumb. Television is the way it is simply because people tend to be extremely similar in their vulgar and prurient and dumb interests and wildly different in their refined and aesthetic and noble interests. It’s all about syncretic diversity: neither medium nor Audience is faultable for quality.
Still, for the fact that individual American human beings are consuming vulgar, prurient, dumb stuff at the astounding average per-household dose of six hours a day—for this both TV and we need to answer. We are responsible basically because nobody is holding any weapons on us forcing us to spend amounts of time second only to sleep doing something that is, when you come right down to it, not good for us. Sorry to be a killjoy, but there it is: six hours a day is not good.
Television’s greatest minute-by-minute appeal is that it engages without demanding. One can rest while undergoing stimulation. Receive without giving. In this respect, television resembles certain other things one might call Special Treats (e.g. candy, liquor), i.e. treats that are basically fine and fun in small amounts but bad for us in large amounts and really bad for us if consumed in the massive regular amounts reserved for nutritive staples. One can only guess at what volume of gin or poundage of Toblerone six hours of Special Treat a day would convert to.
On the surface of the problem, television is responsible for our rate of its consumption only in that it’s become so terribly successful at its acknowledged job of ensuring prodigious amounts of watching. Its social accountability seems sort of like that of designers of military weapons: unculpable right up until they get a little too good at their job.
But the analogy between television and liquor is best, I think. Because (bear with me a second) I’m afraid good old average Joe Briefcase might be a teleholic. I.e., watching TV can become malignantly addictive. It may become malignantly addictive only once a certain threshold of quantity is habitually passed, but then the same is true of Wild Turkey. And by “malignant” and “addictive” I again do not mean evil or hypnotizing. An activity is addictive if one’s relationship to it lies on that downward-sloping continuum between liking it a little too much and really needing it. Many addictions, from exercise to letter-writing, are pretty benign. But something is malignantly addictive if (1) it causes real problems for the addict, and (2) it offers itself as a relief from the very problems it causes.10 A malignant addiction is also distinguished for spreading the problems of the addiction out and in in interference patterns, creating difficulties for relationships, communities, and the addict’s very sense of self and spirit. In the abstract, some of this hyperbole might strain the analogy for you, but concrete illustrations of malignantly addictive TV-watching cycles aren’t hard to come by. If it’s true that many Americans are lonely, and if it’s true that many lonely people are prodigious TV-watchers, and it’s true that lonely people find in television’s 2-D images relief from their stressful reluctance to be around real human beings, then it’s also obvious that the more time spent at home alone watching TV, the less time spent in the world of real human beings, and that the less time spent in the real human world, the harder it becomes not to feel inadequate to the tasks involved in being a part of the world, thus fundamentally apart from it, alienated from it, solipsistic, lonely. It’s also true that to the extent one begins to view pseudo-relationships with Bud Bundy or Jane Pauley as acceptable alternatives to relationships with real people, one will have commensurately less conscious incentive even to try to connect with real 3-D persons, connections that seem pretty important to basic mental health. For Joe Briefcase, as for many addicts, the Special Treat begins to substitute for something nourishing and needed, and the original genuine hunger—less satisfied than bludgeoned—subsides to a strange objectless unease.
TV-watching as a malignant cycle doesn’t even require special preconditions like writerly self-consciousness or neuroallergic loneliness. Let’s for a second imagine Joe Briefcase as now just an average U.S. male, relatively unlonely, adjusted, married, blessed with 2.3 apple-cheeked issue, utterly normal, home from hard work at 5:30, starting his average six-hour stint in front of the television. Since Joe B. is average, he’ll shrug at pollsters’ questions and answer averagely that he most often watches television to “unwind” from those elements of his day and life he finds unpleasant. It’s tempting to suppose that TV enables this unwinding simply because it offers an Auschlanderian “distraction,” something to divert the mind from quotidian troubles. But would mere distraction ensure continual massive watching? Television offers way more than distraction. In lots of ways, television purveys and enables dreams, and most of these dreams involve some sort of transcendence of average daily life. The modes of presentation that work best for TV—stuff like “action,” with shoot-outs and car wrecks, or the rapid-fire “collage” of commercials, news, and music videos, or the “hysteria” of prime-time soap and sitcom with broad gestures, high voices, too much laughter—are unsubtle in their whispers that, somewhere, life is quicker, denser, more interesting, more… well, lively than contemporary life as Joe Briefcase knows it. This might seem benign until we consider that what good old average Joe Briefcase does more than almost anything else in contemporary life is watch television, an activity which anyone with an average brain can see does not make for a very dense and lively life. Since television must seek to attract viewers by offering a dreamy promise of escape from daily life, and since stats confirm that so grossly much of ordinary U.S. life is watching TV, TV’s whispered promises must somehow undercut television-watching in theory (“Joe, Joe, there’s a world where life is lively, where nobody spends six hours a day unwinding before a piece of furniture”) while reinforcing television-watching in practice (“Joe, Joe, your best and only access to this world is TV”).
Well, average Joe Briefcase has an OK brain, and deep down inside he knows, as we do, that there’s some kind of psychic shell-game going on in this system of conflicting whispers. But if it’s so bald a delusion, why do he and we keep watching in such high doses? Part of the answer—a part which requires discretion lest it slip into anti-TV paranoia—is that the phenomenon of television somehow trains or conditions our viewership. Television has become able not only to ensure that we watch but somehow to inform our deepest responses to what’s watched. Take jaded TV-critics, or our acquaintances who sneer at the numbing sameness of all the television they sit still for. I always want to grab these unhappy guys by the lapels and shake them until their teeth rattle and point to the absence of guns to their heads and ask why the hell they keep watching, then. But the truth is that there’s some complex high-dose psychic transaction between TV and Audience whereby Audience gets trained to respond to and then like and then expect trite, hackneyed, numbing television shows, and to expect them to such an extent that when networks do occasionally abandon time-tested formulas Audience usually punishes them for it by not watching novel shows in sufficient numbers to let them get off the ground. Hence the networks’ bland response to its critics that in the majority of cases—and until the rise of hip metatelevision you could count the exceptions on one hand—“different” or “high-concept” programming simply doesn’t get ratings. High-quality television cannot stand up to the gaze of millions, somehow.
Now, it is true that certain PR techniques—e.g. shock, grotesquerie, or irrever
ence—can ease novel sorts of shows’ rise to national demographic viability. Examples here might be the “shocking” A Current Affair, the “grotesque” Real People, the “irreverent” Married… with Children. But these programs, like most of those touted by the industry as “fresh” or “outrageous,” turn out to be just tiny transparent variations on old formulas.
It’s not fair to blame television’s shortage of originality on any lack of creativity among network talent. The truth is that we seldom get a chance to know whether anybody behind any TV show is creative, or more accurately that they seldom get a chance to show us. Despite the unquestioned assumption on the part of pop-culture critics that television’s poor old Audience, deep down, “craves novelty,” all available evidence suggests, rather, that the Audience really craves sameness but thinks, deep down, that it ought to crave novelty. Hence the mixture of devotion and sneer on so many viewerly faces. Hence also the weird viewer complicity behind TV’s sham “breakthrough programs”: Joe Briefcase needs that PR-patina of “freshness” and “outrageousness” to quiet his conscience while he goes about getting from television what we’ve all been trained to want from it: some strangely American, profoundly shallow, and eternally temporary reassurance.
Particularly in the last decade, this tension in the Audience between what we do want and what we think we ought to want has been television’s breath and bread. TV’s self-mocking invitation to itself as indulgence, transgression, a glorious “giving in” (again, not exactly foreign to addictive cycles) is one of two ingenious ways it’s consolidated its six-hour hold on my generation’s cojones. The other is postmodern irony. The commercials for Alf’s Boston debut in a syndicated package feature the fat, cynical, gloriously decadent puppet (so much like Snoopy, like Garfield, like Bart, like Butt-Head) advising me to “Eat a whole lot of food and stare at the TV.” His pitch is an ironic permission-slip to do what I do best whenever I feel confused and guilty: assume, inside, a sort of fetal position, a pose of passive reception to comfort, escape, reassurance. The cycle is self-nourishing.
guilty fictions
Not, again, that the cycle’s root conflict is new. You can trace the opposition between what persons do and ought to desire at least as far back as Plato’s chariot or the Prodigal’s return. But the way entertainments appeal to and work within this conflict has been transformed in televisual culture. This culture-of-watching’s relation to the cycle of indulgence, guilt, and reassurance has important consequences for U.S. art, and though the parallels are easiest to see w/r/t Warhol’s Pop or Elvis’s Rock, the most interesting intercourse is between television and American literature.
One of the most recognizable things about this century’s postmodern fiction has always been the movement’s strategic deployment of pop-cultural references—brand names, celebrities, television programs—in even its loftiest High Art projects. Think of just about any example of avant-garde U.S. fiction in the last twenty-five years, from Slothrop’s passion for Slippery Elm throat lozenges and his weird encounter with Mickey Rooney in Gravity’s Rainbow, to “You” ’s fetish for the New York Post’s COMA BABY feature in Bright Lights, Big City, to Don DeLillo’s pop-hip characters saying stuff to each other like “Elvis fulfilled the terms of the contract. Excess, deterioration, self-destructiveness, grotesque behavior, a physical bloating and a series of insults to the brain, self-delivered.”11
The apotheosis of the pop in postwar art marked a whole new marriage between High and Low culture. For the artistic viability of postmodernism was a direct consequence, again, not of any new facts about art, but of facts about the new importance of mass commercial culture. Americans seemed no longer united so much by common beliefs as by common images: what binds us became what we stand witness to. Nobody sees this as a good change. In fact, pop-cultural references have become such potent metaphors in U.S. fiction not only because of how united Americans are in our exposure to mass images but also because of our guilty indulgent psychology with respect to that exposure. Put simply, the pop reference works so well in contemporary fiction because (1) we all recognize such a reference, and (2) we’re all a little uneasy about how we all recognize such a reference.
The status of Low-cultural images in postmodern and contemporary fiction is very different from those images’ place in postmodernism’s artistic ancestors, e.g. the “dirty realism” of a Joyce or the ur-Dadaism of something like Duchamp’s toilet sculpture. Duchamp’s aesthetic display of that vulgarest of appliances served an exclusively theoretical end: it was making statements like “The Museum is the Mausoleum is the Men’s Room,” etc. It was an example of what Octavio Paz calls “Meta-irony,”12 an attempt to reveal that categories we divide into superior/arty and inferior/vulgar are in fact so interdependent as to be coextensive. The use of Low references in a lot of today’s High literary fiction, on the other hand, serves a less abstract agenda. It is meant (1) to help create a mood of irony and irreverence, (2) to make us uneasy and so “comment” on the vapidity of U.S. culture, and (3) most important, these days, to be just plain realistic.
Pynchon and DeLillo were ahead of their time. Today, the belief that pop images are basically just mimetic devices is one of the attitudes that separates most U.S. fiction writers under c. 40 from the writerly generation that precedes us, reviews us, and designs our grad-school curricula. This generation gap in conceptions of realism is, again, TV-dependent. The U.S. generation born after 1950 is the first for whom television was something to be lived with instead of just looked at. Our elders tend to regard the set rather as the flapper did the automobile: a curiosity turned treat turned seduction. For younger writers, TV’s as much a part of reality as Toyotas and gridlock. We literally cannot imagine life without it. We’re not different from our fathers in that television presents and defines our contemporary world. Where we are different is that we have no memory of a world without such electric definition. This is why the derision so many older fictionists heap on a “Brat Pack” generation they see as insufficiently critical of mass culture is at once understandable and misguided. It’s true that there’s something sad about the fact that David Leavitt’s short stories’ sole description of some characters is that their T-shirts have certain brand names on them. But the fact is that, for most of Leavitt’s educated young readership, members of a generation raised and nourished on messages equating what one consumes with who one is, Leavitt’s descriptions really do do the job. In our post-1950s, inseparable-from-TV association pool, brand loyalty really is synecdochic of character; this is simply a fact.
For those U.S. writers whose ganglia were formed pre-TV, those who are big on neither Duchamp nor Paz and who lack the oracular foresight of a DeLillo, the mimetic deployment of pop-culture icons seems at best an annoying tic and at worst a dangerous vapidity that compromises fiction’s seriousness by dating it out of the Platonic Always where it ought to reside. In one of the graduate workshops I went through, a certain gray eminence kept trying to convince us that a literary story or novel should always eschew “any feature which serves to date it”13 because “serious fiction must be Timeless.” When we protested that, in his own well-known work, characters moved about electrically lit rooms, drove cars, spoke not Anglo-Saxon but postwar English, and inhabited a North America already separated from Africa by continental drift, he impatiently amended his proscription to those explicit references that would date a story in the “frivolous Now.” When pressed for just what stuff evoked this F.N., he said of course he meant the “trendy mass-popular-media” reference. And here, at just this point, transgenerational discourse broke down. We looked at him blankly. We scratched our little heads. We didn’t get it. This guy and his students simply did not conceive the “serious” world the same way. His automobiled Timeless and our MTV’d own were different.
If you read the big literary supplements, you’ve doubtless seen the intergenerational squabble this sort of scene typifies.14 The plain fact is that certain things having to do with fiction pro
duction are different for young U.S. writers now. And television is at the vortex of most of the flux. Because younger writers are not only Artists probing for the nobler interstices in what Stanley Cavell calls the reader’s “willingness to be pleased”; we are also, now, self-defined parts of the great U.S. Audience, and have our own aesthetic pleasure-centers; and television has formed and trained us. It won’t do, then, for the literary establishment simply to complain that, for instance, young-written characters don’t have very interesting dialogues with each other, that young writers’ ears seem “tinny.” Tinny they may be, but the truth is that, in younger Americans’ experience, people in the same room don’t do all that much direct conversing with each other. What most of the people I know do is they all sit and face the same direction and stare at the same thing and then structure commercial-length conversations around the sorts of questions that myopic car-crash witnesses might ask each other—“Did you just see what I just saw?” Plus, if we’re going to talk about the virtues of “realism,” the paucity of profound conversation in younger fiction seems accurately to reflect more than just our own generation—I mean six hours a day, in average households young and old, just how much conversation can really be going on? So now whose literary aesthetic seems “dated”?
The David Foster Wallace Reader Page 85