15/03/2019

Capitalist Buzzwords : Creativity and Innovation




“Creative” and its variants are a versatile part of the vocabulary of contemporary capitalism, able to link imagination, aesthetic practice, and religious faith in the pursuit of private gain. The oldest word on creativity’s family tree is “creation,” whose first meaning was specifically Christian—the term for the divine genesis of the universe. One of the newest is the nominal form of creative, normally an adjective for an original thinker or idea. Creative is now also a count noun (think of the creatives who may be moving into your shrinking rust-belt city) and a mass noun (get creative on the horn, said account services to production). Many people would probably agree that creativity is an essential human trait, crucial to a happy life, though this noun is a relatively new coinage. An even more recent development is the notion that creativity is a trait of capitalist markets. And in the United States, the political phrase “job creators” borrows some of creation’s residual divine light to illuminate the benevolent fiat of the capitalist, who is thought to create jobs out of the formless void.

If creation has been divine, creativity is decidedly human. The most important conflict in the etymological history of “creative” is the struggle between its religious and secular meanings. Before the late 19th century, to the degree that people could participate in something called “creation,” it was only to approximate the purity of the original, capital-C Creation, rather than inaugurating their own. “There is nothing new under the sun,” Ecclesiastes reminded humans inclined to creative hubris. “The created cannot create,” (creatura no potest creare) added St. Augustine, insisting that such power resided only in God, and not in his creations. In the history of the word “creative,” there are actually two decisive rifts: this initial one between the divine and the human creation suggested here, and then, once creativity became a human trait, a division between its aesthetic and productive forms. We can roughly date this latter conflict to around 1875, the earliest example of the word “creativity” given in the OED, from an essay on Shakespeare’s singular brilliance.

Creativity was a work of imagination, rather than production, of artistry rather than labor. One of the consequences of this split between art and work has been to valorize creativity as the domain of an intuitive, singular, historically male genius. Productive creativity, meanwhile, is not art but labor, and thus rarely earns the title of creativity at all; this is the supposedly unimaginative labor of the manual worker or the farmer and the often feminized work of social reproduction. There are obvious class and gender prejudices at work here; coaxing a crop out of stubborn soil or preparing a family meal with limited ingredients are not typically seen as creative acts, whereas cooking a restaurant meal with the harvested crops often is. Other differences are rather arbitrary. Children’s play is not thought to be brilliant in the way Shakespeare is; it may, however, qualify as creative because it is appears (to adults, anyway) to be intuitive.

The widespread popularity of “creative” in economic discourse today suggests that this old breach between the imaginative and the productive has partly been closed. The idiosyncratic, eccentric, even oppositional posture of the creative artist is now an economic asset, chased by real-estate developers and promoted by self-help writers as comparable to the spirit of the entrepreneur. The popularity of “creativity” as an economic value in English can be traced to two major sources—Joseph Schumpeter, 20th-century economist and theorist of “creative destruction,” and Richard Florida, the University of Toronto scholar whose book, The Rise of the Creative Class, became one of the most celebrated and influential urban policy texts of the early 2000s. “The bourgeoisie cannot exist without constantly revolutionizing the instruments of production, and thereby the relations of production, and with them the whole relations of society,” wrote Marx and Engels in the Manifesto of the Communist Party. In his 1942 classic Capitalism, Socialism, and Democracy, Schumpeter agreed, up to a point. Schumpeter shared Marx’s sense of capitalism as a destructive and also transformative historical process, but he reframed Marx’s history of class struggle motivated by exploitation as an evolutionary process driven by visionary entrepreneurs. By opening new markets and breaking down old industrial processes, Schumpeter wrote, capitalism “incessantly revolutionizes the economic structure from within, incessantly destroying the old one, incessantly creating a new one.” This process is what Schumpeter called “creative destruction.”

“Creative” here refers to the work of forging new modes of production, new markets, and new products, but it also had a touch of artistry: the ingenuity, vision, and intuition to make things anew. Even so, creativity still belonged more to the artists’ studio than the corner office until the last half of the 20th century in the United States, when it came to describe a productive aspect of the psychology of individual workers. Sarah Brouillette has shown how much the familiar accoutrements of today’s progressive office culture—foosball tables, bright colors, and other perks ostensibly meant to cultivate employees’ creativity and loyalty—owe to the psychologist Abraham Maslow, famous for his 1943 theory of a “hierarchy of human needs.” Maslow, Brouillette writes, “began imagining all business culture as an outlet for and source of workers’ enterprising individual self-fulfillment.” Florida’s notion of “artistic creativity,” which he regards as integrated with other varieties (economic and technological) is based on the presumption that “art” equals “self-expression,” a historically specific assumption which, like “creativity” itself, is mistaken for a universal and timeless idea.

Indeed, it is this particularly modernist understanding of artistic work as solitary and idiosyncratic, oriented towards the expression of the artist’s unique self, that also came to suit US geopolitical interests in the Cold War, as many scholars of US and Latin American modernism have shown. Rob Pope, in his history of creativity, includes two other examples of creativity’s Cold War deployments. The US psychologist Carl Rogers argued that a prosperous nation like the United States needed “freely creative original thinkers,” not ladder-climbing conformists. If the grey Soviet system encouraged the latter, the multi-colored capitalism of the United States required creative free thinkers. It is this perceived cultural strength, not just newer and better weapons, that will beat Communism, J.P. Guilford argued in 1959.

Florida believes strongly in the naturalness and timelessness of this relatively new idea, creativity; he describes it as “what sets us apart from all other species.” Florida elevates creative capitalists from a social type, which they remain in Schumpeter’s theory of the entrepreneur, to a social class, one that Florida estimates as constituting a third of the American working population. Its members include scientists and engineers, architects and artists, musicians and teachers—anyone, in short, “whose economic function is to create new ideas, new technology, and new creative content.” The creative class shares certain tastes and preferences, like nonconformity, an appreciation for merit, a desire for social diversity, and an appetite for “serendipity,” the chance encounter facilitated by urban life.

A taste for city life, in fact, is one of the creative class’s most treasured preferences, and Florida’s ideas promised to leverage these to repopulate declining urban centers without significant public expenditure on social welfare or infrastructure. Politicians in various postindustrial cities in the global north became eager customers of the consultancy spawned by the success of The Rise of The Creative Class. We will find in Florida’s account scarcely a trace of the breach that Williams described between imaginative and productive creativity. Creativity, Florida writes, is the “font from which new technologies, new industries, new wealth, and all other good economic things flow.” Art, music, and gay-friendliness are no longer independent values of their own, but rather values dependent on their appeal to high-wage knowledge workers.

Why creativity now? Jamie Peck, one of Florida’s most unsparing critics, argues that from the 1970s, deindustrializing cities were faced with a dearth of available economic development options. They began competing with one another not only for increasingly mobile jobs, but for places in a consumer economy in which cities became commodities themselves. Abandoning comprehensive urban planning, city governments focused instead on what Peck calls “urban fragments” with potential in this consumer market—single districts with marketable appeal due to their theaters or arenas, historic architecture, proximity to job-rich downtowns, or some other marketable feature. (Often, these urban fragments become the sort of generic, purpose-built developments the suburban-born creatives were supposedly fleeing, like the new creative-class Potemkin village north of downtown Detroit with the odd name The District Detroit, an ode to urbanity that only a marketer could love, or even understand.) The creative city becomes a place where a mobile middle class can participate in this consumer economy as workers and residents. Since there is a broad affinity between one’s economic and imaginative activity in the Floridian regime of creativity, you “live, work, and play,” in the familiar liturgy of bourgeois urban life, as an economic subject at all hours. Even artistic activity that might have once appeared oppositional or radical—Florida claims to be a big fan of rap music—merely buttresses the “creative index” of the city (a metric he developed).

“Whereas classical liberal doxa assumes that what we are and what we own must not be confused,” writes Brouillette, neoliberals favor a “union of economic rationality and authenticity, this perfect marriage between the bohemian and what had been her bourgeois other.” The extra-economic values of the artist and the priorities of the market are no longer treated as autonomous, much less antagonistic, but harmonious. As Brouillette emphasizes, however, it would be a mistake to read the rise of the creative class as a colonization of the once-pure realms of the artistic imagination by the market. The rise of the so-called creative class is not a heroes-and-villains plot of businessmen corrupting creativity. This would be far too flattering to artists and writers, who are hardly innocent bystanders. Rather, the business world has valorized unexamined ideas of what “artistry” means and turned an individualistic, class-bound idea of “the artist” towards market goals. These meanings of artistry have evolved over the years in complex ways, but the one that circulates in the economic use of creativity dates to the origins of the word “creativity” in the late 19th century. Then, as Gustavus Stadler has shown, creativity was closely aligned with 19th-century ideas of “genius” and “inspiration,” which were seen as the fruit of an “irreducible originality,” rather than a social process. As the cult of the entrepreneur shows, this fantasy of irreducible originality is still with us. Sometimes, there really is nothing new under the sun.

From : Keywords: The New Language of Capitalism.  Published by Haymarket Books, 2019.

How ‘Creativity’ became a capitalist buzzword. By John Patrick Leary. LitHub ,  March 11 , 2019.











In 1959, a group of high school juniors in a Brooklyn classroom took a very bad quiz. The questions, written by a right-wing, pro-business lobbying group, were a mix of multiple-choice and true-false propositions on economic matters, like When a company makes big profits in any one year, it ought to raise wages. (Correct answer: false.) Or: Which of the following do you think should be government owned and operated? a) railroads b) automobile companies c) banks d) steel companies e) oil companies f) power companies or g) none of the above. (Correct answer: g.) Most students failed the quiz; the average score was 45 percent, but they didn’t appear to mind much. One student told the authors of the quiz that they represented “a stinking, reactionary organization.”

And so they did. The quiz was written by the National Association of Manufacturers (NAM), an organization founded in 1895 to fight organized labor and retooled in the 1930s as a leading opponent of Franklin Delano Roosevelt’s New Deal. After World War II, NAM’s eager propagandists began developing free-market economics curricula for U.S. high schools, which they were convinced were hotbeds of pro-labor and communist sentiment. Written student feedback recorded in NAM’s files seemed to confirm those fears. A Brooklyn student wrote on the quiz:


  WORKERS OF THE WORLD UNITE! Down with the capitalist exploiters. These ‘people’ do no work and expect to receive recompense for laziness. To paraphrase Lincoln, who in his second inaugural commented in much this way on slavery—‘the taskmaster wringing his bread from the sweat of another’s brow and the unrequited toil of millions of laborers.’ This system is contradictory to our ideals. The system must be abolished and the worker emancipated.

The results showed the NAM researchers the urgency of their mission. The honor students who flunked their economics quiz were some of the city’s most elite students, and yet, the NAM report concluded, they showed an exceptionally “high degree of economic illiteracy.” Teaching the youth of America to love the “free enterprise system” again would be a generational task.

Looking back from the vantage point of neoliberal America circa 2019, it is striking (and a little cheering) to see the anguish of the defenders of private property at mid-century. In her 2009 book Invisible Hands, Kim Phillips-Fein showed how deeply the fall of the “employer’s paradise” of the 1920s shook the ruling class from the 1930s onward, as workers struck and organized and the public mood turned against industry. By the 1950s, right-wing writers complained that socialists still set the terms of debate and commanded the language of the economy: “55% hold to the Marxist theory ‘from each according to his ability, to each according to his needs,’” according to the NAM’s report on American teenagers during the Eisenhower era. Their schoolbooks uniformly “make the employer look like a piker,” lamented the head of the group’s education department. “We have not learned to speak our own economic language as well as the other fellow has learned his specious patter,” wrote Wilbur Brons in the Chicago Journal of Commerce in 1944. Things were so bad in 1950 that Pierre S. du Pont, scion of the Delaware chemical dynasty and one of America’s richest men, felt himself to be living a kind of tragedy. “Perhaps,” he wrote wistfully to a colleague, “we were born too soon.”


Exiting the social-democratic nightmare meant, in part, learning, and teaching, a new language. The socialistic shibboleths of “security,” “planning,” “economic democracy,” and “full employment” would have to be confronted and countered with new ideals of competition and entrepreneurship. People like the Brooklyn high school juniors had learned, through years of Depression and war, to associate freedom with security from the thing called “the market,” in its various manifestations: a cruel boss, a closed factory gate, a sped-up assembly line. To redeem the free enterprise system, the apostles of private property needed a vocabulary for emancipation through the market. To do this, they needed to make the market the sort of misty abstraction you could never confuse with the “sweat of another’s brow.” If only Pierre du Pont could have known how popular “innovation” would one day become.

One of innovation’s earliest and most cogent modern definitions, in the economic usage that we’ve wearily come to know so well, comes from Joseph Schumpeter. The Harvard economist saw innovation in 1942 as “the entrepreneurial function.” To innovate meant:

     to reform or revolutionize the pattern of production by exploiting an invention or, more generally, an untried technological possibility for producing a new commodity or producing an old one in a new way, by opening up a new source of supply of materials or a new outlet for products, by reorganizing an industry and so on.

He wrote elsewhere that “innovations are always associated with the rise to leadership of New Men,” which pinpoints one of the major paradoxes of the term: its simultaneously functional and utopian usages. That is, the process of innovation—Schumpeter famously called it “creative destruction”—is both a workaday managerial process and also a kind of heroic eruption of market-shaping genius. The ubiquity of the term today may make us think of high-tech abundance, or perhaps the “better living through chemistry” that twentieth-century American firms like DuPont long promised consumers of plastics and nylons. But now, of course, the idea of innovation—less moored to material things than chemistry was—reigns supreme in Silicon Valley, the world’s “most innovative neighborhood,” as a typical description in the business press goes. “Chemistry” implied, at some point in the process, the production of a material object; tech innovation celebrates the decisions of the manager in his office.

Still, where, exactly, did this brand of innovation, in all its world-conquering glory, come from? It is impossible to date the origins of a word, especially one as vague as this one, with any certainty, but “innovation” as we use it now—a spirit of market-serving managerial creativity—begins roughly in 1934. This was the year that saw hundreds of thousands of workers across the country in a variety of industries—from longshoremen in San Francisco and Seattle to teamsters in Minneapolis to textile workers in Alabama—walk off their jobs in a wave of major strikes. And in the following year, the Social Security Act and the Wagner Act were passed, landmark pieces of the New Deal. Reacting with alarm to these events, the nation’s evangelists for private property cast about desperately for a response.

The New Deal’s domestic opponents, scrambling for an edge against Roosevelt, fixed on “bureaucracy.” In the platform of the American Liberty League (a short-lived congress of industrialists the president eagerly denounced as “economic royalists”) the first point was “to combat the growth of bureaucracy.” By “bureaucracy,” they meant the Works Progress Administration, unemployment insurance, labor unions, Social Security, and the kind of public ownership of major industries the Brooklyn students were still prepared to endorse in 1959. The problem, of course, was the widespread popularity of all the above. The defenders of free enterprise were still using someone else’s terms: “freedom from bureaucracy,” also a NAM slogan, simply riffed on Roosevelt’s “four freedoms.” What the American right needed, in other words, was its own positive language of “freedom” in the market.





Innovation became the word for the job. Its very vagueness served as an irresistible selling point. Like Supreme Court Justice Potter Stewart’s definition of obscenity, innovation was something that its ardent modern prophets professed to know mostly when they saw it. And increasingly they came to see it everywhere—particularly in the internet age’s long tail. By now, virtually any new product is described as an innovation, but innovation in its most dominant form today is a kind of spirit, a way of being, an attitude. This helps explain its wide distribution across profit-making and nonprofit fields. The Harvard Business Review, the organ of corporate conventional wisdom, defines innovation in one article as “experimentation, risk-taking, and variety,” which are “the enemy of the efficiency machine that is the modern corporation.” Putting a man on the moon, putting the world on wheels, putting an encyclopedia in your pocket: these breakthroughs are all fruits of this spirit of risk-taking. But even the most banal, disposable products of our lives are awash in it; if, for any reason, you find your way to www.band-aid.com, you will find there an article called “A History of Band-Aid Brand Innovation.” (Sample achievement, 2001: “Band-aid.com is launched.”)

Most institutions ritually invoke it for self-justification: Aramark Correctional Services, the private prison food contractor implicated in Michigan for serving literal garbage to inmates, sponsors a “Jail Innovations of the Year Award” to recognize jailers for “innovative contributions made to their jail or correctional facility.” Politicians of all parties invoke it reverently. There are, by my rough count, at least a half-dozen Innovation Churches currently operating across America, and elsewhere, a “theology of innovation” seeks to update that decidedly old-economy metaphor of the preacher as shepherd in favor of a new, knowledge-economy pastorate. The evangelical megachurch minister Rick Warren, for example, has written that “a theology of innovation always reminds us that God intends us to be creative.” Warren is speaking here of a church’s organizational culture, and he borrows from the hipster capitalist CEOs who advocate replacing boardrooms and cubicles with ping-pong tables and reclining lounge chairs. Submission to God, in Warren’s church, no longer requires Job’s prostration in the face of the Lord’s unfathomable desires. Instead, Warren writes, “it’s when I get in a totally prone position”—he means sitting in a recliner—“that I can be the most creative and can discover what God would have us do.”

Our children are also taught to be innovators, from inner-city Atlanta to affluent San Diego, and from pre-kindergarten through college. A school in Las Vegas promises to nurture the “entrepreneurs and creators inherent in us all,” starting with twelve-month-olds in day care; entrepreneurship summer camps teach middle-schoolers to write a business plan and pitch investors; colleges and universities increasingly market themselves as laboratories of innovation and entrepreneurship.

After factories, prisons, schools, and churches, there isn’t much left over in twenty-first century American life. Across these various institutions, innovation is so widespread and its goodness so seemingly self-evident that questioning it might seem bizarre and truculent, like criticizing beauty, science, or penicillin—things we think of either as universal human virtues or socially useful things we can scarcely imagine doing without. Where did this ubiquitous concept come from—so imprecise, so vapid, in neither its verb or noun form describing any coherent action or object with clarity or consistency? And what does its popularity say about us?

Here the postwar usage of the term is especially telling. Along with “entrepreneurship,” a related term, innovation became a response to the malaise of bureaucracy typically shorthanded by the title of William Whyte’s 1956 best-seller, The Organization Man. Whyte argued that Americans in the fifties—by which he meant middle-class, professional white men—were “imprisoned in brotherhood.” A stifling spirit of consensus was sapping the country’s vital and independent energies. Middle-class, flannel-suited office workers lost in bureaucracies and marooned in suburbs threatened the vitality of an economy compromised by its own stability. The old economic virtues had lost their power. What good was a Horatio Alger story about the importance of thrift in an economy built on endless consumption? What did “pluck” matter to a middle manager?

The college-educated youth of the late 1950s were interested in job security, not trailblazing, Whyte argued. He quotes an economics professor who says his students “do not question the system. . . . They will be technicians of the society, not innovators.” In an example of the evergreen nature of blaming social problems on Kids Today, a youth marketer lamented in 1960 that young men were not the “pioneer stock” of their grandparents. “I think that a lot of large corporation heads know it and are alarmed about it,” he said. “They get the organization man rather than the entrepreneur.” One of the first articles centered on the topic of innovation in the Harvard Business Review, from 1962, echoed this suspicion of the “technician” as a compliant facilitator, rather than a dynamic creator. “In our striving to create and maintain order and stability in large enterprises,” wrote John J. Corson in “Innovation Challenges Conformity,” “it has become increasingly clear that we discourage and limit the initiative and creativeness required for innovation.”

Free enterprise, then the most popular euphemism for capitalism, described a system and a structure; innovation was becoming a way to describe a free individual’s way of thriving inside that structure. Advocates of the new discipline of management began to emphasize what the MIT business professor Douglas McGregor called in 1960 “the human side of enterprise”—moral and personal traits like empathy and innovation, instead of classic administrative traits like discipline and authority. In his 1954 book The Practice of Management, Peter Drucker, an Austrian exile often called the “father of management theory,” described the calling of managers as nothing less than to integrate operations “so as to utilize the special properties of the human being.”



Drucker, beginning with his 1939 book The End of Economic Man, was preoccupied by the organization as a cultural and political problem—he argued that a fetish for organization as such defined the appeal of Nazism. But in his 1985 book, Innovation and Entrepreneurship, Drucker sounded an optimistic note. “Where are all the young people who, we were told fifteen years ago, were turning their backs on material values, on money, goods, and worldly success, and were going to restore to America a ‘laid-back,’ if not a pastoral ‘greenness’?” For the management theorists who came to prominence in the 1960s, the business world was not the place of one-dimensional men, from which laid-back eccentrics fled. Quite the contrary: “innovation” was the recovery of the more fully human, the irrational, the creative, the idiosyncratic in business and in all aspects of culture influenced by it (which was, Drucker argued, every sphere of human culture). And so innovation could offer a sense of purpose to workers swallowed by bureaucratic consensus. As Nils Gilman has argued about Drucker, practical business advice—the workaday techniques of innovation and entrepreneurship—became the solutions to the crises of capitalism that erupted violently at the beginning of Drucker’s career. The innovator, as Schumpeter had said, was both bureaucrat and hero.

This sort of idealism still resounds in the vocabulary of the business world today: it is individualized, moral qualities of innovation and “passion” that ostensibly drive success, and whimsical fetishes like “leadership” and “design thinking” package a species of individual deliverance in a cubicle. The freedom from bureaucracy and order-following that innovation came to promise belongs to what Drucker called in 1968, with now-quaint optimism, “the knowledge economy.” This was an economy driven by “ideas, concepts and information rather than manual skill or brawn.” No longer a rule-following drone in the industrial society, the worker in the knowledge economy was driven by her independent choices. We have moved on from an economy of “predetermined occupations into one of choices for the individual,” Drucker claimed. The knowledge-economy worker can make a living doing “almost anything one wants to do and plying almost any knowledge.” “This,” Drucker insisted, defying the biblical warning, “is something new under the sun.”

For the historian of such an idea—the supposedly ageless idea of newness itself, of creativity, of human initiative—the warning in Ecclesiastes against hubris makes a great deal of sense. What is very old in the literature of business, self-help, and corporate public relations is constantly made new again. Theirs is a history of expired prophecies and long-defunct game-changers: from “quality circles” and mind cures to synergy and emotional intelligence, the fantasy of the transformational new idea—the better intellectual mousetrap—has long seduced writers chasing riches and prestige by telling others how to chase riches and prestige. For example, the number of articles, books, and lectures asserting that no, “innovation” is not the same as “invention,” might suggest that the issue has been definitively settled. But business literature is also a history of endless repetition, of already thin gruel reheated and sold—and sold, and sold, in a vast publishing market—as fresh nourishment. Innovation, in fact, is one such example. It’s one of the oldest new ideas in the corporate lexicon, a novelty that never ages. Forty years after Schumpeter praised the “creative destruction” of the innovative entrepreneur, the management consultant and best-selling author Tom Peters paid tribute to these hero-functionaries in his book In Search of Excellence—the most widely held monograph among libraries in the United States from 1989 to at least 1997, according to the library database WorldCat. “Small, competitive bands of pragmatic bureaucracy-beaters,” he called the apostles of excellence, “the source of much innovation.”

Because it is an ideal of managerial decision-making, innovation is critical to a vernacular of neoliberalism that renders invisible most forms of labor performed by most people on Earth—hot, fast, exhausting, repetitive, alienating, caretaking, unwaged. Its emergence in the 1950s and 1960s was conditioned by mass culture and the legacy of the New Deal. Influential theories of entrepreneurship and innovation in these decades were also invested in questions about the fate of the non-white world, which was poised as never before to disrupt the world system.

Drucker said that what distinguished an “‘underdeveloped country’—and keeps it underdeveloped—is not so much a shortage of capital as it is shortage of innovation.” Walt Rostow, the best-selling political theorist whose 1952 book, The Process of Economic Growth, was an argument for the steps an underdeveloped society needs to “take off” into development, listed a willingness “to accept innovations” as a major criteria. This helps explain how “innovativeness” can become a measure, not just of an industry’s profitability, but of social worth and national character. When The Economist last October defended the Honduran migrants who were confronting Trump at the southern U.S. border, they did not invoke civil rights, the dignity of the individual, or pan-Americanism. They spoke, instead, of the danger to “American dynamism and innovation.” Central American migrants include valuable fruit-pickers and home health aides but they can also be “entrepreneurs and coders,” useful resources for American industry. If innovation is, as Jill Lepore has argued, the nineteenth-century ideal of progress “scrubbed clean” of its horrors and “relieved of its critics,” this is once again proof that Ecclesiastes was probably right: there is nothing new under the sun.

But in the years since Rostow, Drucker, and even Tom Peters theorized the mystic inner workings of innovation, the term has taken on a more abstract, even metaphysical cast. The Harvard Business School luminary Rosabeth Moss Kanter asserts that “to create a culture in which innovation flourishes takes courage,” and advises the courageous to put “innovation at the heart of strategy, and tout it in every message.” It can also take the form of incantatory rituals (to become more innovative, management consultant Abhijit Bhaduri suggests “weekly conversations with millennials to understand how they dream”) and Orientalist fantasies (one former Microsoft executive listed his job title as “Innovation Sherpa” on his LinkedIn profile, and the title of “innovation guru” abounds in the world of business consulting). Often it takes the puzzling form of tautology. To be more innovative, says Harvard Business School professor Clayton Christensen, the most revered prophet of the gospel of “disruptive innovation,” you have to “‘think differently,’ to borrow a slogan from Apple.” But Christensen here refers to a slogan that advertises Apple’s capacity for innovation, which lies precisely in the company’s alleged ability to, uh, think differently. In other words, Christensen is arguing that to innovate, you have to be innovative. Tautological certainties like this abound in tech-industry and innovation discourse, whose proselytizers wander in closed conceptual circles in which capitalism, or a world and a history outside of it, is never imagined, much less questioned. Where these modern prophets of the innovation gospel originally hailed it as the means by which capitalism reinvented itself, and disrupted ossified markets and backward social practices, the innovator’s prerogative has gradually morphed into the self-evident aim of capitalist development.

And this means, much like other varieties of secularized worship, the innovation cult accrues greater complements of unquestioned social power. The brand of innovation is everywhere and nowhere. Primary schools, liberal arts colleges, and business schools all claim to nurture it; entire airports could be filled with the business-advice tomes that have claimed to teach it; municipal, state, and federal governments compete to subsidize it; and there are few products that are not advertised to the consumer as delivering the taste, sound, or experience of it.

Back here in the actually existing social world, though, the widespread circulation of the innovation ideal allows us to employ it as a sort of marker to track the major distinguishing features of contemporary capitalist ideology: its celebration of knowledge, rather than labor, as the driving force of the world economy; its ostensible disdain for hierarchy and bureaucracy; its unskeptical celebration of technology; and its reframing of the loss of job “security” as the laudable increase of “flexibility.” The innovator who emerges from this complex history is a contradictory figure defined by an oscillation between what seem like contradictory poles: imagination and production, rebellion and convention, progress and reaction, the common good and private wealth. These paradoxes, too, are part of the idea’s power, since the underlying vagueness of innovation-for-innovation’s sake permits it to mean all things to all people. It cultivates the open-ended creativity typically identified with the artist, and turns it to profit-making ends, and cultivates it in large firms; it is the twenty-first century’s theory of progress, but it is a heroism of office work. It is a theory of novelty that is perpetually repeating itself. And in its political origins, it is a theory of the new for those outraged by the New Deal.

Its current power shows, however, that the leaders of NAM were on to something. To change the way people think about capitalism, you have to start with how they talk about it.

The Innovator’s Agenda. By John Patrick Leary. The Baffler , March 4, 2019. 







When General Motors laid off more than 6,000 workers days after Thanksgiving, John Patrick Leary, the author of the new book Keywords: The New Language of Capitalism, tweeted out part of GM CEO Mary Barra’s statement. “The actions we are taking today continue our transformation to be highly agile, resilient, and profitable, while giving us the flexibility to invest in the future,” she said. Leary added a line of commentary to of Barra’s statement: “Language was pronounced dead at the scene.”

Why should we pay attention to the particular words used to describe, and justify, the regularly scheduled “disruptions” of late capitalism? Published last week by Haymarket Books, Leary’s Keywords explores the regime of late-capitalist language: a set of ubiquitous modern terms, drawn from the corporate world and the business press, that he argues promulgate values friendly to corporations (hierarchy, competitiveness, the unquestioning embrace of new technologies) over those friendly to human beings (democracy, solidarity, and scrutiny of new technologies’ impact on people and the planet).

These words narrow our conceptual horizons — they “manacle our imagination,” Leary writes — making it more difficult to conceive alternative ways of organizing our economy and society. We are encouraged by powerful “thought leaders” and corporate executives to accept it as the language of common sense or “normal reality.” When we understand and deploy such language to describe our own lives, we’re seen as good workers; when we fail to do so, we’re implicitly threatened with economic obsolescence. After all, if you’re not conversant in “innovation” or “collaboration,” how can you expect to thrive in this brave new economy?

Leary, an English professor at Wayne State University, brings academic rigor to this linguistic examination. Unlike the many people who casually employ the phrase “late capitalism” as a catch-all explanation for why our lives suck, Leary defines the term and explains why he chooses to use it. Calling our current economic system “late capitalism”suggests that, despite our gleaming buzzwords and technologies, what we’re living through is just the next iteration of an old system of global capitalism. In other words, he writes, “cheer up: things have always been terrible!” What is new, Leary says, quoting Marxist economic historian Ernest Mandel, is our “belief in the omnipotence of technology” and in experts. He also claims that capitalism is expanding at an unprecedented rate into previously uncommodified geographical, cultural, and spiritual realms.

Keywords was inspired by a previous work of a similar name: the Welsh Marxist theorist Raymond Williams’s 1976 book Keywords: A Vocabulary of Culture and Society. Williams’s goal, like Leary’s, was to encourage readers to become “conscious and critical” readers and listeners, to see the language of our everyday lives “not a tradition to be learned, nor a consensus to be accepted, [but as] . . . a vocabulary to use, to find our own ways in, to change as we find it necessary to change it, as we go on making our own language and history.” Words gain their power not only from the class position of their speakers: they depend on acquiescence by the listeners. Leary takes aim at the second half of that equation, working to break the spell of myths that ultimately serve the elites. “If we understood... [these words] better,” Leary writes, “perhaps we might rob them of their seductive power.”

To that end, Leary offers a lexicon of about 40 late capitalist “keywords,” from “accountability” to “wellness.” Some straddle the work-life divide, like “coach.” Using simple tools — the Oxford English Dictionary, Google’s ngram database, and media coverage of business and the economy— Leary argues that each keyword presents something basically indefensible about late capitalist society in a sensible, neutral, and even uplifting package.

Take “grit,” a value championed by charter school administrators, C-suite execs, and Ted Talkers. On the surface, there’s nothing objectionable about insisting that success comes from hard work sustained in spite of challenges, failure, and adversity. It can even seem like an attractive idea: who doesn’t want to believe, as author of the bestselling Grit: The Power of Passion and Perseverance Angela Duckworth puts it, that success rests “more on our passion and perseverance than on our innate talent” — or the race and income of our parents?
  
What discussions of “grit” scrupulously avoid, Leary writes, is “the obviously central fact of the economy”: poverty. Duckworth and other proponents of grit nod to the limited horizon of opportunity presented to those living in poverty, but insist that grit can help people “defy the odds.” Implicitly, they accept that most will fail to do so: they simply promise elevation to the hard-working, the deserving, the grittiest — that is, to the very few.


“Grit offers an explanation for what exists,” Leary writes, “rather than giving us tools to imagine something different.” Rather than attacking the conditions that make “grit” necessary, the word’s proponents ask women, people of color, and the poor to overcompensate for the unjust world into which they’ve been born. While the need for “grit” is most often preached to urban schoolchildren and people in poverty, its “real audience,” Leary writes, is “perched atop the upper levels of our proverbial ladder,” a position from which inequality doesn’t look so bad.


Leary divides his keywords into four broad categories: first is “late-capitalist body talk,” which imbues corporations with the attributes of human bodies, like nimbleness or flexibility, and shifts focus away from the real human bodies whose labor generates its profits. “Much of the language of late capitalism,” Leary writes, “imagines workplaces as bodies in virtually every way except as a group of overworked or underpaid ones.”

Then there’s the “moral vocabulary of late capitalism,” which often uses words with older, religious meanings; Leary cites a nineteenth-century poem that refers to Jesus as a “thought leader.” These moral values, Leary says, are generally taken to be indistinguishable from economic ones. “Passion,” for example, is prized for its value to your boss: if you love what you do, you’ll work harder and demand less compensation. Some are words, like “artisanal,” that reflect capitalism’s absorption of the countercultural critique that it failed to provide workers with a sense of purpose and autonomy. Finally, there is the category of words that reflexively celebrate the possibilities of new technologies, like “smart”: smart fridge, smart toaster, smart toilet.

As Leary shows, these keywords reflect and shore up the interests of the dominant class. For the tech overlords of Silicon Valley, an “entrepreneur” is someone innovative and savvy, who “moves fast and breaks things.” The entrepreneur alone creates his company’s exorbitant wealth — not his workers, nor any taxpayers who may fund the innovations his company sells. (Elon Musk, for example, has received nearly $5 million in government subsidies). It’s a very useful concept for billionaires: after all, why redistribute that wealth, through taxes or higher wages, to those who didn’t create it?

In these short essays, Leary undermines what Soviet linguist Valentin Voloshinov describes as the aim of the dominant class: to “impart an…eternal character to the ideological sign, to extinguish or drive inward the struggle between social value judgements which occurs in it. ” And in the case of “entrepreneur,” for example, Leary shows that quite a lot of struggle between social judgements is contained in the word.

First defined around 1800 by French economist Jean-Baptiste Say as one who “shifts economic resources . . . into an area of higher productivity and greater yield,” the word was given a dramatically different inflection by political economist Joseph Schumpeter. According to Leary, our contemporary view of entrepreneurship comes from Schumpeter, who believed that the entrepreneur was “the historical agent for capitalism’s creative, world-making turbulence.” When we talk about “entrepreneurs” with an uncritical acceptance, we implicitly accept Schumpeter’s view that wealth was created by entrepreneurs via a process of innovation and creative destruction — rather than Marx’s belief that wealth is appropriated to the bourgeois class by exploitation.

By demonstrating how dramatically these words’ meanings have transformed, Leary suggests that they might change further, that the definitions put in place by the ruling class aren’t permanent or beyond dispute. As he explores what our language has looked like, and the ugliness now embedded in it, Leary invites us to imagine what our language could emphasize, what values it might reflect. What if we fought “for free time, not ‘flexibility’; for free health care, not ‘wellness’; and for free universities, not the ‘marketplace of ideas”?

His book reminds us of the alternatives that persist behind these keywords: our managers may call us as “human capital,” but we are also workers. We are also people. “Language is not merely a passive reflection of things as they are,” Leary writes. “[It is] also a tool for imagining and making things as they could be.”

 The language of capitalism isn’t just annoying, it’s dangerous. By Rebecca Stoner. The Outline.  December 7, 2018. 





Talk with John Patrick Leary on his book.  RisingUp With Sonali, February 15 ,  2019



  

More on John Patrick Leary’s website Keywords for Capitalism






























No comments:

Post a Comment