The realities of social structure are more blurred […] One special case are the procedures of social research, which to an increasing extent are being built in as an accepted part of the validation and legitimation procedures in current American society, for better or worse [White Reference White1965: 10].
Several stories might be, and are, told about eigenvectors [Breiger Reference Breiger2000: 110].
Social network analysis, broadly construed, is performative in much the same theoretical sense and, increasingly, on much the same empirical scale as the economic and financial models examined by “social studies of finance” [Callon Reference Callon and Callon1998; MacKenzie and Millo Reference MacKenzie and Millo2003; MacKenzie Reference MacKenzie2006; Muniesa and Callon Reference Muniesa and Callon2007]. That is the main claim of this paper. I argue that there are three principal reasons for believing it. First, the two cases are alike. Considered as intellectual and practical projects, the structure and trajectory of social network analysis and finance theory show striking similarities in form. Second, the standard of proof is within reach. While the evidence for weak or moderate versions of the performativity thesis is quite good, its strongest formulations are only circumstantially supported by the best evidence we have at present. Thus, the case for a performative dimension to network theory has about the same level of empirical support—which is to say, pretty good—as that for economics. Whether we should interpret this as a mark in favor of the performativity thesis, or a strike against it, is an open question. Third, and more positively, the evidence is relatively new. Paradigmatic cases of the performativity of economics go back some decades, but the performative moment of network analysis is more recent and much less well-explored. I offer several examples.
Network concepts and images are by now central not just to the economic sociology of markets, but to theories of social exchange in general. The importance of network ties to market exchange is a core aspect of what, 25 years ago, was beginning to be called the new economic sociology [Granovetter Reference Granovetter1985]. We visualize the circulation of people and goods in society as flow in a dynamically evolving structure of network relations. As a disciplinary project, network theory has grown from a peripheral position in the early 1970s—or, more charitably, from its niche as a respected but specialized subfield—into a central project within contemporary sociology. Network theory has proliferated and diffused across the intellectual landscape over this period with great success. Its ability to cash out some of its most important theoretical concepts and images in formal methods and usable tools has been a vital part of this process. As in the case of theories of finance and their expression in financial markets, we see network theory and its analytical toolkit embedding and extending themselves in a range of settings. The growth of network theory within sociology, in other words, has been accompanied—and is perhaps by now overshadowed—by its practical embedding in the world at large.
A secondary claim of this paper is that what we are seeing “performed” in these settings is quite different from the model of economic action investigated by MacKenzie, Callon and others, and also quite different from an earlier (and quite successful) effort to evangelize the network gospel. An older public image of the importance of networks in exchange was highly instrumental. It emphasized the benefits of “networking” for individual careers, and the strategic advantage to managers of having well-structured networks with respect to their employees, their suppliers, or their competitors. Despite the success of network imagery as a vehicle for entrepreneurial advantage in competitive markets, the network technologies and models that are being implemented most widely (at least in the public arena) are those that emphasize network effects in their aspect as channels of generalized reciprocity, connectivity, and community-building. While these aspects are not inimical to profit opportunities, in this sense the performativity of networks trends against that of economics.
In what follows, I lay out the performativity thesis in more detail and assess the parallels between the cases of finance and network theory, providing a series of examples for the latter case. I then return to the question of the usefulness of performativity as a concept.
The performativity thesis
Economic sociologists argue about whether it makes sense to think of economic knowledge as performative. Originally articulated by Callon [Reference Callon and Callon1998] and refined by MacKenzie and Millo [Reference MacKenzie and Millo2003], and MacKenzie [Reference MacKenzie2006], the performativity thesis is that economics produces a body of formal models and transportable techniques that, when carried out into the world by its professionals and popularizers, reformats and reorganizes the phenomena the models purport to describe. This is a suggestive idea, and one that admits of stronger and weaker interpretations. In its strongest form, the performative process brings the empirical phenomena into line with the original model. Of particular interest in this approach is the focus—inherited from science studies—on how performative projects are accomplished by way of practical technologies, reproducible models and portable algorithms. The success of economics is not just a matter of a particular conception of rationality serving as a ceremonial gloss on social action [Meyer and Rowan Reference Meyer and Rowan1977]; nor is it a simple instance of ideological indoctrination [Marwell and Ames Reference Marwell and Ames1981]. Rather, tools implementing formal models of action—“calculative devices”—are put in the hands of social agents by the model-builders or their representatives. These devices act as “cognitive prostheses” that enable actors to accomplish calculative tasks previously beyond their reach, but which are required by the theoretical models. When incorporated into the everyday work of market agents, these devices allow real settings to better approximate the original models, and their assumptions. This is the “performative loop” in its most interesting, so-called “Barnesian” form.
Granovetter argued that a social theory of markets should begin with a view of actors as embedded in an evolving structure of concrete social relations. From this perspective, neoclassical economics is fundamentally misconceived, either because “the fact that actors may have social relations with one another has been treated, if at all, as a frictional drag that impedes competitive markets”, or because, when they are examined, models “invariably abstract away from the history of relations and their position with respect to other relations” [Granovetter Reference Granovetter1985: 485-486].Footnote 1 The strong version of the performativity thesis, by contrast, is a kind of backhanded compliment from sociology to economics. Complimentary because it acknowledges the success of economics in prosecuting its claims to objective knowledge of the economy, but backhanded because it claims that this success is not what it seems. Economics turns out to create rather than discover its subject matter. This idea has a beguiling appeal to sociologists. If the story is right, then it seems that sociology, the politically and institutionally weaker field, wins out over economics in the end, like a canny David besting a lumbering Goliath. The natives’ account—something like, “These models work because they are correct, or very good approximations to the truth”—is shown to be wrong, and the sociological account ends up encapsulating and explaining the success of the economic one.Footnote 2
Or so one version of the story might go. The most careful and detailed study of the performativity of economic theory, Donald MacKenzie’s An Engine, Not a Camera [MacKenzie Reference MacKenzie2006], takes a much more circumspect line. The book’s subtitle is not “How Markets Perform Financial Models” but rather “How Financial Models Shape Markets.” To shape is not to create or determine. MacKenzie frames the book by claiming his argument is quite different from an economic sociology built on the concept of embeddedness, but when making his case he is scrupulous in his handling of the data, and much more conservative in his claims. Much of the empirical detail of the book fits quite comfortably into a more conventional institutionalist account of entrepreneurs working hard to consolidate a new field of economic action, yielding a market populated by densely-networked actors with a lot of local knowledge about the particular financial instruments they manage [Healy Reference Healy2006].
Nevertheless, stronger versions of the performativity thesis are increasingly common. MacKenzie [Reference MacKenzie2006, 18-19] distinguishes three kinds of performativity: “generic,” “effective” and “Barnesian” (together with the latter’s negative complement, “counterperformativity”). Generic performativity means the active use of some bit of theory not just by economists but also by economic agents, policy makers and the like. Effective performativity requires that the use of theory not just be window-dressing: it must “make a difference” in practice. Barnesian performativity [named for Barnes Reference Barnes1983, Reference Barnes1988] requires that the use of economics actively alters processes “in ways that bear on their conformity to the aspect of economics in question.” That is, the model or theory must bring participants into line with its picture of the world. In that case the model helps make itself true, in the sense that before its public appearance the system did not behave in accordance with the model’s predictions, whereas subsequently it does. Naturally, it is also possible that a model might undermine the real-world viability of the process it describes. That would be a “counterperformative” effect.
MacKenzie argues that performativity can be thought of as part of “a more general phenomenon: the incorporation of economics into the infrastructures of markets.” He examines four cases from the field of finance theory where ideas developed (mostly) by academics might have had performative effects on the structure and practice of financial markets. These are the Modigligani-Miller “irrelevance” propositions for capital theory; portfolio selection theory and the closely related Capital Asset Pricing Model (capm); random walk models and the efficient market hypothesis (emh); and the Black-Scholes-Merton (bsm) formula for option pricing. Taken together, these ideas form the core of modern finance theory, and they contributed to a revolution in financial markets that began in the late 1960s. The cases are mixed in their support for the performativity thesis. The Modigliani-Miller propositions helped launch modern finance theory but did not have strong or immediate practical consequences. In retrospect, they could be seen as providing intellectual support for less negative attitudes towards debt-financing, and as foreshadowing somewhat the financialization of corporate governance which took place in the 1980s. The capm’s effects were also ambiguous, as its operationalization presented practical problems and its results agreed only fairly well with the data. The use of the model in practice did not improve its fit. The emh’s effects were more direct. It allowed researchers and investors to systematically identify market anomalies. Using the capm as a baseline, a series of studies investigated the existence of investment opportunities that offered excess risk-adjusted returns.
Finally, there are the bsm equations for option pricing. This is one of finance theory’s crown jewels and also MacKenzie’s best case for performativity. If the emh provided an overall vision of how the market should work, then Black, Scholes and Merton provided a technique that could be put to work within the market itself. Here MacKenzie sees Barnesian performativity in action, because, he seeks to show, the method that traders used to identify the discrepancies in option prices was the same, in essence, as the one academic researchers used to assess the accuracy of the model itself.
MacKenzie refuses to oversell his findings, and is very reluctant to say that he has found more than strongly circumstantial evidence that his cases are performative in the strongest sense, where the use of a model in practice actively alters how things work “in ways that bear on their conformity to the aspect of economics in question” [MacKenzie Reference MacKenzie2006: 18]. He does believe that some nontrivial version of the performativity thesis is true. But he only rarely speaks as though it has been empirically demonstrated, saying that “Black-Scholes-Merton option-pricing theory was enacted at the Chicago Board Options Exchange,” or remarking that “what is now performed in Chicago is no longer classic option-pricing theory” [MacKenzie Reference MacKenzie2006, 202]. Much more often, he says that “there will often be an element of conjecture” to such claims, or that there is “a lack of conclusive evidence” in crucial cases, or that the sociological processes “cannot be distinguished” from the strictly economic ones, or that “there is no way of being certain” that the practical adoption of theoretical models led to an improvement in the observed fit of those models [MacKenzie Reference MacKenzie2006: 18; 194; 237; 256].
In considering these claims, economic sociologists have been variously supportive or skeptical. But they have confined themselves to case studies of economic expertise and its effects on the world. Debate has centered on the proper way to characterize economics and economists, and a three-sided argument developed between Performativists, Virtualists, and Granovetterians [Carrier and Miller Reference Carrier and Miller1998; Miller Reference Miller2002; Callon Reference Callon2007; Finch Reference Finch2007]. There are several case studies across different economic sectors or in different national settings. But the idea that a social science discipline other than economics might have performative effects was slow to surface as a problem within the debate in economic sociology.Footnote 3 Sociologists have remained outside the performative bubble, hoping to pop it.
To the contrary, I argue that in its organization, content and effects, a current wave of social-organizational innovation is being built out of network metaphors and with network tools, and that it satisfies the criteria for “performativity” at least as strongly as is documented in existing case studies of economic models and their effects. Network analysis has a very high degree of “performative potential” because of its combination of a strong theoretical framework, its portable toolkit of methods, and its ability to be embedded in the architecture of exchange systems. As it happens—probably not by coincidence—its developmental trajectory is also similar in key respects to the main exemplar of performativity in markets, finance theory.
The parallels to the performative processes described by MacKenzie, Callon and others are direct and robust. Like the early days of finance as a discipline, network theorists begin in small numbers, languish in relative obscurity and have to wait before the force of their work is widely recognized. Like the intellectual products of finance theory, network researchers develop models notable initially for their relatively high degree of abstraction, their commitment to mathematical formalization under strong assumptions, and a combination of intuitively plausible force and lack of validation by any substantial body of data. As time passes, things improve in both camps. Like the early intellectual missionaries of finance theory, network theory begins to find a home in business schools, and versions of bits of it come to be propagated by consultants. Eventually, I argue, in both cases the theory begins to have significant effects on the kind of processes it nominally seeks to describe.
The growth of network analysis
A detailed intellectual history of social network analysis remains to be written. Crudely, we can say that throughout the 20th century a number of researchers and small working groups developed methods and ideas that, in retrospect, are precursors of contemporary approaches. But these groups tended not to reproduce or institutionalize themselves, and neither did they develop into a full-fledged field. One of the few histories of the field describes much of the 20th century as the “dark ages” of network analysis [Freeman Reference Freeman2004]. The intellectual consolidation of network analysis as a self-conscious enterprise began in earnest in the early 1970s as a recognizable theory group established itself [Mullins Reference Mullins1973; Leinhardt Reference Leinhardt1977; Holland and Leinhardt Reference Holland and Leinhardt1979]. Harrison White’s research group at Harvard was an epicenter of innovation, with other important work being done at Irvine, Chicago and also in the Carolinas. Network theory’s diffusion as a practical technology followed on a decade or so later. Network researchers became increasingly familiar in sociology departments and, importantly, in business schools. The presence of network analysis in these settings was accompanied, more or less simultaneously, by “network entrepreneurs” who developed products to sell to firms, and consulting practices to convey them. Often these were the same people who were teaching in the business schools.Footnote 4
Within academic sociology, specialists in theory (itself a formerly high-status specialization in occupational decline during this period [see Lamont Reference Lamont2004 on this point]) gradually came to recognize network analysis as a serious contender in the space of general social theory, rather than, as had previously been the case, a more narrowly middle-range and primarily methodological enterprise. The appearance of Identity and Control [White Reference White1992] helped consolidate this process, as it provided a distinctive theoretical statement encapsulating one of the central lines of work of the previous two decades. Papers such as Emirbayer and Goodwin [Reference Emirbayer and Goodwin1994] and Emirbayer [Reference Emirbayer1997] are indicative of this shift. The resurgence of economic sociology was another important aspect of the process. In its early days, the new field was more or less an alliance of network structuralists and organizational institutionalists. Many of its foundational statements and research exemplars are built up from network foundations. The field still bears the marks of its birth, with the network analysis of firms in markets remaining at the core of research activity [Convert and Heilbron Reference Convert and Heilbron2005; Fourcade Reference Fourcade2007].
The consolidation of network analysis within the academy and initial diffusion into the economy in the guise of firm-level consulting or managerial advising, while important, still leaves a gap between strong forms of performativity in economics and the network case. Surely the sheer scale of the financial instruments and asset positions that might be in part “performed” by financial models have no counterpart in the world of social network analysis. Where is the equivalent of the Chicago Board of Exchange? Or Long Term Capital Management? Or Salomon Brothers? While the analogy works in principle, one might object that there is simply no comparison in scale.
The point is not without force. The institution-building resources at the disposal of the pioneers in financial modeling were quite substantial, as MacKenzie documents, and considerably larger than those available to their counterparts in network analysis. But it is important to properly situate the relative timing of each field’s performative moment, and once we do so the comparison does not seem so implausible. Finance theory got off the ground quite slowly, beginning in the 1950s and mostly in place by the late 1960s. It began to make its way into market settings in the 1970s, not without resistance from established interests. By the late 1980s it was in a very strong position. Network theory really got moving perhaps 20 years or so later. Despite many precursors and lone wolves in its intellectual history, consolidation was quite late in coming.Footnote 5 The tools of the approach began to make their way into the academic mainstream and the world of consulting in the mid to late 1980s, rather than (as was the case for finance) in the late 1960s. And until the 1990s, their practical impact was largely still confined to the world of managerial advice and personal strategy.
The timing of the formation of a viable social network subfield remains to be fully explained. But once the field began to develop, some of the resistance it encountered was due to the difficulties in making its approach tractable on its intended scale. The papers put out by White’s research group, for example, were directly concerned in principle with the theory of large-scale social structure. They explicitly differentiated their approach from a previous generation of sociometric studies of small groups. They also were committed to developing formal tools—“satisfactory methods for aggregating networks among individuals” [White, Boorman and Breiger Reference White, Boorman and Breiger1976, 734]—that could be applied to real data. In triangulating these goals, a consequence was that their analytical breakthroughs far outran the computational resources available to apply them. They were also constrained by the limited availability of data in appropriate form. The theoretical discussion in White, Boorman, and Breiger [Reference White, Boorman and Breiger1976] keeps the macro-structure of role systems for whole societies squarely in view, but four of the five applications of the blockmodels developed in the paper have fewer than 25 cases each.
White’s group was well aware of this problem. Boorman and White [Reference Boorman and White1976, 1441] argued that “[t]he present work tries to take seriously what Durkheim saw but his followers did not: that the organic solidarity of a social system rests not on the cognition of men but rather on the interlock and interaction of objectively definable social relationships.” But, on the other hand, they acknowledged that their reach exceeded their (considerable) grasp: “We see at present no intelligent way to develop role interlock for open networks extending through large populations, even though this topic is much closer to the heart of sociology than is small-group structure.” This problem may have contributed to resistance to network theory as a serious theoretical program. Sociologists outside the fold of network theory during that time might have been forgiven for wondering whether quite so much mathematical firepower needed to rain down on the heads of 18 hapless monks, or whether the social habits of 18 women really required elegant matrix operations in order to be satisfactorily described [Breiger Reference Breiger1974]. To object on those grounds was to miss the point, but in an understandable manner. As the availability of network data gradually increased, such objections became less and less tenable.
White’s group and its descendants were by no means alone in facing this problem. Much the same problems of computational tractability and infeasible data collection faced two other streams of research in the incipient field. While the blockmodeling approach focused on the global structure of networks, with the aim of providing role theory with new analytical foundations (and real tools for analysis), work on “small world” problems of pathways through networks [Sola Pool and Jochen, 1978]—made famous by Milgram [Reference Milgram1967] and descending from Rapoport [Reference Rapoport1953a; Reference Rapoport1953b]—was in much the same position.Footnote 6 A third approach, built around metric methods [Davidson Reference Davidson1983] was better able to handle somewhat larger data sets, but rather than obtaining a complete picture of the network, ego-network information was used to generate multiple measures of social distance, to which scaling methods were then applied to produce a few key dimensions of inferred structure [Laumann and Pappi Reference Laumann and Pappi1973; Laumann and Pappi Reference Laumann and Pappi1976; Laumann, Marsden, and Galaskiewicz Reference Laumann, Marsden and Galaskiewicz1977].
Network analysis developed and extended its methods and range of applications through the late 1970s and into the 1980s [see e.g. Freeman, White, and Romney Reference Freeman, White and Romney1989; Marsden and Lin Reference Marsden and Lin1982; White Reference White1981], and the macro-structural approach was pushed forward by work that connected network theory to other structural and ecological accounts of exchange, affiliation and stratification [Blau Reference Blau1977; Breiger Reference Breiger1981; Cook Reference Cook1982; McPherson Reference McPherson1983]. At the same time, though, formal organizations increasingly became key sites for the development and application of network techniques, because more or less complete data on certain sorts of networks could be collected inside them.Footnote 7
The performativity of networks
For a model to be properly performative, there must be something in the world for it to hook on to. A distinctive feature of economics is, as Foley [Reference Foley2004: 82] explains, “that the very phenomena it studies take a quantitative form (market transactions, accounts) produced directly from the phenomena.” Thus,
The profit and loss of a company, or the net worth of a household, are quantities that are inherent in the existence of the company or household. The economist may have to take some trouble to collect and organize it, but is not required, like the physicist or biologist, to devise instruments to represent the phenomenon studied […] in a quantitative form.
Economic sociologists might properly reply that a considerable amount of work needs to be done—and political conflicts resolved—before something like the accounts of a company are socially available in a way that seems to be “produced directly from the phenomena” [e.g. Carruthers and Espeland Reference Carruthers and Espeland1991; Espeland and Stevens Reference Espeland and Stevens1998]. It is nevertheless true that, as presently constituted, firms and other economic entities do yield such data as a matter of course. This provides the necessary substrate for a performative process to take hold. The same might be said of record-producing systems run by bureaucracies, and especially the state: they too can, under the right circumstances, generate formally analyzable data about their subjects as a matter of routine. Indeed, there is a close (and complex) connection between the development of the state’s capacity to collect formal data, the specific form that data takes, and the growth and consolidation of economics and statistics as professions [Fourcade Reference Fourcade2009; Schweber Reference Schweber2007]. Until recently, no such data-generating substrate existed in most other social settings, or for many other kinds of interaction. This is now changing, as an increasing variety of social exchanges leave digital records in their wake. It is the increasing availability of widely distributed, high-volume, and above all more or less automatically-generated relational data that has enabled the transition from generic network imagery to more effective forms of performativity.
The rapid development of computing power, the infrastructure of the Internet, and the protocols of the World Wide Web, together transformed the capacity to construct, visualize, analyze and build networked systems in practice. They were also accompanied by a major shift in the cultural salience of network imagery. The broad outlines of these developments were not unanticipated, but it is fair to say the system of connections that grew up over the course of the 1990s surprised most observers. Because its constituent elements were built up at different speeds, observers did have time to predict what was likely to be possible a few years ahead, and in some cases pursue the project of building out those possibilities themselves. The growth of the web was greeted with understandable excitement by social network analysts. Here was an amazing demonstration of the power of network effects, unfolding before everyone’s eyes in real time, and on a very large scale. And here also, at last, was the potential to collect, visualize and analyze absolutely vast amounts of data on truly gigantic network structures. That potential had remained a distant possibility for a long time. Within the space of a few years, the size of actually-existing network data sets that were (in principle) accessible jumped several orders of magnitude. The upper limit went from perhaps a few hundred nodes to millions or tens of millions, as in the case of something like the centrally managed aol instant messenger database or, more recently, into the billions with Facebook. It is during this period that we can begin to see the performative potential of network analysis realized. I argue that generic and effective varieties of performativity are widespread. Barnesian performativity is more difficult to establish, but that is true for the finance case as well.
Generic network imagery
Computing power grew up first. It allowed for the accumulation of large databases and the potential to analyze the contents of this information, and also revealed the prospects for surveillance by those in control of information collection and storage. These implications began to be articulated in the 1970s. By the early 1980s, a series of well-publicized breaches of government and corporate computer systems by youthful hackers introduced the public to the idea that database systems were reachable over the phone, possibly by anyone.
Commentators saw that the availability of data presented opportunities for new kinds of data analysis and that this would have broader social effects. A Washington Post article from 1984, for example, discusses the effects on privacy of “a world in which employers are monitoring workers” and “friends and neighbors are prying into one another’s private affairs,” especially “the young computer generation.” Scott Boorman is quoted in the article, and his comments are of interest for the two dimensions of the issue he identifies. The first is the prospect of managers being able to identify “patterns of association between individuals” by way of electronic records [Aplin-Brownlee Reference Aplin-Brownlee1984].Footnote 8 The connection to the business consulting wave of network analysis is clear enough, and in its way recalls the opening image of Burt [Reference Burt1992]. The use of network tools in the consulting business has typically been rather less Orwellian, at least in retrospect, than envisaged here. Network methods are typically sold as tools to reveal the structure of a work group or organizational division to its own members, rather than just to upper management, and thus enhance productivity, creativity and effectiveness.
At around the time that network analysis was becoming more common as a consulting product, in the business literature network imagery and networking metaphors became increasingly prominent also. An early example of the genre is Welch [Reference Welch1980], Networking: the Great New Way for Women to Get Ahead. Other examples include Fields [Reference Fields1983], Youngs and Boe [Reference Youngs and Boe1989], Raye-Johnson [Reference Raye-Johnson1989], Krannich and Krannich [Reference Krannich and Krannich1989], Baber and Waymon [Reference Baber and Waymon1992], and Fisher and Vilas [Reference Fisher and Vilas1992], amongst many others. Looking at these books from the 1980s onward, it is striking to see how many of them are written by and for women. In the business paperback market, as distinct from corporate consulting, networking was often presented as a way for women to connect with the right people and succeed in business by taking advantage of their skills at interpersonal communication.
At this point, the idea of networking in the business world was well-established. Similarly, people were aware that large electronic databases containing detailed personal information—including network information—were held in both public and private hands; and that this data might be accessible to enterprising hackers over the phone. But although the business side emphasized network imagery (connections, payoffs, brokerage, and so on), discussion on the computing side still tended to be framed in “information society” terms [Bell Reference Bell1973] where the revolutionary potential of computers and information technology was acknowledged but its encapsulation and control by existing organizational and institutional forms tended to be assumed.Footnote 9 Boorman’s comments on this point in Aplin-Brownlee [Reference Aplin-Brownlee1984] are of interest:
I would say that the concept of privacy is profoundly changing,” Boorman said. “In the old days, 10 or 15 years ago, an invasion of privacy meant that someone had somehow gotten at some personal secret of yours and had revealed it to some third party or to the world at large.” But large new data bases of “very mundane information” about individuals […] make it “possible to characterize one’s life history on an almost minute-to-minute basis” on and off the job and to use this information for “something much more interesting than ferreting out particular secrets […] I think this goes well beyond the immediate, classic problem of government agencies exceeding their statutory mandate […] In a funny way the people we are most vulnerable to is our own immediate employer.
To be clear, a newspaper article is not the place to find a fully articulated view of this topic (Boorman’s least of all). But the discussion represents a point when the social possibilities of networked information technologies were beginning to come into focus. They foreshadow a shift from a concern with the effects of “computers” as such to an emphasis on the connections between repositories of information, and beyond that to the prospect of detailed personal data being used in more interesting ways than just exposing secrets. The growth of data collection is acknowledged, and the implications for the concept of privacy are picked out sharply. But the social units involved are still the individual, the organization, and the state. Data repositories are seen as allowing for the ex post reconstruction of pre-existing social structure through formal analysis. The networked dimension of the data itself is conceived in much more limited terms as the problem of unauthorized users hacking their way into systems. What is missing is the idea of a network form of organization built out of the flow of quantified but “very mundane information” that might “characterize one’s life history,” and which might be actively constructed by users themselves rather than collected by some supervisory entity.
The core infrastructure of the Internet was already in place by the 1980s, though not especially widespread [Abbate Reference Abbate2000]. The development of the world wide web protocol and browsing software in the early 1990s gave people the basic tools to connect across it, and pushed its growth out beyond government and educational institutions. In the early days of the World Wide Web (before 1995-96 or so), the marvel was the sheer fact of connectivity, the ability to follow threads through a huge network, like a speeded-up version of Milgram’s six-degrees letter experiment. The fact that a network of this sort even existed, was more or less freely navigable, contained a motley assortment of content made available by all manner of people and organizations, and which one could contribute to easily, was remarkable in itself. The dominant metaphors of the period emphasized the flow of information across the network (“the information superhighway”) and its abstracted, slightly ethereal quality (“cyberspace”). The first wave of investment in dot-com startups, however, funded all manner of ill-advised efforts to get people to buy various products online or provided unwanted alternatives to things already available elsewhere.Footnote 10
Otherwise, few websites did anything useful. Retail sites were catalogs. The most widely-used navigational tools were catalogs, too, structured along the lines of the Yellow Pages (recall that Yahoo! was originally an acronym where the “h” stood for “hierarchical”), and early search engines did not perform well. Although the majority of websites were reachable from one another, and the existence of links between them made the network imagery more than just metaphor, the structure of the web had no semantic content built into itself. Prognosticators envisioned a world of much richer information flow and connection [Dyson Reference Dyson1997; Dertouzos Reference Dertouzos1997], but the web during its first boom period did not exemplify it.
Data generation and the rise of affiliation engines
So-called “Web 2.0” services came closer to the forecasters’ ideal. The label originated as the name of a conference [O’Reilly Reference O’Reilly2005], and there has been some disagreement about whether it refers to anything more than a particular round of venture capital investment. Skeptics tend to agree that even if the label is something of a misnomer it still refers to a real phenomenon.Footnote 11 The key innovation in these technologies was not the network infrastructure or basic protocols for data transmission—those were by then taken for granted—but the ability to encode, extract and make useful much of the semantic content of data that had previously been untapped. This has been characterized as a transition from “information silos”—networks of sites that are formally connected but substantively isolated—to “architectures of participation” that are more interactive. Network imagery and concepts were once again brought to the fore, as they were during the 1990s expansion, but with a different emphasis.
Beginning around 2003, new kinds of websites and services appeared that made it much easier for people to assemble multifaceted networks devoted to the cataloguing and exchange of all kinds of information. These sites typically embedded network-analytic techniques in their service, and made an effort either to reveal the network to users or otherwise harness its structural properties in order to do something useful. The emphasis began to shift from websites to internet-enabled services. The precise techniques varied, but were closely related to one another.
Content aggregation and filterting led the way. Pioneered by sites like Slashdot, Kuro5hin, and Metafilter, the idea that the readership of a website could act as the site’s own quality filter had been around for some time. Slashdot instituted a moderation system for comments on stories. Other sites extended the idea to the evaluation of most or all of a site’s content, with users being allowed to vote on whether a story should be retained, highlighted or otherwise promoted. More recently, sites like Reddit generalized the approach and combined it with the ability to record individual preferences. Reddit, for instance, uses its voting mechanism to control which stories are assigned to the front page, and each individual vote from a particular user also contributes to that user’s personal filter for stories they would prefer to see.
The principle of aggregation and rating of news items or comments is easily extended to the more general case of collaborative filtering. Here the preferences of individuals, expressed through choices and ratings about choices, are aggregated using a similarity metric. The goal is to predict whether and which users will like the next item that comes along. Thus, Amazon has long collected information on what its customers buy, asked them to rate what they buy on a five point scale, and then tried to make recommendations on the basis not just of the past behavior of consumers, but the behavior of putatively similar consumers. Collaborative filtering is a hard problem because it is focused on prediction of future likes and dislikes. Its strongest base at present is in computer science, specifically the field of machine learning. But its approach is closely related to algebraic methods for the discovery of categories, where the goal is to cluster individuals or items into subsets in order to make predictions about what people will like.
Rather than rating or upvoting items, users might be invited to tag or classify them instead. Tagging systems allow users to associate labels with objects, such as webpages, photographs, book or article records, and so on. At the individual level the result is an informal classification system where categories usually emerge more or less haphazardly over time, and in any event based on the needs of a particular user. When such systems allow users to see or search each other’s tags, and objects can accumulate tags from many users, the result is a “folksonomy” [Wal Reference Wal2007; Shirky Reference Shirky2004], or a structure defining classes that emerge from multiple instances of labeling. Beyond their utility to users, tags are easily thought of as creating an affiliation network that ties together individuals, items, and categories assigned to items. This makes it possible to suggest items to readers based on their patterns of classification and the classifying work of similar others.
The use and refinement of these methods continues down to the present. However, the rise of huge social-networking platforms, together with the extremely rapid diffusion of mobile computing, has displaced them while also expanding the performative arena even further. The most popular networking services, especially Facebook, rely less on a specific method and more on persuading people to join and build a social network from the bottom up. While the social dimension of services like Facebook, Instagram, or Twitter is clear, the relevance of the methods and tools that are our focus here may be less obvious. Facebook’s success was partially predicated on making it easy for users to find people they knew and, subsequently, making it a routine to observe the ebb and flow of activity on one’s network one or two degrees out. The periodic controversies over the degree of information that Facebook makes available, and the ease with which users may reveal information about themselves or their friends, reflect the organization’s desire to see what, if anything, might define the upper limit of users’ willingness to feed large quantities of mundane but intrinsically relational data to it.
Tellingly, Facebook has become very attractive to social scientists partly because it implements a comparatively rich model of affiliation, and partly because (unlike the Internet as a whole) it can in principle provide direct access to very large quantities of data about its users and their social relations. In this regard Facebook and its competitors appeal to network analysts in much the same way as corporate settings did in the 1980s, but now on a vastly larger scale. Lewis, Kaufman and Christakis [Reference Lewis, Kaufman and Christakis2008] and Lewis et al. [2008] are early examples of research conducted directly on Facebook. Wilson, Gosling, and Graham [Reference Wilson, Gosling and Graham2012] provide a review of the first wave of research based on Facebook data. Subsequently, the company itself has begun to generate public research of its own [Bakshy et al. Reference Bakshy, Rosenn, Marlow and Adamic2012, Bakshy et al. Reference Bakshy, Messing and Adamic2015]. Work at this new scale has moved quickly from detecting network effects [Coviello et al. 2014] to inducing them [Kramer, Guillory and Hancock Reference Kramer, Guillory and Hancock2014]. The data used in such studies are produced by a system that implements a particular model of affiliation, which is then subject to examination, intervention, and refinement. This implementation provides a platform for further social interpretation by users, and investigation by experts.Footnote 12 Naturally, this opens the way to the strongest sort of performativity.
Network analysis and the nature of network data
Formal methods of network analysis are by now ubiquitous in the operation of online services. These methods enable the kinds of network-building, category-making, equivalence-discovering features of these websites to work. The examples covered demonstrate how well-established the generic employment of network imagery is, and how effective these methods are in enabling useful things to happen. That leaves the third and strongest kind of performativity. Recall that Barnesian performativity, as defined by MacKenzie, requires that the use of economics actively alters processes “in ways that bear on their conformity to the aspect of economics in question” [MacKenzie Reference MacKenzie2006: 19]. That is, the model or theory must bring participants into line with its picture of the world.
This is a straightforward thesis to state but it is difficult to specify or observe in practice. MacKenzie’s case for it rests on the career of the Black-Scholes-Merton formula for pricing options. MacKenzie documents the development of the theory behind option pricing and its subsequent application in practice. He argues that the appearance and application of the bsm formula had three main effects. First, the model’s power and elegance legitimated the business of options trading: “it undermined the long-standing cultural association between options and gambling” [MacKenzie Reference MacKenzie2006: 158]. The fact that the basic equations were published work rather than private methods eased their acceptance. Second, Fischer Black sold elegantly-constructed sheets containing Black-Scholes values for options (and associated information) that traders could use on the floor while doing their work. They simplified the process of making trades, though traders using them were occasionally mocked as not being man enough to work without them. Third, MacKenzie argues that the model was put to use in “spreading,” the identification of pairs of options on the same underlying stock where one member of the pair was underpriced with respect to the other. Here MacKenzie sees Barnesian performativity in action, because the method that traders used to identify the discrepancies in option prices was the same, in essence, as the one academic researchers used to assess the accuracy of the model itself:
The most thorough tests of fit were conducted by Mark Rubinstein (1985) […] [In essence] Rubinstein checked whether the graph of implied volatility against strike price was a flat line, as it should be on the model. There was thus a homology between the econometric testing of the Black-Scholes-Merton model and the trading-floor use of the model in “spreading.” When spreaders used the model … it would be precisely deviations from that flat line that they would have identified and that their activities would have tended to “arbitrage away.” It seems, therefore, that the model may have been helped to pass its central econometric test […] by the market activities of those who used it [MacKenzie Reference MacKenzie2006: 165].
This is MacKenzie’s strongest example of Barnesian performativity “a direct performative loop between ‘theory’ and ‘reality’” [MacKenzie Reference MacKenzie2006: 165]. The mechanism here is of great interest because it is not what we typically mean when we say that economic theory has the capacity to make itself true by successfully implanting itself in our minds. Some critics have worried, for example, that the spread of the rhetoric of commodification makes people forget that their motives and actions are not all that well-described by the self-interested vocabulary they use. And experiments in social-psychology and behavioral economics have found that exposure to the lessons of undergraduate economics makes people more selfish (or rational, if you prefer), and more likely to behave like homo economicus. Nothing like this is happening in the case of option pricing. The setting is already a market, and the self-interested motivations of traders do not change a bit. Rather, the model is put to use prospectively in essentially the same way that a researcher would go about testing it retrospectively. It is adopted in practice in a way that mirrors its assumptions and prescriptions. This correspondence is what causes the gap between theory and practice, between economics and reality, to narrow. Moreover, the narrowing happens from the side of practice: by employing the formula to identify and exploit profit opportunities, market actors moved observed prices closer to what the model predicted should be observed.
Is there a parallel in the network case? In the previous section we saw a range of web services that put calculative devices in the hands of users in interesting ways. These devices act as “cognitive prostheses,” in Callon’s phrase—they allow users to do things they were unable to do before, such as easily see three or four degrees out of their social network, or discover which of thousands of strangers is most similar to them in their taste in books or music, or quickly locate people with similar financial goals, and so on. It is a relatively short step from here to taking advantage of these tools in ways that bear on actors’ conformity to some aspect of network theory. To take a simple but significant example, Facebook uses its data on the structure of social relations to routinely suggest lists of “people you may know” to users, with the goal of encouraging users to add those people to their network. In this way, the application works automatically to encourage the closure of forbidden triads in people’s social networks—something which, in theory, should be the case anyway. This is likely also to increase the degree of measurable homophily in the network. Were a complacent analyst subsequently to acquire some Facebook data and run some standard tests on the network’s structure, they would find—to their satisfaction—some confirmatory results about the structure of “people’s social networks.” Moreover, they would be able to claim that these results were plausible partly because of the scale of the data used for the analysis.
The case of homophily also provides a suggestive potential example of counterperformativity. As affiliation engines build social networks, developers and site managers may find homophily to be a problem. Not enough interesting things might be happening if everyone in your network is much like oneself. Site developers therefore have to figure out how to keep things varied. One commentator puts it this way:
If you don’t buy into homophily completely, what can you do? Recommendations increase your pool of interest in very short steps. To break homophily, recommend something for reasons other than “this meshes very tightly with your profile”. This seems heretical at first: the whole logic behind recommendations is to guess at items the user will probably like. But it has to happen. For you to identify their complete region of interests, you necessarily have to show them things in and out of that region […] Doing this creates serendipity: pleasantly surprising the user […] Another way to build in serendipity is to have pivotal navigation: tags, top ten lists, and Flickr’s interestingness measure are all ways to break people out of whatever group they’re in and take them to something new. Links are at the heart of this: we’ve all been lost in clicking our way through a drunkard’s walk of the Internet at one point or another. Inspire that in people: build those links and the metadata behind them into your site from the get-go. Your challenge for this week: spot the social software features of a site you use that encourage homophily, and figure out two ways to break that homophily [Torkington Reference Torkington2006].
What is counterperformative from one theoretical perspective might be seen as performative from another: a recommendation to break homophily by way of manufacturing conditions for serendipity might be thought of as a move from Lazarsfeld and Merton [Reference Lazarsfeld and Merton1954] to Merton and Barber [Reference Merton and Barber2004].
Consider the case of Google. It is important to note that although it is older than the “Web 2.0” wave of startups, Google was a pioneer in the methods now associated with the latter. As Graham [Reference Graham2005] remarks:
Suppose you approached investors with the following idea for a Web 2.0 startup: “Sites like del.icio.us and flickr allow users to ‘tag’ content with descriptive tokens. But there is also huge source of implicit tags that they ignore: the text within web links. Moreover, these links represent a social network connecting the individuals and organizations who created the pages, and by using graph theory we can compute from this network an estimate of the reputation of each member. We plan to mine the web for these implicit tags, and use them together with the reputation hierarchy they embody to enhance web searches.” How long do you think it would take them on average to realize that it was a description of Google?
As is well known, the PageRank algorithm allows Google to order search results based on a calculation of the reputation of pages it finds, where reputation is roughly the number of links a page received, but with the importance of each incoming link itself weighted by the reputation of the page it comes from. In the early papers defining the PageRank method [e.g. Page et al. Reference Page, Brin, Motwani and Winograd1998], Google’s founders explicitly link their approach to network-based methods of citation counting in library science and social network analysis. PageRank began life as a method of effectively treating webpages as a giant adjacency matrix and calculating their eigenvector centrality. As Google’s scale and scope increased, the methods and their implementation became rather more complicated. But this idea remains the central reason Google outperformed every other search engine in the early days of its life. The patent for PageRank [Page Reference Page2001] cites Mizruchi et al. [Reference Mizruchi, Mariolis, Schwartz and Mintz1986] and Doreian [Reference Doreian, Batagelj and Ferligoj1994]. It is straightforwardly a network measure.Footnote 13
In the case of the Internet, Google is the first and still the major example of the power of figuring out the right metric, or measurement technique. Its methods take advantage of properties of the network structure itself, so there is much more at work than just the generic use of network imagery. The PageRank system helps reveal the implicit network structure of reputation to users, and allows users to do things with it.
Effective search is like safe payment methods. In the absence of either, many routine online actions and services would not be doable at all, or could not be done nearly as efficiently. Indeed, it is worth rereading Page et al. Reference Page, Brin, Motwani and Winograd1998 to see just how poorly its main competitors at the time performed by comparison. Once it became known that Google’s results were also effectively a measure of reputation, other uses for its results quickly suggested themselves.
A key problem for the Barnesian version of the performativity thesis is to explain the success of a formal model in a given market. MacKenzie sees this difficulty but does not fully resolve it. In particular, it is unclear what the relationship is between the sociological feedback loop MacKenzie identifies and the substantive accuracy of the model in practice. As he notes, not just any bit of algebra could have been “performed” successfully: if a formula containing a serious mistake was put to work on the trading floor, it would have created exploitable errors and quickly been driven out. But if we grant that, on balance, the circumstantial evidence tends to favor Barnesian performativity in the case of BSM, then it is equally plausible to say that Google’s PageRank formula was performative in much the same fashion.
PageRank boiled the complexity of searches on the network down to an algebraic expression, on the one hand, and an effective implementation, on the other. It began life as an academic exercise to discover a better method for finding what you wanted online. When released into the wild it was fantastically successful on its merits. It was rapidly incorporated into the online practices of a majority of Internet users, and then became the focus of efforts to further assess its quality, and indeed to probe and exploit its weaknesses. In this respect, the method transformed public and professional understandings of what search results were, shifting them from a paradigm rooted in an analogy to entries in a catalog or directory to one that understood high ranking in search results as the outcome of a reputational process within a network.
Academic analysis acknowledged its practical dominance, on the one hand, and its place in a family of well-known related methods, on the other [Langville and Meyer Reference Langville and Meyer2005]. The broad outlines of the PageRank method were public from Google’s beginnings (though the details of its fine-tuning have never been), and so knowledge that Google’s method was robust and based on reputation-weighted links may well have led users to adjust—or at least become more self-conscious—about the rank of their own sites, and that of those they sought links from. It gave birth to a mini-industry euphemistically named “Search Engine Optimization” that in essence attempted to beat Google’s methods by leveraging them, in a way very similar to the actions of traders looking for anomalies in options pricing. These efforts led to the discovery and spread of several “corner” cases such as “Googlewhacking” and “Google Bombing,” for instance. And once the anomalies were exposed and exploited, the PageRank algorithm was periodically modified in order to make particular techniques less effective, to eliminate them altogether, or to encourage site administrators to adopt new kinds of behaviors that Google approved of.
In all of this, the calculation of modified eigenvector centrality was leading a social life not very different in principle from the bsm equation. The intense focus of derivative markets at the Chicago Board of Exchange is one kind of highly concentrated performativity. The net effects of billions of separate but interdependent web searches, and the actions that flow from them, is another more diffuse but hardly less significant variety.
The performativity of what?
What should we make of the performativity thesis in the light of these network cases? Those who have applied the idea to financial markets might see the network case as evidence for the robust generalizability of the thesis. Network theorists might argue that their methodological innovations allow pre-existing relations to be discovered, rather than created. Institutionalists could plausibly claim that performativity is best understood as a species of normative isomorphism. I suggest that none of these interpretations is entirely satisfactory.
If there is a strong disanalogy between the performativity of economics and the performativity of social network analysis, it might lie in differences in the connections of each to their respective intellectual precursors. In the case of economics, one might claim that the methods put into practice in market technologies were rooted fairly strongly in that discipline, and there is a direct connection between the work of particular economists—Modigliani and Miller, Treynor, Sharp, Black et al.—and the transformation of financial practice. The same does not seem to be true of network analysis. The methods discussed above, from PageRank to the equivalence methods at the heart of many services, have diverse roots. Some come from computer scientists in the world of machine learning; some from researchers in library or information science; some grew out of statistical methods for the identification of clusters and the reduction of high-dimensional data; some are based on the analysis of complex systems; and some are rooted in social network analysis as practiced by sociologists. Formally, many of these methods are very similar. All of them have been given a major push by the rise of cheap computing. They are part of a general toolkit of applied statistics rather than the product of a particular field’s needs.
This argument is not quite right, however. It rests on an illegitimate winnowing of the disciplinary history of economics and finance in order to sharpen the contrast with the heterogeneous origins of network methods. The technical methods of economics and finance have variegated roots, too, and do not form a single line of development going back to, say, the first marginalists. Indeed, the relationship between the technical core of modern economics and the toolkit of other technical fields in the 19th and early 20th centuries is a subject of some controversy [Mirowski Reference Mirowski1989, Reference Mirowski2001]. Moreover, while modern finance is closely connected with economics, it is not an accident that the former is a separate discipline, with its own system of qualifications, and not a subfield of the latter (“ketchup economics” arguments like Summers Reference Summers1985 notwithstanding).
Consider the “what” question with respect to the models implied by the embedding of network methods in online services. As the history of social network analysis itself makes clear, formal methods can be placed in the service of quite different visions of social process. The consolidation of social network analysis in sociology came out of an effort to place classical concepts of social roles and solidarity on a new footing. But the expansion of the field and its first successes in the world came out of an interpretation that stressed the instrumental benefits in terms of strategic power and sheer profit that the cultivation of certain sorts of networks might bring.Footnote 14 Today, the data-generating capacity of online services has allowed the original, more comprehensively sociological vision of network analysis to return to the fore, as practices (such as the formation and maintenance of friendship ties) that previously left little in the way of quantitative material to work with now leave analyzable records as a matter of routine. Network metaphors and methods are being built in to social practices in ways that at once provide new, theoretically-informed tools for social actors, and notionally “raw” data for social network analysts.
The comparison with economics is instructive. Sociologists often decry the insidious diffusion of economic models into all aspects of life. Quantification is seen as the handmaiden of instrumentalization; system consumes lifeworld one bit of formal measurement at a time. The performativity of economics is the latest and one of the more sophisticated accounts of this process. Economics is seen to take a world posited in theory and implement it in practice by way of a cadre of trained professionals and their box of methodological techniques. And yet, as argued here, a family of powerful quantitative methods and models exists which treats data and metadata in network terms. These methods are being used not just in the retrospective analysis and description of the Internet’s structure, but also prospectively in its construction and ongoing expansion.
Accounts that stress the undoubted power of economic models and their realizability in practice should not blind us to the enormous expansion in the performative capacity of the network toolkit over the past 15 years. That capacity is increasingly expressed in the construction, extension and visualization of systems that emphasize affiliation, connectivity and the flow of generalized exchange. Indeed, one of the main problems faced by many internet services is the disconcerting gap between the technical ability to harness people’s desire for sociability and the business possibilities for making sustainable profits from it. In this sense, the performativity of networks might work in opposition to the performativity of economics. At the same time, as relational methods become pervasive, more and more sources of “raw” data about structures of affiliation will be generated as a matter of routine by network-analytic tools. Sociology might soon find itself in a similar position to finance, testing and validating its methods by analyzing a world increasingly built with tools of its own devising.
Acknowledgments
I thank Paul DiMaggio, Marion Fourcade, Gabriel Rossman, and several colloquium and conference audiences for helpful comments on this article. I also acknowledge my debt to Ron Breiger, Joe Galaskiewicz, Miller McPherson, Linda Molm, and Lynn Smith-Lovin, whose corridor-based network induced the idea in this article.