1. Introduction
States of whole systems often constrain the behavior of their parts. Conditions in a cell, such as its phase in the cell cycle, constrain which genes are expressed. The same molecule can have different effects on components of a cell (e.g., promoting or inhibiting apoptosis) depending on the conditions in the cell. Phenomena such as these are often characterized as involving downward or top-down causation (Noble Reference Noble2006). They are contrasted with cases of bottom-up causation in which a state of a component of a system partially determines the state of the whole (e.g., a genetic mutation impairs fatty-acid metabolism of a cell or bonds between actin and myosin produce muscle contraction). While most researchers do not find bottom-up causation mysterious (it is invoked in reductionistic explanation of properties of a system in terms of its parts), many have found top-down causation to present a puzzle: how does a whole system have effects over and above those of each of its components?
In an attempt to defuse the mystery, Craver and I (Craver and Bechtel Reference Craver and Bechtel2007) distinguished constitution and causation and proposed treating constitution as a relation between lower-level parts and higher-level mechanisms and causation as a relation between entities at the same level. On our proposal, parts constitute wholes; they do not cause their properties. By viewing causation as occurring at all levels, not just the lowest one, we presented ourselves as offering an account that would capture what is characterized as top-down causation without engendering conceptual problems such as those posed by Kim’s (Reference Kim1998) exclusion argument. When one thing acts on a whole mechanism and its components are also modified, we proposed that higher-level causation was involved in producing the effect on the whole mechanism. Since the parts are what constitute the whole mechanism, one or more of them would themselves be changed in that process. No additional causal processes were involved between the whole and the part, although additional causal processes might then ensue in the mechanism as a result (thereby altering the state of the system as a whole).
However, the examples Craver and I introduced fail to bring out in what sense higher levels are involved in producing these effects at the lower level. In offering an example in which a person’s activity (playing tennis) results in a change within the person’s body (altered metabolism), we emphasized the role of lower-level causal relations: “In this and many similar cases, a change in the activity of the mechanism as a whole just is a change in one or more components of the mechanism which then, through ordinary intra-level causation, causes changes in other components of the mechanism” (559). This presents the challenge: why should one treat the whole mechanism as a higher-level entity rather than just a collection of lower-level entities, with all the causality operative between these lower-level entities? This is a version of Kim’s exclusion argument against top-down causation, and a number of critics (see, e.g., Soom Reference Soom2012; Rosenberg Reference Rosenberg2015) have objected that in the end Craver and I, like Kim, allow causation only at the lowest level. In pushing a similar objection, Fazekas and Kertész (Reference Fazekas and Kertész2011) argue that constitution should be expressed in an identity claim, and this undermines any sense of autonomy for higher-level causation.
My goal in this article is to unpack Craver’s and my distinction between causation and constitution by explicating more clearly (1) when an entity or activity that is regarded as at a higher level enters into causal relations such that the causality should be treated as at a higher level and (2) the relation that holds between the state of the mechanism as a whole and the state of its components. While bottom-up causation has seemed less mysterious, the same problem arises in cases of supposed bottom-up causation: how do operations of parts of a system have effects on the whole when what they seem to have is effects on other parts, which together constitute the whole? Whereas top-down causation raises the question why the whole is considered the cause, bottom-up causation raises the question why effects are assigned to the whole.
Crucial to any account of interlevel causation is the notion of level that is invoked. Craver and I dissociated our treatment of levels from many in the literature, such as the notion of levels of science that were invoked in the theory-reduction literature or notions of levels defined in terms of the size of entities (e.g., molecules, cells, organs, organisms; for a frequently cited example, see Churchland and Sejnowski [Reference Churchland and Sejnowski1992]). In the context of discussions of top-down causation, neither of these senses brings out what many find to be problematic. There is nothing problematic with objects studied in one discipline having causal effects on those studied in other disciplines or for large entities to have causal effects on small entities. Craver and I restrict our account to mechanistic levels in which a mechanism consists of appropriately organized parts performing operations. The notion of a mechanism causing the state of its parts does bring out what many have found problematic in talk of top-down causation since the state of the parts and the state of the mechanism are not independent in the sense required for one to cause the other.
What is crucial to the mechanistic conception of level is the idea of an entity being constituted by its parts and operations. The notion of constitution is what must be explicated, and section 2 will highlight this challenge by developing an example biological mechanism that I will use through the rest of the article. Then, in section 3, I introduce a framework for thinking about systems that has been extensively applied in systems biology (Barabási and Oltvai Reference Barabási and Oltvai2004) and neuroscience (Sporns Reference Sporns2010): graph-theoretic representations of networks. Appealing to graph theory may appear paradoxical as graph-theoretic representations do not explicitly advert to levels—all nodes are represented on a plane. Many graphs that characterize biological systems, though, involve modules in which there is high clustering of nodes often around one or more hubs. Many of these modules, on the account I am offering, constitute higher-level mechanisms. One must show not only that high-level mechanisms can be identified but that the effects on the parts when the whole mechanism is affected correspond to what has led to talk of top-down causation. That is, the condition of the whole mechanism must result in different behavior of the part than would occur when the conditions in the whole mechanism are different. This requires that we turn from the structure of modules to their functioning. The clustered nodes in biological networks are typically not ordered sequentially; rather, they are connected in such an interconnected manner that one can identify multiple feedback loops. When the operations corresponding to the edges in graphs are nonlinear, interconnected modules can exhibit complex dynamic behavior such as oscillations. As a result of dynamic behavior within the module, the parts themselves do not respond to external inputs in the same manner on all occasions. How they respond depends on the current conditions of the module. This is a diagnostic symptom of top-down causation.
The graph representations that I introduce in sections 3 and 4 help make it clear when it is appropriate to identify mechanisms as higher-level components: mechanisms are highly interactive modules within a larger network that are capable of exhibiting complex dynamics. But as a result of treating all interactions, those within the module and those between modules, as edges, the graph representation does not make manifest why the effects of mechanisms on their constituent components are different from the propagation of causal effects throughout the network. Instead, the graph representation may reinforce the perception that all causation is at the lowest level represented by individual nodes in the graph. To address this issue, in section 5 I will deploy a distinction between constraints and dynamical laws that has been introduced into theoretical biology from physics (Pattee Reference Pattee1971; Hooker Reference Hooker2013). Constraints reduce the degrees of freedom that are left open by dynamical laws alone by, for example, establishing correlations between variables or restricting the range of variables. Although the imposition of constraints and the propagation of effects of constrained systems involve diachronic activity, the constraints exercised by a whole mechanism on its parts are synchronic: they are realized through the organizational relationships between the components at a time. Given the organization of the mechanism, the responses of components that are altered when the mechanism itself is affected by external causes are constrained by the whole.
2. When Top-Down Causation Seems Problematic: Levels of Mechanisms
The notion of mechanism invoked in the conception of levels I am addressing arises from the practice of biologists over the past several centuries. In the life sciences, investigators developing explanations often (1) begin by identifying the mechanism responsible for a specific phenomenon to be explained, (2) proceed to decompose the mechanism into its parts and the operations they perform, and (3) finally recompose the mechanism to show how, as a result of the organized parts orchestrating their operations, the mechanism generates the phenomenon. While this practice has been pursued for several centuries, it has assumed a central place in philosophical accounts of explanation in the last couple decades (Bechtel and Richardson Reference Bechtel and Richardson1993/2010; Machamer, Darden, and Craver Reference Machamer, Darden and Craver2000; Bechtel Reference Bechtel2006, Reference Bechtel2008). The steps of decomposition and recomposition are what invite employing the word levels. Parts are, in a rather natural sense, at a lower level than the mechanism constructed out of them.
This conception of level can be illustrated using an example to which I will return throughout this article, the phenomenon of circadian rhythmicity in mammals (i.e., daily oscillations in behaviors and physiological functions that are endogenously generated but entrainable to day/night cycles in the local environment). The responsible mechanism resides in individual cells. The genes Per and Cry and the proteins synthesized from them, PER and CRY, are major parts.Footnote 1 PER and CRY form a dimer and, after being transported back into the nucleus, inhibit their own transcription by interfering with the activators, BMAL1 and CLOCK.Footnote 2 These genes and proteins occupy a lower level than the mechanism itself.
Using this example, we can now see what is sometimes regarded as problematic about top-down causation. Whether Per and Cry are transcribed and translated into the proteins PER and CRY depends on the phase of oscillation the host cell is in. The oscillatory phase, a state of the whole cell, is determining the behavior of its parts, seemingly in accord with accounts of top-down causation. However, the phase of the oscillator at a time just is the concentrations of PER, CRY, and a cadre of related proteins that make up the mechanism. Treating the concentrations as causing the phase, or vice versa, seems to violate many intuitive aspects of causation. Causation is often understood as involving contact action or a propagated signal (Hitchcock Reference Hitchcock2003). Moreover, causes are assumed to precede their effects. Both conditions require that causes and effects be wholly distinct (see Lewis [Reference Lewis2000] for detailed arguments as to why causes and effects must be distinct). The phase of the oscillation in the cell is not distinct from the concentrations of PER and CRY. More generally, the parts and wholes of a mechanism are not distinct—each requires the existence of the other. There is no possibility for transmission between parts and wholes since they are not distinct. Neither the states of the parts nor the state of the whole comes before the other—they occur simultaneously. Talking of top-down causation in these cases also seems to engender the problem of redundant causation highlighted by Kim (Reference Kim1998). While we might attribute the change in concentration in PER or CRY to the phase of the oscillator, it can also be attributed to specific molecular events, namely, that the proteins have attached themselves to two other proteins, BMAL and CLOCK, and removed those proteins from the E-boxes on the Per and Cry genes, terminating the synthesis of more PER and CRY.
Craver and I sought to make a virtue out of these problems for top-down (or comparable problems for bottom-up) causation by treating the interlevel relation as constitution and noting that a state of the whole mechanism involves at least some of its parts being in appropriate states. Any phenomenon that seems to exhibit bottom-up or top-down causation can be accommodated by viewing causes as operating either (1) between the whole mechanism and other entities outside it or (2) among the parts within the mechanism. The fact that the parts constitute the whole has the consequence that the state of the whole mechanism is changed whenever the state of one of its parts is changed and the state of at least one of the parts is changed whenever the state of the whole is changed. Craver and Bechtel referred to how the constitution relation mediates between changes to the parts and changes to the whole as yielding mechanistically mediated effects.
A major shortcoming in Craver and my account is that by merely pointing to a constitution relation, we left unspecified what it is for parts to constitute a mechanism such that the mechanism is at a higher level. The need to address this question can be seen by examining the three-tier diagram Craver (Reference Craver2007) uses to present his view of levels. At the top an arrow on the left terminates at a darkened oval that represents a mechanism. Another arrow leaves from the mechanism. This is intended to show that when the mechanism receives an input, it generates an output. Dotted lines connect that oval to the one below, indicating that what is below is an expansion of what is above. In this middle tier, the arrow on the left is shown as entering into the mechanism to terminate at one of the ovals within it. The four ovals within are connected by arrows, culminating in one from which the arrow on the right now exits. One of the ovals in the mechanism is again shown in black, and it is expanded in the third tier in the same manner. The figure makes clear the sense in which this account of mechanisms is supposed to give rise to a view of levels: the inner ovals in tier 2 are inside the larger oval, which is then shown in black above it on the upper tier. The question the diagram poses is: what do the ovals represent? Put another way, why is the oval in the middle tier around all four inner ovals, and not some other possible combination of ovals (e.g., just two of the inner ovals in tier 2 or one of these and another outside of the oval shown)?
Motivating the ovals in figure 1 requires an answer to the original question: when do components constitute a mechanism? A mechanism is not just an aggregation of parts, each of which performs an operation. In their characterization of mechanisms, Machamer et al. (Reference Machamer, Darden and Craver2000) emphasize productive continuity between the parts involved in generating the phenomenon. This is critical but not sufficient. A pond is a collection of entities, many performing operations, with productive continuity between operations. But it is not what scientists would call a mechanism. One thing that is central in all accounts of a mechanism is that the parts are those entities whose activities or operations are responsible for the phenomenon attributed to the whole mechanism. In the mechanism literature, the specification of the phenomenon to be explained is often invoked to determine which entities belong to a mechanism and which, even though they are located among the other entities, do not. With respect to the pond, if one identifies a phenomenon, such as maintenance of pH, then one can search for the mechanism—the productively linked operations of entities in the pond that contribute to that phenomenon.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20220105125021191-0154:S0031824800009685:S0031824800009685_fg1.png?pub-status=live)
Figure 1. Craver-style representation of three levels of a mechanism. Adapted from Craver (Reference Craver2007).
As useful as appealing to a phenomenon to be explained is in identifying candidate parts of a mechanism, it is insufficient to fix the boundaries and hence pick out a mechanism as a higher-level entity existing as such in the world. One reason is that the same entity may be involved in the production of several different phenomena. To accommodate this, we require an account in which a mechanism can share parts with other mechanisms. Another reason is that the range of entities or activities that can affect a given phenomenon is not sharply bounded. Often entities very distant in time and space can play critical roles in the generation of a phenomenon (Bechtel Reference Bechtel2015). The first entity in tier 2 of Craver’s diagram has an arrow coming into it from somewhere else, but that entity is, for some unexplained reason, not counted as part of the mechanism.
The issue of where to draw boundaries around a mechanism is in fact a crucial issue in biology. As research proceeded on circadian rhythms, researchers identified multiple feedback loops in addition to that involving PER and CRY. For example, BMAL1, which serves as an activator of Per and Cry transcription, is produced by a feedback loop in which its synthesis is regulated by the nuclear receptors RORα and REV-ERBα, while it together with CLOCK are activators of their synthesis. Figure 2 presents the conception of the core clock mechanism as it was understood around 2005. It presents the circadian clock mechanism as a well-bounded set of components. (The figure itself does not show any inputs or outputs, but in fact there are input signals that serve to entrain the clock to the day/night cycle and output signals that serve to regulate the expression of a wide variety of other genes.) Since then the emergence of new techniques, such as knocking down expression of genes through use of small interfering RNAs, has revealed more than 300 additional genes that are both expressed in a circadian manner and exert effects on the phase or amplitude of circadian rhythms (Zhang et al. Reference Zhang, Liu, Hirota, Miraglia, Welch, Pongsawakul, Liu, Atwood, Huss, Janes, Su, Hogenesch and Kay2009). Many of these genes were already identified as components of other cellular mechanisms. The last dozen years have revealed extensive interactions, for example, between components of basic metabolic mechanisms and core components of the circadian clock. The questions of where to draw the boundary of the clock mechanism and why to draw it there have assumed prominence (for examples, see Bechtel [Reference Bechtel2015]). If Craver’s account is to provide a principled characterization of entities in terms of mechanistic levels, one needs a procedure for limning the boundaries of the mechanism and distinguishing it from other entities that affect the phenomenon but are not identified as parts of the mechanism.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20220105125021191-0154:S0031824800009685:S0031824800009685_fg2.png?pub-status=live)
Figure 2. Major parts and operations of the mammalian circadian clock as understood circa 2005. Not shown are the various kinases that phosphorylate the proteins and determine whether they are broken down or transported into the nucleus.
3. Flattening Levels: Using Graphs to Identify Mechanisms
I will return to the question of what the ovals represent in Craver’s diagram below. First, I turn to a second question posed by his diagram. At the top level, arrows just contact the black oval representing the mechanism. At the middle tier, the corresponding arrows contact the smaller ovals inside the larger oval. At the lowest tier, the arrows penetrate the larger oval that corresponds to one of those small ovals and contacts yet inner ovals. What the penetration of arrows into the mechanism at lower tiers suggests is that the causal effect is not on the higher-level unit as a whole, but on one (or possibly several) of its components. If in the bottom tier, instead of expanding an inner oval, Craver had chosen to expand the initial oval from the middle tier, the point would have been even clearer. Then on the three tiers the arrow representing input to the mechanism would connect, respectively, with the highest-level oval (the mechanism), one of the inner ovals (a component of the mechanism), and one of the ovals within that oval (a component of the component). At the lowest tier, the arrows in and out are no different from those between components, and the idea of causation at multiple levels seems to be lost. In many cases, such reduction to lower levels might seem to be appropriate: when light exposure entrains circadian rhythms, photons affect the melanopsin molecules in the intrinsically photoreceptive retinal ganglion cells. As a result, when neurotransmitters are released, they initiate a signaling cascade within cells in the suprachiasmatic nucleus, resulting in increased Per transcription. All of these events seem to be at a single level. But then in what sense is the clock, a higher-level entity, entrained? And how does its phase, as a result of entrainment, affect the behavior of its components?
This exegesis of Craver’s diagram suggests that the critics who viewed Craver and my account as rendering higher levels epiphenomenal were right. It suggests a highly reductionistic picture of levels according to which causal relations that were supposed to be between entities at higher levels of organization dissolve into causal interactions at the lowest level considered. To bring out this point, I have presented a flattened representation of Craver’s diagram in figure 3. It preserves the relationships between components at the middle and lowest tiers in Craver’s diagram and fills in possible decompositions of the three ovals in the middle tier that were not expanded in Craver’s diagram. To identify what were supposed to be higher-level entities, I have included dotted ovals around the lower-level ovals and a dashed oval around the whole set.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20220105125021191-0154:S0031824800009685:S0031824800009685_fg3.png?pub-status=live)
Figure 3. Flattened redrawing and elaboration of Craver’s diagram, with dotted ovals grouping nodes that correspond to units at the middle and top levels in Craver’s diagram.
If one ignores the dotted ovals, figure 3 corresponds to a graph-theoretic representation of a network. A graph representation uses nodes (ovals) to represent entities and edges to represent relations (perhaps causal) between nodes. In constructing a graph representation, one has to settle on which entities are to be represented as nodes. Often in systems biology the nodes represent relatively low-level entities such as genes or proteins. Here I have simply taken the entities in the bottom level in Craver’s diagram as nodes. This does not presuppose that there is a lowest level (or, as I will discuss below, that all entities are at the same level). As inquiry proceeds, researchers may elect to treat the entity represented as a node as a mechanism and replace the node with a set of nodes constituting its constituents. Edges may represent either structural connections or functional connections (for purposes of this article I will assume they are functional). In graph representations more generally, the edges in a graph may be directed or undirected; but in a graph representation of a mechanism, the edges are directed and represent operations in which one node exercises a causal effect on another.
By showing all edges as between nodes, the graph representation brings out what is challenging in explicating top-down or bottom-up causation: there does not seem to be any principled criterion for identifying levels. The dotted and dashed ovals that correspond to entities at higher levels in Craver’s diagram appear to be purely arbitrary impositions on the graph representation. One could draw ovals that group nodes in different ways. Moreover, since they are not the endpoint of edges, the dotted and dashed ovals or whatever they represent seem to be causally inert. Rather than providing an account of causation at multiple levels, it appears that all higher levels appear to have been rendered epiphenomenal.
I will argue, however, that there is a way to use graph-theoretic analysis to identify structures within networks that correspond to the sort of entities traditionally viewed as residing at higher levels and to show in what respect these entities constrain causal processes. To do so, I need to introduce three measures graph theorists have introduced to characterize graphs:
1. Mean shortest path length is a measure of the average number of edges that must be traversed on the shortest path between two nodes.
2. The clustering coefficient is a measure of how connected to each other the nodes linked to a given node are.
3. Degree distribution is a measure of how the number of connections from different nodes is varied.
In terms of these measures, one can describe different network topologies. A randomly connected network will have a small mean shortest path length (there are, on average, short paths between any two units), allowing for rapid transmission between nodes, but exhibit little clustering of units. Lattice or near-neighbor structures exhibit high clustering, allowing nodes to combine operations to produce collective effects, but have long mean shortest path lengths. Highly clustered units are often referred to as modules. Networks that Watts and Strogratz (Reference Watts and Strogratz1998) designated small worlds retain the short mean shortest path length of random networks but contain modules of clustered units more typical of near-neighbor networks. Figure 4 presents a toy example network in which nodes represent entities and arrows causal interactions between them. It exhibits small-world organization: there are distinct modules of interconnected units but a number of connections linking components in different modules. As discussed further below, the interconnections within these modules enable them to function as higher-level units in which the response to an input from outside depends on the state of the module.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20220105125021191-0154:S0031824800009685:S0031824800009685_fg4.png?pub-status=live)
Figure 4. Graphical representation of modules organized around provincial hubs and integrated through a connector hub.
When these network topologies were being explored, most researchers assumed that node degree would be distributed normally. However, Barabási and Albert (Reference Barabási and Albert1999) found that in many real-world networks, degree distribution corresponds to or approximates a power law: most units have few connections to other units, but some units are very highly connected to other units. Networks with power-law degree distribution are referred to as scale-free as there is no characteristic scale on which to describe the network. Highly connected nodes are commonly referred to as hubs. Provincial hubs are ones that exhibit high clustering and so serve as integrators of activity in modules. Connector hubs, on the other hand, often have low clustering and serve to integrate activity between modules. Although the number of nodes is too small to produce a truly scale-free network, the graph in figure 4 illustrates how modules can be organized around provincial hubs and integrated together through a connector hub.
What is important for present purposes is that the modules in these networks are differentiated on principled grounds: the nodes within a module are more interconnected with each other than they are with nodes elsewhere. The extensive interactions between the nodes constituting a module enable them to work together as a unit. This does not mean that the connections to components outside the module are not important; inputs from them may figure critically in regulating some behaviors of the module and enabling the output of the mechanism to affect other processes. But the enhanced connectivity enables greater coordination within the module, allowing the components to work together, and justifies treating them as constituting a mechanism when they produce a phenomenon we seek to explain. To return to Zhang et al.’s (Reference Zhang, Liu, Hirota, Miraglia, Welch, Pongsawakul, Liu, Atwood, Huss, Janes, Su, Hogenesch and Kay2009) discovery of an additional 300 genes that affect circadian rhythms when knocked down, it is noteworthy that they did not view them as core clock components since they are not nearly as interconnected as the components shown in figure 2. Rather, they viewed the proteins expressed by these genes as external factors that interact with the mechanism. One of the interesting consequences of the application of network analyses in systems biology is that often researchers identify clusters or modules within them that correspond more or less closely to mechanisms that have been identified using more traditional approaches in cell and molecular biology. Ravasz et al. (Reference Ravasz, Somera, Mongru, Oltvai and Barabási2002) provide one example of this. The researchers constructed an overlap matrix for the metabolic network in E. coli and identified substrates that form clusters or modules based on edges corresponding to metabolic interactions. The modules identified through this analysis closely approximated those traditionally identified as constituting mechanisms but often included components not previously identified.
4. From Graphs to Dynamics: Coordinated Dynamics in Mechanisms
So far I have tried to show that biological mechanisms more closely correspond to interconnected modules in scale-free small-world networks (fig. 4) than to ovals in Craver’s diagram (figs. 1 and 3). To appreciate the importance of interconnected modules, we must move beyond graph-theoretic representations and consider the types of dynamic behavior such organization supports. When one starts with one part and traces connections in interconnected networks, such as those shown in figures 2 and 4, one often discovers that the parts affected by the operation themselves perform operations that directly or indirectly affect the operation of the part from which one started. That is, the operations feed back onto parts that were envisioned as earlier in the process. Such feedback does not involve backward causation, since the effects are not on current but future operations of the part, but over time it can give rise to coordinated dynamical behavior that constrains the behavior of the parts of the module.
Understanding feedback has proven challenging for humans. Although we know it was already employed by Ktesibios of Alexandria in the third century BCE to regulate the flow of water in his water clock, negative feedback did not become recognized as a general design principle until the twentieth century, when cyberneticists such as Rosenblueth, Wiener, and Bigelow (Reference Rosenblueth, Wiener and Bigelow1943) and Wiener (Reference Wiener1948) presented it as a principle for enabling engineered and natural systems to achieve target outcomes. Prior to that it had been rediscovered many times, including by Watt, who employed it in his governor for the steam engine, which inspired Maxwell’s mathematical analysis of governors (Mayr Reference Mayr1970). But even when negative feedback was recognized as a design principle for regulating systems to pursue target outcomes, many theorists did not attend to the fact already recognized by engineers that negative feedback could generate oscillations.
Oscillations in biological organisms were often concealed by such techniques as examining mean behavior and not attending to time series data. Physiological processes were envisioned, in accord with Machamer et al. (Reference Machamer, Darden and Craver2000, 3), as proceeding “from start or set-up to finish or termination conditions.” Variation in activity was regarded as noise. But increasingly through the twentieth century researchers came to recognize that a broad range of physiological processes, from glycolysis to neural processing, generate oscillations. Following on models, such as one advanced by Goodwin (Reference Goodwin1965), negative feedback was recognized as a design principle for generating endogenous oscillations, and when oscillations were discovered in biological organisms, investigators proposed negative feedback mechanisms. Circadian rhythms were no exception, and circadian researchers were on the hunt for a feedback mechanism for several decades before Hardin, Hall, and Rosbash (Reference Hardin, Hall and Rosbash1990) provided critical empirical evidence.Footnote 3 Using cloning, they demonstrated that the mRNA and proteins produced by the first identified circadian gene, Period (Per), oscillated with a circadian period, with the concentration of the protein peaking several hours after the mRNA. On this basis, they proposed the feedback loop described above in which PER inhibits the transcription of Per. Mental simulation of the operations proposed reveals how the mechanism could oscillate: when PER levels are low, synthesis of new PER proceeds; but as PER accumulates, it inhibits further synthesis. Only when sufficient PER has broken down can synthesis resume. Mental simulation cannot determine whether the oscillations will dampen or be sustained indefinitely (assuming a sufficient supply of free energy); accordingly, Goldbeter (Reference Goldbeter1995) constructed a computational model, inspired by Goodwin’s, that demonstrated that the proposed negative-feedback mechanism could produce sustained oscillations under physiological conditions.
What is important for thinking about top-down causation is that the overall state of the circadian mechanism determines how components within it behave and how they will respond to perturbations arising outside the mechanism. This state of the whole mechanism might be described in terms of the states of some components of the mechanism (e.g., PER concentrations in the nucleus are high). The effect on Per transcription at that time is determined by conditions generated within the mechanism, not from outside it. This point, however, extends far beyond the circadian example. Any network in which the edges are not all in one direction is subject to complex dynamical behavior, such as oscillation, in which the behavior of parts is constrained by the behavior of other parts in the network.
The constraining effect of the whole on the parts is clearly seen in how the parts respond to external inputs differentially depending on the state of the mechanism. This is manifest in the process by which the circadian clock is entrained to local conditions, especially light conditions. Light has different effects at different times of day, as exhibited in the phase-response curve shown in figure 5 (time is shown in circadian time, according to which 0 corresponds to dawn). During the early part of the night a light pulse delays the phase of the oscillation, but a light pulse late at night advances the phase. During daytime light pulses have no effect. Once details of the clock mechanism in mammals were worked out, researchers determined that entrainment to light resulted when a light signal from the retina to the suprachiasmatic nucleus functions to increase Per transcription. Now researchers could understand why light exposure had different effects on the phase depending on time of day. If Per transcription is already at its maximum, as it would be during anticipated daytime, light input could have no further effect. At dusk and early night, as Per concentration is diminishing due to the feedback mechanism, a light signal can counter the increasing inhibition and keep transcription going longer. This has the effect of delaying the phase of the oscillator. On the other hand, later in the night as dawn is approaching, Per concentration is again beginning to increase due to the endogenous oscillation. A light signal at this phase will speed up transcription, resulting in reaching daytime levels in PER concentration earlier and advancing the phase of the oscillator.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20220105125021191-0154:S0031824800009685:S0031824800009685_fg5.png?pub-status=live)
Figure 5. Typical phase response curve for rodents kept for total darkness and then administered light pulses at different times. Circadian time 0 is the time of expected dawn.
The circadian example makes clear the functional importance of the type of integration of nodes found in modules in network representations of biological systems. The operations of individual components (e.g., transcribing genes) depend on the state of other parts of the mechanism (e.g., the concentrations of the PER protein in the nucleus). As a result of the interconnectivity of the parts, especially the feedback loops, the module identified as the mechanism functions as a unit, with the operations of the individual parts of the mechanism determined by other parts of the mechanism. If there were no feedback loops, the parts that receive the input would not be sensitive to conditions elsewhere and would always respond in the same manner to the input. There would also be little reason to identify the parts as constituting a module. The feedback loops are responsible for the state of the module modulating the responsiveness of its parts to external input. Even though the input to the mechanism may affect only one or a few components, the mechanism as a whole is the relevant unit due to the interconnections that run through the mechanism.
As network analyses have been pursued in fields such as systems biology and neuroscience, it has become apparent that modules often form a hierarchy: modules often have far more connections to some modules than others. The toy network in figure 4 shows interconnections among all the modules that render the whole into a module. The interconnections that give rise to these larger modules also result in dynamics that affect the behavior of the individual modules. A last circadian example provides an illustration of such a hierarchy and how it contributes to control over individual components multiple levels below. An important feature of oscillatory systems, already recognized by Huygens in the case of pendulum clocks, is that a very weak signal is capable of causing the oscillations to synchronize. I noted above that the suprachiasmatic nucleus (SCN), a bilateral structure with about 10,000 neurons on each side, serves as the master circadian clock in mammals. By dispersing SCN neurons, Welsh et al. (Reference Welsh, Logothetis, Meister and Reppert1995) determined that individual neurons sustain oscillations but with widely varying periods. When they are connected in the SCN or in slices, they synchronize with each other, thereby enabling sloppy oscillators to generate a highly regular signal to transmit to the rest of the organism (Aton and Herzog Reference Aton and Herzog2005). Synchronization of components depends on the long-distance connections exhibited in figure 4. It begins with operations in which molecules such as the hormone VIP are synthesized in individual cells. These are dispersed out of the cell to bind to receptors on other cells, initiating a signaling cascade in those cells that ultimately affects the processes of Per transcription and translation within these cells.
To understand synchronization, one must identify modules at two different levels and understand how conditions in modules at each level affect both their own components and those at still lower levels. The interactions between SCN neurons determine the rate and efficiency of synchronization among them, while the interactions within neurons determine how they respond when VIP binds to one of their receptors. Although the individual operations are described at the molecular level, they are constrained by the current conditions in the large-scale modules.
5. Top-Down Effects Due to Constraints Imposed by Networks Dynamics
My strategy in the previous sections has been to use network analyses to identify modules and the dynamic behavior that arises in some modules with rich interactive connections to characterize the constitutive relation that holds between parts and a mechanism. When network analyses are applied to biological systems, one frequently finds small-world organization in which node degree is distributed according to a power law. This has the result of creating modules of highly interconnected nodes in which, nonetheless, several nodes still receive inputs from outside the module. These modules often correspond to biological mechanisms that have been identified through more classical techniques.
In Craver’s diagram (figs. 1 and 3), the ovals drawn around mechanisms seemed arbitrary, but on the network account there are principled reasons for picking out modules. What renders a group of entities into a module are the interconnections and interactions between them. When the entities work together to produce a phenomenon, they count as a mechanism. The interconnections and interactions often yield dynamic behavior in which components in a mechanism behave differently at different times due to activity elsewhere in the mechanism. As a result, determining the organization and dynamics of the module is crucial for understanding the behavior of its parts when they receive external inputs.
A concern I raised earlier about Craver’s and my treatment of top-down causation is that it rendered all causal relations at the lowest level. To first appearances, graph-theoretic representations of networks seem to reinforce that concern. In cases in which things outside the mechanism exercise effects on the mechanism, it will typically be by affecting one or more parts of the mechanism. These altered parts then causally modify other components of the mechanism. Even endogenous activities such as oscillations are explained in terms of the feedback between components of the mechanism. The network representation seems to favor a highly reductionistic account that represents all activity at one lowest level.
This interpretation, however, is mistaken. First, the nodes in a network need not belong to a common level in any of the standard senses. In some cases a graph representation of a network is developed by identifying interactions between entities that might be taken to be at a common level on grounds such as common size or common type of entity. Gene regulation and protein interaction networks are examples. In other cases the nodes correspond to entities that are structurally or functionally connected independently of whether they are situated at what is regarded as a common level. The only sense of level that is explicitly embodied in a network diagram is between nodes and modules that constitute them. This is the familiar mechanistic sense of level. As Craver and I argued, however, this provides only a very local conception of level: the parts of a mechanism that interact are at the same level. This account provides no guidance for judging whether the nodes of different modules correspond to entities at the same level or whether the subparts of the parts belong to the same level. On this conception of level, there are no grounds for treating the nodes in a given graph as at a common level.
Second, although in any graph representation there will be a set of nodes that correspond to what are taken as the basic entities, they should not be treated as representing entities at some base level. At best they represent the entities at which the graph representation bottoms out. Just as researchers investigating a mechanism have a choice as to whether to characterize the behavior of the mechanism as it interacts with other entities, many of them mechanisms (thereby elaborating the characterization of the phenomenon), or to decompose it and appeal to its parts and operations to explain its behavior, those developing a graph representation have a choice as to whether to represent a whole mechanism as a node or decompose it into other nodes representing the parts of the mechanism. On many occasions researchers seek to decompose one part of the mechanism, leaving others untouched. The graph representation will show the components into which the one mechanism has been decomposed interacting with the other mechanisms that have not been decomposed.
Third, in developing a graph representation, one might deliberately represent a set of entities as a single node. Researchers may have already identified the components of a mechanism but deem them not to be relevant to their analysis. For example, in analyzing the interactions between SCN neurons, researchers may choose to treat the neurons as units, ignoring the interactions of molecules within them. The graph representation will employ units for neurons and edges for the connections between them. There are contexts in which the whole SCN becomes a single node and the edges are the connections between the SCN and the various organs that the SCN regulates and that, in many cases, send inputs back to the SCN. The graph representation format does not privilege a lowest level but represents as nodes those entities whose interactions are deemed relevant for understanding the phenomenon of interest.
Although a graph representation provides a convenient way to identify modules in networks, by using the same type of arrow to relate nodes within a module and those between nodes in different modules, it glosses over an important distinction relevant to understanding top-down causation in mechanisms. Looking diachronically, both the edges between nodes inside a module and those between modules may represent causal relations. But in the moment when the mechanism receives causal input from outside by having the state of one or more of its parts altered, the relation between the parts and the mechanism as a whole is not diachronic but synchronic. At a given time, the mechanism is constituted of its various parts. The natural language to use to talk about how synchronically the parts are affected by the whole is that of constraint: being situated in the mechanism constrains the behavior of the part.
The notion of constraint has roots in mechanics. Fundamental laws such as those proposed by Newton characterize possible ways a system might evolve given initial conditions. But in a given situation constraints restrict those possibilities, foreclosing some while leaving others open (Hooker Reference Hooker2013). Constraints such as an inclined plane limit the motion a marble can take as a result of the gravitational attraction between it and the earth. In terms of mathematical representations of the evolution of a system, constraints are captured in relations added to dynamical equations that limit the degrees of freedom available through which the system might evolve.
Some constraints, such as an inclined plane or an electric wire, are fixed with respect to the system in question. They provide a minimalist notion of top-down causation. The path a marble takes after being deposited on an inclined plane is governed by the angle of the plane. Other constraints change, often in response to other activities in the system. A switch in an electrical system can direct electricity along different paths in a circuit. More interesting sets of constraints, and more interesting examples of top-down causation, arise as parts of a system constrain each other in ways that change dynamically. In the much-discussed example of Bénard cells, after the application of heat, molecules begin to move and exert constraints on each other. Eventually coordination between the molecules results in a macro-scale convection pattern in which individual molecules are constrained.
The notion of constraint has been applied to biology by a number of authors (Pattee Reference Pattee1971; Rosen Reference Rosen and Rosen1985; Hooker Reference Hooker2013; Moreno and Mossio Reference Moreno and Mossio2015). A key foundation of their thinking is that biological organisms exist far from thermodynamic equilibrium, and in order to build and maintain themselves, they must constrain the flow of free energy to perform work that builds the structures that perform the constraint role. The different mechanisms constituting a biological organism each play a role in such a process by restricting the range of activities that can occur. The circadian mechanism that provided my example throughout this article figures in such a network of constraints as it constrains various physiological processes (typically by regulating expression of other genes) to occur at times when they can work together to maintain the organism. The parts of the mechanism itself are generated by the operation of the mechanism (the protein PER is synthesized when BMAL1 binds to the promoter on the Per gene) and perform operations in it (PER inhibits the ability of BMAL1 to activate the transcription of the Per gene).
The notion of constraint provides a way to understand how the constitution of a mechanism results in the phenomenon referred to as top-down causation. Appealing to graph-theoretic representations of systems, I have identified mechanisms as dynamical systems that can arise in modules. The degrees of freedom available to entities of such interactive dynamical systems are reduced, and so they behave differently than they would if not part of the system. Especially when the constraints are changing as a result of the dynamical activity, these entities may exhibit different behavior on different occasions.
Thinking in terms of constraints also facilitates a response to the assumption of the closedness of the physical that lies at the foundation of Kim’s exclusion argument. Fundamental dynamical laws do apply universally, and as long as one can specify initial conditions for each entity in the system, one can determine its behavior by invoking these laws. But even to address problems of analytic mechanics, physicists recognize the need to invoke constraints. These constraints are not derived from the laws themselves but must be ascertained empirically and added to the laws to determine behavior. The set of possible constraints is not closed. Yet, only in terms of constraints can one predict or explain outcomes. Moreover, constraints are determined locally. In biology, the constraints imposed in a mechanism are specific to the conditions in the living system. From this perspective, the physical is far from closed but rather is extremely open-ended. Wherever one finds a set of components organized into a module with sufficient interactions, one will encounter constraints that limit the behavior of the components and how they respond to external inputs. The phenomenon described as top-down causation is not unusual, but common.
6. Conclusion
Craver and I proposed that bottom-up as well as top-down causation could be understood by limiting causal processes to within levels of mechanisms and treating the constitution of mechanisms as mediating between levels. We did not, however, provide an account of what it was about constitution of mechanisms that motivates appeals to effects or causes at other levels. Invoking graph representations, in this article I have characterized mechanisms as modules that appear in many graphs of biological networks. Modules consist of nodes that are highly clustered and in that respect are distinguished from other nodes (with which they still have a number of edges). Particularly important in integrating nodes into modules that can exhibit what are taken to be top-down effects are feedback connections that are common in biological systems. With feedback, modules are capable of endogenous activity that results in their being in different states at different times. When this happens, some of their components are in different states. Depending on the state, the component will produce different responses to inputs. This is the sort of behavior that has characteristically motivated talk of top-down causation. In developing this account, I have employed flat graph representations of systems in which modules can be identified. Their dynamics can then be analyzed. This has allowed me to pick out mechanisms in which phenomena traditionally characterized as top-down causation can be understood without directly appealing to levels. But this explication of levels also allows us, if desired, to reintroduce reference to levels and characterize what is meant by top-down or bottom-up causation in nonproblematic ways. Finally, by characterizing how a mechanism affects parts in terms of constraints, I have differentiated how the mechanism restricts the behavior of its parts from the way the parts are affected by external inputs and provided a way to resist the assumption of the closedness of the physical underlying Kim’s exclusion argument.