1 Introduction

Woodward (e.g., 2008, 2010) defended the view that true causal claims, satisfying a minimal “interventionist” criterion for causation (Sect. 3 below), can differ in the extent to which they satisfy other conditions relevant to their use in explanatory theorizing—these include stability and proportionality, among others. The stability of a causal relationship has to do with whether it would continue to hold under changes in background conditions. Proportionality, on the interpretation described below, has to do with the extent to which a causal claim fully captures conditions under which variations in some phenomenon of interest occur. Depending on empirical considerations and our target explananda, causal claims at different “levels”Footnote 1 or framed in different vocabularies may satisfy stability and proportionality requirements to differing degrees and thus differ in their appropriateness or informativeness for explanatory purposes. This can help in finding the most appropriate level (and associated vocabulary) at which to frame explanatory claims. Sometimes this framework favors “upper-level” over lower-level explanations, although in other circumstances, depending on the empirical facts, it favors lower-level explanations.

Recently these ideas have been criticized by several writers. Franklin-Hall (2016) criticizes Woodward’s formulation of proportionality (as well as, by implication, other formulations in the literature such as Yablo (1992)). She also argues that, even when suitably reformulated, proportionality cannot be used to motivate a “non-pragmatic” preference for upper-level explanations over those provided by some lower-level theory. (See Sect. 2 for what “non-pragmatic” means.) Related objections to proportionality have been advanced by Shapiro and Sober (2012), and Maslen (2009). Both Franklin-Hall and Weslake (2010) also object to Woodward’s appeal to stability considerations in support of upper-level explanations. They argue that a focus on stability, combined with other features of Woodward’s account, instead leads to the conclusion that upper-level explanations are always “non-pragmatically” inferior to those provided by lower-level theories—particularly those of “fundamental physics”.Footnote 2

This paper responds to these objections and also attempts to place them within a more general framework for thinking about the status of upper-level explanations. I begin (Sect. 2) with some brief remarks about the overall strategy of Weslake’s and Franklin-Hall’s arguments. I then turn to the notion of proportionality, arguing, in agreement with Franklin-Hall, that current formulations of this notion including the formulation in Woodward (2008, 2010), are inadequate. Sections 3, 4 reformulate this notion and make its underlying rationale more transparent. Section 5 then introduces the notion of conditional irrelevance, which I argue is central to the justification of upper-level explanations and to whatever “autonomy” they possess. Section 6 discusses an alternative proposal by Weslake concerning what the “non-pragmatic superiority” of upper-level explanations consists in.

2 Background: non-pragmatic superiority and W-questions

I begin with some remarks about the overall strategy of Weslake’s and Franklin-Hall’s arguments. This will be crucial to some of the more detailed arguments that follow. As noted, both authors organize their discussions around the question of whether it is possible for “upper-level” explanations to be “non-pragmatically” superior to lower-level explanations of a sort that (they suppose) are provided by fundamental physics. They agree, as does virtually everyone, that there are “pragmatic” reasons (where these are taken to include reasons having to do with our epistemic and calculational limitations—hereafter just “limitations”) that support employment of “upper-level” explanations over lower-level alternatives. Because of these limitations, we cannot actually construct or exhibit explanations of many features of upper-level phenomena (e.g., the behavior of the stock market or short term memory) from fundamental physics. They ask, however, whether in addition to these pragmatic advantages, there are further non-pragmatic features that show upper-level explanations to be superior. In effect, they consider situations in which we have available a complete fundamental physics and in which we abstract away from all epistemic and calculational limitations standing in the way of constructing explanations of upper-level phenomena from fundamental physics. They then ask whether in this situation, there would be additional reasons of a non-pragmatic sort for preferring upper-level explanations. Weslake argues that there are such additional reasons, calling this view Autonomy and contrasting it with its denial, Fundamentalism. Weslake claims that interventionism is committed to Fundamentalism and criticizes it accordingly. Franklin-Hall does not explicitly argue that there are non-pragmatic reasons why upper-level explanations are superior to fundamental explanations but she describes this as a widely shared view and faults interventionism for failing to capture it.

An important part of the background to this discussion is Woodward’s (2003) and Hitchcock’s and Woodward’s (2003) accounts which connect the “explanatory depth” of an explanation to its ability to answer a range of questions (w-questions) about the circumstances in which its explananda would have been different. In doing this, the explanation exhibits a pattern of dependence between explanans and explanandum.Footnote 3 My view, as argued below, is that this idea can be thought of as helping to provide an underlying motivation for the use of proportionality and stability in explanatory assessment and hence in finding an appropriate level for framing upper-level explanations. By contrast, both Weslake and Franklin-Hall argue that, putting aside pragmatic considerations, explanations in terms of fundamental physics always answer more w-questions than upper-level alternatives and hence (mistakenly) will be judged superior by the w-question criterion.

I find this overall line of argument unpersuasive for a number of reasons. Here I focus on just one of these, which is crucial to understanding how the w-question criterion (as well as the notions of proportionality and stability) should be interpreted. When interventionism speaks of an explanation answering a wide range of w-questions it means just that—that the explanation actually displays or exhibits a pattern of dependence that provides answers to such questions, showing us how changes in the factors cited in the explanans are associated with changes in the explanandum. I will not try to provide a complete account of what “displays or exhibits” means since what will matter most for my discussion is that we be able to recognize examples in which there is a clear failure of this condition. However, the basic idea is that the candidate explanation should provide information from which one can read off or access the claimed pattern of dependency of the explanandum on the explanans. In physics this typically involves exhibiting a derivation or calculation or solving an equation, either analytically or via perturbation methods, where this connects the explanandum with the explanatory premises one employs. On the interventionist view, there is thus an important difference between, on the one hand, the claim (supposing this to be true) that such a derivation or an answer to a w-question “exists in principle” or is “implicit” (Weslake 2010) in some theory and, on the other hand, actually writing down an explicit set of governing equations and exhibiting solutions to them or displaying the steps in a derivation leading from the explanans to explanandum which show how the latter depends on the former. On the interventionist view, only the latter counts as “exhibiting” or “displaying” an explanation or “providing” answers to w-questions. For example even assuming that the claim that the behavior of the stock market is “implicit” in the standard model of fundamental physics or is “in principle derivable” from this model is true, such a claim does not, in the relevant sense, provide answers to w-questions about stock market behavior and does not amount to the exhibition of an explanation for such behavior.Footnote 4

In contrast to the situation envisioned above, in other contexts the exhibition of a pattern of dependency may simply involve the presentation of a causal claim. To use an example discussed below, the causal claim

(2.1) The presentation of a red target causes a pigeon to peck

will be naturally interpreted by many as implying that a claim that a pattern of dependence exists between whether or not the target is red and whether or not the pigeon pecks. I take this to be a case in which (2.1) exhibits a pattern of dependence because this pattern is readily accessible to those who understand (2.1). This claim of accessibility is supported by recent psychological research—see, e.g., (Sloman and Lagnado 2005) which shows that people readily associate claims like (2.1) with interventionist counterfactuals expressing dependency relationships.

Returning to the distinction between establishing that (1) an explanation exists, in the sense in which it might be claimed, e.g., that an explanation of stock market behavior in terms fundamental physics “exists” or is “implicit” in fundamental physics and (2) actually displaying or exhibiting an explanation of stock behavior, I take it to be uncontroversial that there is some distinction of this sort that can be drawn. However, it is a further question why it matters whether a researcher accomplishes (2) rather than just establishing (1). Why adopt an account of explanation that is sensitive to this distinction? One important reason is that (2) is an important goal of inquiry in its own right that is not achieved just by establishing (1). Even if it is true that, abstracting from our limitations, answers to questions about stock behavior are derivable in principle from fundamental physics, no serious researcher would think that just observing or establishing that this derivability claim is true is an end point in inquiry. Instead a central goal of researchers is to exhibit factors on which the movement of stock prices depends and to explicitly connect the movement of stock prices to these factors. Our current science is such that doing this requires “upper-level” theories of the sort found in economics and finance, rather than appeals to fundamental physics. Some philosophers may think that this is a temporary state of affairs but even if this turns out to be true (a possibility I would regard as extremely unlikely), it remains true that these upper level theories do or at least aspire to do something [described by (2)] that is both valuable and distinct from merely establishing that explanations “exist”.

With this motivation in mind, I will follow Woodward (2003), (especially pp. 157–161, 175–181, see also Woodward 2016b) in understanding the notion of explanation in such a way that an explanation is something that displays or exhibits dependency relations.Footnote 5 This follows ordinary usage (no researcher thinks it is an explanation of the behavior of the stock market to observe that such behavior must be derivable in principle from fundamental physics) but more importantly reflects the idea, developed in more detail in Sect. 5 and “Appendix 1”, that we should think of explanation as a goal that can play a methodological role in guiding inquiry and in theory construction.Footnote 6 It is exhibited explanations that can function in this guiding role.

As another illustration of this idea, consider Richard Feynman’s remark in the mid-1950s that he did not understand why superconductivity occurred (Anderson 2011). Feynman made this remark despite having helped to create the fundamental theory (QED) which governs the behavior of electrons and which (he knew) describes the fundamental physics that “underlies” superconductivity and from which, given the assumptions that Weslake and Franklin-Hall make, superconductivity is presumably “derivable in principle”. As I understand the notion, an “explanation” of superconductivity requires much more than the truth of this derivability in principle claim—it requires actual exhibition of difference-making relations relevant to the upper-level behavior (as in, for example, the Bardeen, Cooper, and Schrieffer theory). It is this that Feynman did not possess and that Bardeen, Cooper and Schrieffer won the Nobel Prize for.

More generally, in what follows I will understand the notion of an explanation answering a w-question in such a way that only explanations that exhibit answers to such questions count as providing answers to them. Similarly stability and proportionality will be understood as criteria for choosing among explanations and causal representations that we are able to exhibit or formulate, rather than criteria for the evaluation of supposed explanations that will never be exhibited.

There is more to be said about non-pragmatic superiority and about how the w-condition criterion should be understood but in order not to obstruct the flow of my discussion, I have relegated additional discussion to two appendices.

3 Interventionism and proportionality

Consider a causal claim of form “X causes Y in background circumstances B” where “causes” is understood to mean that a type level relation of causal relevance holds between X and Y, where X and Y are variables. The following is a slight modification of a proposal in Woodward (2008) which I label (M*)Footnote 7

(M*) X causes Y in B if and only if there are distinct values of X, x1 and x2, with x1 ≠ x2 and distinct values of Y, y1 and y2 with y1≠ y2 and some intervention such that if that intervention were to change the value of X from x1 to x2, then Y would change from y1 to y2.Footnote 8

“Intervention” should be understood along the lines described in Woodward (2003).

When a variable X satisfies (M*) with respect to Y, I will sometimes say that X is a “difference-maker” for Y and that there is a “dependency relation” between X and Y. I will also use the phrases “causal claim” and “causal explanation” interchangeably in what follows, assuming that a causal explanation for some variable Y is simply an assembly of information about the causes of Y (understood in accordance with (M*)) perhaps along with a more or less explicit statement of how Y depends on these causes, along the lines described in Woodward (2003) and Hitchcock and Woodward (2003).Footnote 9

As Woodward (2008) notes, (M*) provides, at best, a kind of minimal condition for causation, in the sense that a claim can satisfy (M*) and yet seem, in other respects, less preferable (less perspicuous, informative, illuminating or explanatory) than other true causal claims, also satisfying (M*), that might be formulated about the system of interest.Footnote 10 In other words, different causal claims can each satisfy (M*) and yet differ along other, additional dimensions apparently relevant to their assessment. Proportionality is one such dimension.

Turning now to this notion, Yablo’s informal gloss is that proportionality requires that causes be “just enough” for their effects, neither omitting relevant detail nor being overly specific in the sense of containing irrelevant detail.Footnote 11 (We will see shortly that this motivating idea is potentially misleading.) In one of Yablo’s illustrations,Footnote 12 a pigeon is trained to peck at targets of any shade of red and only such targets. The pigeon is presented with a scarlet target and pecks. Consider the following two claims:

(3.1) The scarlet color of the target causes the pigeon to peck.

(3.2) The red color of the target causes the pigeon to peck.

Yablo claims (3.2) is superior to (3.1) on the grounds that (3.1) is “overly specific” and fails to be “proportional” to its effect. Yablo does not take proportionality to be a necessary condition for the truth of causal claims (1992, p. 277) but a number of other writers either interpret Yablo as claiming this (e.g., Shapiro and Sober 2012) or advocate this position themselves (e.g., List and Menzies 2009).

To represent (3.1) and (3.2) within an interventionist framework we need to express them in terms of claims about variables. (3.2) does not explicitly display the alternative possible values of the variables figuring in it but it is very natural to interpret it as employing a cause variable RED that is conceptualized as a binary variable that can take either of two values, {red, not-red}, and an effect variable PECK conceptualized as a binary variable capable of taking the values {peck, not-peck}. (3.2) might then be represented as:

(3.3) RED causes PECK

where this is unpacked as implying that

(3.3a) An intervention that sets RED = red is followed by PECK = peck (alternatively, RED = red causes PECK = peck)

(3.3b) An intervention that sets RED = not red is followed by PECK = not peck (alternatively, RED = not red causes PECK = not peck)

Obviously (M*) counts (3.2)–(3.3) as true.

How about (3.1?) One possible way of representing it within an interventionist framework is to employ, along with PECKS, a variable SCARLET that takes the values {scarlet, not scarlet}. On this interpretation, (M*) also counts (3.1) as true. This is because, given the causal structure of the pigeon’s situation, an intervention that sets the target color to scarlet is followed by pecking and there exists an intervention that sets the target to some different color (e.g., blue) that is followed by non-pecking. In my idiolect, I’m inclined to count (3.1) as true, in agreement with the judgment of (M*). Other examples, described below, suggest even more strongly that satisfaction of a proportionality requirement should not be regarded as a necessary condition for a causal claim to be true. Nonetheless it seems to me that Yablo is clearly correct in thinking that (3.1) is in some way inferior to or less perspicuous than (3.2) in describing the pigeon’s situation. I take this to motivate the general idea that (3.1) is deficient but that this deficiency is not a matter of (3.1) being false. Instead, following Woodward (2010), I will take this deficiency to have to do with an explanatory limitation in (3.1)—it is inferior qua causal explanation in comparison with (3.2) and fails to convey or exhibit explanatorily relevant information in comparison with (3.2).

To further spell out this idea, note that, intuitively, within an interventionist framework, there are at least two different ways in which a causal explanatory claim might be deficient:

  1. (1)

    It might falsely claim that some dependency relationship is present when it is not—call this falsity.

  2. (2)

    It might fail to represent dependency relations that are present—call this omission.

When (3.1) is interpreted (in accord with (M*)) as the claim that there exist interventions in which the target is set to non-scarlet which are followed by non-pecking it does not make false claims about the existence of dependency relations. However, under this interpretation, there are facts about how the occurrence of pecking depends on the color of the target (in particular that pecking occurs for non-scarlet but red targets) that are not conveyed by (3.1). In this sense there are dependency relations present in the example that (3.1) fails to capture or represent, so that (3.1) when so interpreted exhibits failures along the omission dimension (2). I will suggest below that many failures of causal claims to fully satisfy proportionality have this feature—they involve claims that correctly represent some dependency relations that are present but that fail to represent others.

Since the extent to which an explanation fails to convey information about dependency relations that are present seems in many cases to be a matter of degree, we should expect this also will be true of proportionality—it too will come in degrees, rather than being an all or nothing matter. (Since truth and falsity don’t come in degrees this is another reason why proportionality should not be regarded as a necessary condition for the truth of causal claims.)

In an attempt to capture this line of thought, Woodward (2010) proposed the following proportionality-like condition (meant to capture the extent to which a causal or explanatory claim satisfies proportionality). Unlike Yablo’s formulation, this was intended to apply to causal claims that relate non-binary variables and that are not necessarily related as determinates and determinables. Causal claims also can vary in the degree to which they satisfy (P):

(P) There is a pattern of systematic counterfactual dependence (with the dependence understood along interventionist lines) between different possible states (or values) of the cause and the different possible states (values) of the effect, where this pattern of dependence at least approximates to the following ideal: the dependence (and the associated characterization of the cause) should be such that (a) it explicitly or implicitly conveys accurate information about the conditions under which alternative states of the effect will be realized and (b) it conveys only such information—that is, the cause is not characterized in such a way that alternative states of it fail to be associated with changes in the effect. (Woodward 2010: p. 298)

One way in which (P) is inadequate is brought out by the following counterexample due to Franklin-Hall (2016). Suppose the causal facts involving the pigeon are as described earlier and consider a new binary variable V, which has two values, scarlet and cyan, with the pigeon pecking when the target is scarlet but not when it is cyan. Consider

(3.4) V’s taking the value = scarlet causes PECK = pecking

(3.4) satisfies both (a) and (b) in P. But (3.4) seems defective in comparison with (3.2) and also seems to violate what Franklin-Hall calls the “spirit” of proportionality, which, following the discussion above, I take to involve the extent to which dependency relations that are present are represented. (3.4) fails to represent some dependency relationships that are present (those involving red but non-scarlet and non-cyan but non-red targets) despite satisfying (P). Note also that (3.4) also satisfies both (M*) as well as (M**) (footnote 7)—thus further illustrating the need for an additional condition on explanatory assessment that goes beyond (M*) and (M**).

Another counterexample is due to Shapiro and Sober (2012). Their immediate target is the following characterization of proportionality:

A statement of the form ‘C caused E’ obeys the constraint of proportionality precisely when C says no more than what is necessary to bring about E. (p. 89)

They consider a case in which real-valued variables X and Y are related by some non-monotonic function F which maps two different values of X—e.g., 3 and 22—into the same value of Y (y = 6), with other values of X being mapped into different values of Y. There is an obvious sense in which X = 3 is not “necessary” (or is “overly specific”) for Y = 6. So proportionality, understood as above, fails in this case. Assuming (as Shapiro and Sober do) that proportionality is put forward as necessary for causation, it seems absurd to deny that, in the example as described, X = 3 caused Y = 6 or to claim that the relationship between X and Y is not a causal relationship. (Note that this relationship is causal according to (M*).) Moreover, Sober and Shapiro’s example seems to tell equally against clause (b) in (P): it seems misguided to contend that a causal claim described by a non 1–1 function is for that reason less preferable or explanatory than a causal claim described by some 1–1 function, but the requirement that “alternative states of the cause are associated with changes in the effect” seems to have just this implication.

Note that it is natural to think about (3.1)–(3.2) as well as (3.4) as raising issues about variable choice—about which variables we should employ in formulating causal relationships. (For example, V in (3.4) is, intuitively, a less “good” variable than RED.) As we see from these examples, different variables can lead to the formulation of different causal claims that, even if true, differ along the omission dimension (2) described above. I take one of the intuitions behind proportionality to be that, other things being equal, we should choose variables that allow for the formulation of causal claims that, in addition to satisfying (M*), do a better rather than a worse job of satisfying (2).

Now consider another example. Suppose that smoking S causes both lung cancer L and yellow fingers Y but that neither L nor Y cause the other. Suppose we represent this causal structure as:

S → Y

Unlike (3.1) and (3.4) the problem is now not that we have the “wrong” variables but rather that we have omitted a variable and an arrow (S → L). In this way we failed to represent a dependency relation that is present. Alternatively, suppose that we adopt a representation in which there is an arrow from Y to L. Here the mistake is a representation of a non-existent dependency relationship, but again this is not because we have the wrong variables.

Discussions of proportionality in the philosophical literature have generally focused on examples like (3.1) versus (3.2 or 3.4), in which problems arise because of deficient variable choice (and in particular, choice of variables with the wrong grain, leading to failure to accurately represent dependency relations). In the interest of consistency I will follow this, distinguishing failures of proportionality (which I will take to always involve failures associated with variable choice) from omitted arrows. I acknowledge some arbitrariness to this decision since it appears that the root defect is the same in both cases—failing to represent a dependency relation that exists. On the other hand, connecting proportionality to this root defect helps to make its underlying rationale more transparent.

With this as background, I suggest the following formulation of a proportionality requirement as a replacement for the formulation in Woodward (2010).

(P*) Suppose we are considering several different causal claims/explanations formulated in terms of different variables and representing different claims about patterns of dependency relations involving some target effect or explanandum E and where all of these satisfy some minimal interventionist condition like (M*). Then, other things being equal, we should prefer those causal claims/explanations that more fully represent or exhibit those patterns of dependence that hold with respect to E.

The notion of representing or exhibiting should be understood along the lines described in Sect. 2. That is, (P*) is to be understood as applying to choices among causal claims/explanations we are actually able to produce or exhibit. Thus (P*) is better satisfied to the extent that causal claims and explanations are formulated in terms of variables and dependency relations that both fully capture those relationships on which E depends that do exist and do not imply relationships that do not exist. (P*) should also be understood as applying to a fixed or pre-specified E. This specification, in addition to the empirical facts that obtain, fixes the range of possible variation that the effect or explanandum phenomenon (or explananda phenomena) exhibits and in turn what the causal claims/explanations we are assessing are required to account for. For example, in the various pigeon examples, E is specified by the variable PECKS which (we assume) can take just two possible values. As an empirical matter, the pigeons will peck in some circumstances and not others and it is this we are trying to account for. Obviously a candidate cause might do well in satisfying (P*) with respect to one effect or explanandum E and less well with respect to some other explanandum E*—this is one reason why in applying (P*) we need to specify what explanandum or explananda we have in mind.Footnote 13

Returning to the examples considered earlier in this section, I take it to be obvious that (P*) judges (3.2) to be superior to (3.1). To apply (P*) to the CYAN example (3.4), note that (3.4) both fails to convey the information that other shades of red besides scarlet will lead to pecking and also fails to convey the information that non-red colors besides cyan will lead to non-pecking. So (P*) correctly judges (3.4) to be deficient in comparison with (3.2) even though (3.4) is true according to M* (and (M**)). Note that if we restrict ourselves to the variable V in the CYAN example, (3.4) does fully convey the dependence of pecking on the values of that variable—the problem with (3.4) is instead that there is another variable RED that can be used to better capture dependencies that (3.4) misses. This illustrates the point that (P*) is to be applied comparatively to alternative claims framed in terms of different variables in the cause position.

Turning next to Sober and Shapiro, we should be able to see that (P*) is not subject to their counterexample since (P*) does not say that to satisfy proportionality (or to come closer to satisfying proportionality) the functions describing causal relationships must be monotonic or 1–1. The function F does an entirely adequate job of representing the full range of dependency relations in Sober and Shapiro’s example and only those relations and so satisfies (P*). In other words, as far as (P*) goes, there is nothing wrong with causal claims such that some variations in the values of a cause variable are not associated with differences in the values of effect variables in a 1–1 manner, as long as the actual patterns of dependence of the values of the effect on its causes are accurately and fully described.Footnote 14 (Of course if no variations in the cause variable are associated with any changes in the effect variable, then the cause is irrelevant to the effect.)

To enlarge on this, consider another variant on the pigeon case which will play a role in Sect. 5: a fine-grained color variable G takes distinct values for each of a number of distinct shades of red (scarlet, maroon etc.) and also different values for a large number of distinct shades of non-red colors, in this way covering the full color spectrum. This is accompanied by a specification of a functional relationship F which maps all the color values corresponding to the coarse-grained color red into the peck value of the PECK variable and all of the remaining values of G into the non-peck value of PECK. From the point of view of capturing the full range of dependency of PECK on target color, an appeal to

(3.5) F, along with fine-grained information about the target color presented on different occasions

seems to do as good a job as (3.2) in accounting for pecking behavior [and satisfies (P*) just as well as (3.2)].

The most that might be said against (3.5) is that it employs a characterization of the cause-variable that is more fine-grained than is necessary given the effect-variable and in this respect is a less efficient or economical representation than (3.2). But this does not show that (3.5) fails to capture the full pattern of dependence of PECKS on target color or that it misrepresents that dependence—that is, that it fails to conform to (P*). Although (3.5) employs a variable that involves discriminations among its values that are overly fine-grained or irrelevant as far as pecking goes, it “corrects” for this by specifying exactly which of those values leads to pecking and which do not, so that the resulting overall dependency relation is just what is conveyed by (3.2). On the other hand (to anticipate my discussion below) note also that for the example as described, if we are interested in accounting just for contrast between the pigeon’s pecking versus not pecking, (3.2) and the RED vocabulary do just as good a job with respect to (P*) as the more fine-grained vocabulary associated with G. As far as (P*) goes, this is enough to entitle us to use (3.2) rather than the explanation that appeals to G. This illustrates how proportionality can license the use of coarse-grained variables, even if it does not require this. In fact, as we shall see, much of the interest of proportionality lies in this fact—in its licensing or permissive role.

4 Lumping and proportionality

After suggesting that we might deal with her CYAN counterexample (3.4) to proportionality by supplementing this notion with the requirement that the variables employed “exhaust” the possibility space (so that a variable whose only values are “scarlet” and “cyan” is not a good variable because it is non-exhaustive, since there are other possible colors), Franklin-Hall advances an additional objection to a proportionality constraint. Suppose that in the pigeon example, in addition to presentation of a red target causing the pigeon to peck, tickling the pigeon’s chin or electrically stimulating its cerebellum also causes pecking. Now consider

(4.1) The presentation of a red target or provision of food or tickling of the chin or electrical stimulation of the cerebellum (other value: none of the above) causes the pigeon to peck (other value: not-peck).

Franklin-Hall claims that (4.1) does an even better job of “exhausting the causal possibility space” than (3.2) (the RED/PECKS explanation) and should be judged superior to (3.2) by a plausible proportionality requirement. More specifically, she claims that (4.1) shows that proportionality combined with exhaustivity “recommends maximally disjunctive accounts, those citing causes that effectively lump together, into a single explanatory factor, every single means by which the effect might, in principle, have been brought about.” (2016, p. 568). Franklin-Hall contends that “such [lumping] accounts are absent from the explanatory annals, presumably in part for their genuine explanatory inferiority; they are pitched at such great heights as to induce a kind of explanatory hypoxia, specifying far too little about what actually brought about the explanandum event to be very explanatory of it.” (2016, p. 568)

In assessing this claim, we need to pay attention to a crucial distinction—the distinction between (different) variables and different values of those variables.Footnote 15 This distinction yields two different possible readings of (4.1). On one reading—call it (4.1*)—three distinct variables are described in the cause position of (4.1): a variable corresponding to whether the target is red, a variable corresponding to whether there is stimulation and a variable corresponding to tickling. On a second reading (4.1**) these variables are collapsed into a single “disjunctive” variable V* with two possible values—“true” if either the target is red or there is stimulation or tickling and “false” otherwise. (4.1**) is then interpreted as claiming that V* causes pecking. I believe this second interpretation is the one Franklin-Hall intends (it is the one that involves “lumping”), but, as we shall see, she also ascribes features to the example that are more consistent with the interpretation (4.1*).

It may seem tempting to suppose that there is no real difference between (4.1*) and (4.1**)—that both are equally good representations of the same causal structure and that in general it is a matter of indifference whether we employ a number of distinct variables in our causal representation or collapse these into a single variable, as (4.1**) does. One powerful reason for thinking otherwise is that the use of directed graphs to represent causal relationships and the methodologies for causal discovery associated with these depend crucially on there being such a distinction.Footnote 16 The graphical representation of (4.1*) is:

figure a

By contrast the graphical representation of (4.1**) is

figure b

Causal discovery algorithms like those described in Spirtes et al. (2000) depend on there being a difference between these two representations. (4.1*G) describes a collider structure which licenses certain inferences (e.g., that RED and TICKLES are dependent conditional on PECKS). (4.1** G) obviously does not imply this. Similarly, when S (in the smoking example) is represented as the common cause of the distinct variables Y and L, this allows us to apply the Causal Markov condition and conclude that Y and L are independent conditional on S. If we collapse Y and L into a single (presumably four-valued) variable X, we have S causing X and we can no longer apply the Causal Markov condition in this way.

Assuming that the choice of graphical structure to represent causal relationships is not completely arbitrary, there must be some basis for decisions about when it is preferable to represent a causal structure by means of distinct variables and when it is permissible (or a good strategy) to lump or collapse these into a single variable. I will not propose a general account of this but, following Hitchcock (2012) and Woodward (2016a), I take one relevant consideration to have to do with a distinction between variables and their values: A single variable can take any one of a number of different values for different systems or on different occasions but when a variable is predicated of a single unit or object, this variable cannot take two different values at the same time—e.g., a particular cannonball cannot have a mass of 10 kg and 20 kg at the same time. (Here the force of “cannot” is logical or conceptual, rather than causal.) On the other hand, for systems of the sort described in Franklin-Hall’s example, it is plausible that if two variables—call them V1 and V2—are fully distinct, all pairwise combinations of their values should be logically or conceptually possible (although there may of course be causal relationships among them that causally exclude certain possible combinations of valuesFootnote 17). In other words, if V1 takes the value v11, this should not constrain, for logical or conceptual reasons, the value taken by some fully distinct variable V2. (One might take this as a proposal for what it means for variables to be fully distinct.) Variables that are not distinct in this sense cannot stand in causal relationships with each other. For similar reasons, if one holds that variables can be experimentally manipulated independently of each other or that causal relationships involving these variables can be independently changed or interfered with, one is treating these variables as “distinct” in the sense just described.Footnote 18 Representations in which distinct variables are represented as such can thus allow us to capture facts about what would happen if causal relations in which those variables stand are independently disrupted (which involve answers to one kind of what-if-things-had-been-different question)—facts that may not be captured by alternative representations.

It seems clear that Franklin-Hall’s description of (4.1) involves the assumption that it is possible for an experimenter (or nature) intervene to set the values of each of the cause variables independently of values of the others—e.g., by varying the color of the target from red to not red and, independently of this, deciding whether or not to tickle the pigeon. She thus treats these cause variables as distinct. Moreover, Franklin-Hall explicitly assumes that the “pathway” by which presentation of a red target causes pecking is distinct from the pathway by which other causes of pecking operate, and that because of this distinctness, these pathways may have independent conditions of breakdown and that it may be possible to interfere with one pathway independently of the other. To use Franklin-Hall’s example, it may be possible to interfere with the RED →PECK pathway by blindfolding the pigeon without disrupting the TICKLES>PECK pathway. Moreover, whatever might be meant by “distinct pathway”, it seems to be an uncontroversial necessary condition for pathway distinctness that the RED variable be distinct from the TICKLES variable in the sense described above. By contrast if we use a single variable V* as in (4.1**) to represent the cause of pecking, we do not capture the facts about independent variability of values of and independence of pathways just described. That is, (4.1**) does not represent the fact that we (or nature) can alter the value of the variable RED independently of the value of the variable TICKLES (and what would happen under such an alteration) in the way that (4.1*) does. Thus the three-variable representation (4.1*) provides a better representation of how Franklin-Hall understands her example than (4.1**).

Putting this in terms of graphical representations, (4.1*G) allows us to represent the result of interfering with the RED  PECK pathway while the other pathways remain intact as a matter of replacing 4.1*G with

figure c

This (arrow-removal) representation (and its non-graphical counterpart) correctly capture what would happen to PECK if we disrupt the RED → PECK pathway and simultaneously alter whether there is tickling and so on. By contrast (4.1**) and (4.1**G) do not represent this information. Thus in comparison with (4.1*), (4.1**) fails answer some what-if-things-had-been-different questions associated with this example and fails to represent dependency relations present in the example. One may think of this feature of (4.1**) as indicating that it satisfies proportionality (understood along the lines of P*) less well than (4.1*). In other words (4.1*) is preferable to (4.1**) from the point of view of proportionality.

For these reasons, if the facts are as described by Franklin-Hall, I see nothing problematic about (4.1) when understood as (4.1*)–(4.1*G). Instead, the judgment implied by the w-question criterion that (4.1*) is preferable to (4.1) seems correct: (4.1*) conveys a wider range of information about what PECKS depends on. Moreover, this judgment is reflected in aspects of scientific practice. For example, if we wanted to construct a regression model of the pigeon’s pecking behavior, it would generally be thought preferable to construct a multivariate equation which corresponds to the graphical representation (4.1*G) with the variables all explicitly related to the dependent variable PECKS rather than a bivariate regression equation with a single independent variable.Footnote 19 There are legitimate questions in such cases about the criteria for when one should stop adding independent variables to a regression equation (see “Appendix 2”) but it is not always regarded as methodologically objectionable to add them. Once we add requirements concerning the conditions under which variables are distinct, we see that neither (P*) nor the w-question criterion license an indiscriminate collapsing of distinct variables into a single general variable, with an accompanying “explanatory hypoxia”.

5 Autonomy and conditional irrelevance

In this section, I want to use some of the ideas from previous sections to explore some larger issues about the status of upper-level explanations and how considerations having to do with proportionality and stability can help us in finding the right level or variables for framing such explanations.

Let me begin by being a bit more explicit about the ways in which considerations of proportionality can guide variable choice. First, a cause or explanans variable may be such that its values cannot be used to provide a full accounting of the range of conditions under which the explanandum takes its values and some alternative candidate for the cause variable may do better on this score in which it case it should be preferred. This is the reason for preferring RED over SCARLET in the pigeon example. A second possibility is more subtle and has to do with the way in which in certain situations considerations having to do with (P*) can license the use of more coarse-grained or macroscopic variables over more fine-grained microscopic variables.

To motivate this idea, I begin with a striking empirical fact: to an extent that may initially seem surprising, the difference-making features cited in many lower-level, fundamental theories sometimes can be absorbed into variables that figure in upper-level theories without a significant loss of difference-making information with respect to the explananda of those upper-level theories. This fact is crucial for understanding how upper-level explanation is possible. To capture this, I begin with the simplest possibility, which I acknowledge is a limiting case, and then relax some of its characterizing assumptions. Suppose it is possible to find a set of variables Xi which are causally relevant to explanandum E holding for system S and which are such that, given the values of those Xi, further variations in some other set of variables Yk characterizing S are irrelevant to (do not make a difference for) E, even though the Yk have much higher dimensionality or degrees of freedom than the Xi. In the example from Sect. 3, given the values of the RED variable, further variations in the values of the more fine-grained color variable G in 3.5 are irrelevant to whether the values taken by the PECKS variable. As another illustration suppose (as is almost but not quite true) that variations in the values of various thermodynamic variables like temperature (the Xi variables above) are difference-makers for those aspects of the macroscopic behavior of a gas that are described by thermodynamic variables (E), and that further variations in its microscopic state (the Yk variables above), as described by the positions and momenta of the individual molecules making it up that are consistent with the values taken by the macroscopic variables are irrelevant to the behavior of the gas.

To further spell this out, let us say, following (M*), that a set of variables Xi is unconditionally relevant (alternatively, irrelevant or independent) to E if there are some (no) changes in the values of each Xi when produced by interventions that are associated with changes in E. A set of variables Yk is irrelevant to variable E conditional on additional variables Xi if the Xi are unconditionally relevant to E, the Yk are unconditionally relevant to E, and conditional on the values of Xi, changes in the value of Yk produced by interventions and consistent with these values for Xi are (unconditionally) irrelevant to E.Footnote 20 In other words, changes in the variables Xi are causally relevant to E in the sense captured by (M*) and conditional on the values taken by Xi, further variations in the Yk make no difference to E.Footnote 21 We can think of this as a generalization of the “screening-off” idea used by Yablo to characterize proportionality described in footnote 11—the Xi screen off the Yk from E.Footnote 22

In a case of this sort, as noted in Sect. 3, we can satisfy the demands associated with (P*) just as well by citing the Xi to explain E as by citing Yk: Thus proportionality understood as (P*) does not drive us in the direction of always preferring more fine-grained or microscopic variables: if we want to explain E, we can just cite the variables Xi, and ignore the lower level variables Yk. As it is often put, one collapses the many degrees of freedom or the high dimensionality associated with variations in Yk into the much smaller number of degrees of freedom associated with Xi. Put differently, the Xi are a permissible coarse-graining of the Yk with respect to E. Thus one does not need to model the system in terms of the Yk, the Xi do just as well. And of course, this is a very good thing if, as a practical matter, there is no possibility of actually constructing or exhibiting explanations that appeal to the variables Yk; and associated laws, but one can exhibit explanations that appeal to Xi.

As I have said, this is an ideal case but I take it to provide an illustration of how and in what circumstances upper-level explanation is possible and how such explanation can be “autonomous” from lower-level details—autonomy here just means that the upper-level variables are relevant to the explanandum E and that the variables figuring in lower-level or more fine—grained theories are conditionally irrelevant to E given the values of the upper-level variables.Footnote 23

This framework differs from the vindication of the non-pragmatic superiority of upper level explanations sought by Weslake and Franklin-Hall in several important ways. First, there is no attempt to argue that the explanation of E in terms of the upper-level variables Xi is “better” than the explanation in terms of the lower-level variables Yk given the contrary-to-fact supposition that one can exhibit the latter explanation. Rather, the idea is that the explanation of E in terms of Xi is no worse than the explanation in terms of Yk even assuming we were able to construct the latter. Second, there is a clear sense in which this justification for the use of Xi is not “purely pragmatic”. Pragmatics does play some role in the justification since in applying (P*) we choose among explanations we are able to formulate and this will reflect facts about our limitations.Footnote 24 However, given that our choice is restricted in this way, pragmatics (in the sense under discussion) does not enter into the evaluation of these explanations in any further way that is disturbing. Thus although our limitations do play a role in our decision to employ the upper-level explanation, the fact that the upper-level explanation is no worse (and that the appropriate conditional irrelevance relations hold) is not just a matter of pragmatics—this fact depends on what the world is like and its significance for explanation depends on the adoption of an account of explanation in which conditional relevance and irrelevance play the role just described.Footnote 25 In this connection, we should also note that, on the account proposed, the invocation of “pragmatics” by itself is not sufficient to explain why upper-level explanations are sometimes possible. In some cases very high dimensional lower-level variables may be relevant to some upper-level explananda E and there may not exist more coarse-grained upper-level variables (of a sort that we are able to formulate or measure) that satisfy the screening-off requirements just described. In such cases, actually exhibiting an explanation of E may be impossible, however strong our pragmatic reasons for wanting to do this may be. This is the “model chaos” described by Goldenfeld and Kadanoff below. What we want in an account of upper-level explanation is not just a role for pragmatics, but also (1) some insight into conditions in the world that sometimes support or allow for the successful construction of upper-level explanations as well as (2) an account of explanation that illuminates how such explanations work. My claim is that this has to do with the obtaining of the right sort of conditional irrelevance relations.

I acknowledged above that the sort of complete conditional irrelevance of all lower-level detail from an upper-level explanandum just described may not be common, although it may not be highly uncommon once we restrict its deployment to explanations and relationships we are able to exhibit. We can weaken the notion of conditional irrelevance in various ways. One possibility is that although there may be rare or exceptional values of the Yk that are conditionally relevant to E, even given the values of Xi, this may not be true for most or “almost all” values of the Yk—for most or almost all such values, the Yk are conditionally independent of or irrelevant to E, given Xi even if there are a few Yk for which this is not true. Or perhaps conditional irrelevance holds for all values of the Yk and Xi within a certain large interval, including those values most likely to occur (at least around here right now). In such cases, standard explanatory practice often is to explain E just by citing the Xi, again especially if it is impossible to actually construct or exhibit an explanation of E in terms of the Yk. Again, we can think of this as providing a justification for the use of the upper-level explanations that is not purely pragmatic—the justification also has to do with the facts about “almost irrelevance” just described.Footnote 26

Here are some examples. I noted above that to the extent that our target explananda involve thermodynamic variables describing the macroscopic behavior of a gas like temperature, pressure and volume, it is almost but not quite true that microscopic variations consistent with the values taken by these thermodynamic variables are irrelevant. What is more nearly true is that this is so for almost all—in the sense of all but a set of Lebesgue measure zero—values taken by those microscopic variables. Thus given this target, we can replace the enormous number of variables (and degrees of freedom) necessary to characterize the full microscopic state of the gas with a much smaller number of variables while still satisfying a proportionality requirement like (P*) as well as making use of stable relationships that bear on what we want to explain.Footnote 27

As a second illustration, consider the following remarks of Goldenfeld and Kadanoff (1999) concerning a simple computational model that reproduces real features of fluid flow despite omitting most details concerning the micro-behavior of the constituents of the fluid. They write:

For physicists it is delightful, but not surprising, that the computer generates realistic fluid behavior, regardless of the precise details of how we do the coding. If this were not the case, then we would have extreme sensitivity to the microscopic modeling—what one might loosely call “model chaos”—and physics as a science could not exist: In order to model a bulldozer, we would need to be careful to model its constituent quarks! Nature has been kind enough to have provided us with a convenient separation of length, energy, and time scales, allowing us to excavate physical laws from well-defined strata, even though the consequences of these laws are very complex (87)

Again, the point is that for many aspects of the systems of interest, variations in microscopic variables either don’t matter at all, given the values taken by certain macroscopic variables, or matter only in certain unusual cases. This is very fortunate since it makes it possible to model or explain important aspects of the behavior of those systems without adverting to these microscopic details. If we had to appeal to these details (which would be the case if even near conditional irrelevance relations fail to hold), the exhibition of an explanatory model would be hopeless. As Goldenfeld and Kadanoff suggest, this possibility (of neglecting various low-level details) is closely linked to the physical fact of the “separation of scales” which in an interesting range of cases, has the result that phenomena occurring at length, time and energy scales S are largely or entirely conditionally independent of what is going on at other scales S* with more degrees of freedom. If we can actually construct (display) a model that explains a range of explananda at scale S in terms of some coarser grained theory T and we cannot do this in terms of some finer-grained theory T* (which, if we could construct it, would also allow for the explanation of phenomena at scale S*) and conditional irrelevance relations of the form described above hold, then T satisfies (P*) and the w-question criterion with respect to these explananda.

Goldenfeld’s and Kadanoff’s view is very similar to the view recommended in this paper. They don’t try to argue that an upper-level explanation of bulldozer behavior would be superior to one in that appeals to quantum chromodynamics even if we could derive bulldozer behavior from this theory. Instead they argue that we don’t need to appeal to QCD because of facts about what nature is like (separation of scales) and this is very fortunate, since modeling in terms of QCD (in the sense of actually exhibiting such a model) is impossible.

We may compare Goldenfeld’s and Kadanoff’s remarks with the following claim from Franklin-Hall:

… judged by [an interventionist] standard of excellence, high-level explanations are uniformly impoverished; they explicitly represent fewer features of the world on which the explanandum depends than do lower–level “micro” explanations, limiting the range of w-questions they can answer. (2016, p. 554)

Even putting aside the point that often it is not true that lower-level explanations “can answer” w-questions about upper-level phenomena in the sense of actually exhibiting such answers, this claim of “uniform impoverishment” seems an exaggeration when “depends” is interpreted in terms of conditional dependence, which I suggest is the most obviously relevant interpretation. It is true that the micro-explanation of, e.g., the behavior of a gas contains a representation of many features not represented in the macro-explanation, but the crucial point is that the macroscopic features of the gas that we are interested in explaining may not depend (or may not depend in almost all circumstances) on these microscopic features, conditional on the other macroscopic variables characterizing the gas. Franklin-Hall’s discussion does not adequately reflect the role of this fact in allowing us to (justifiably) omit reference to these features in constructing an explanation of the macroscopic behavior of the gas.Footnote 28

I acknowledged above that on the account I have defended many upper-level explanations are not fully autonomous. Some readers may find this disturbing or at least disappointing—they were perhaps hoping for some stronger result. I disagree: Insofar as there is some feature of good explanatory practice that needs to be captured or explained, the feature in question is not that (1) upper level explanations that eschew appeal to all lower-level considerations are always or even usually superior to those that do, so that a non-pragmatic account of upper-level explanations must vindicate or support (1). Instead, a non-pragmatic account of explanation that implies (1) is misguided for the reason just described. Here the tendency to identify “lower-level” with fundamental physics (or to think in terms of a simple dichotomy between “upper’ and “lower” level) misleads us. It may well be true, for example, that explanations of behavior appealing only to “psychological” variables are unlikely to be improved by incorporation of information from fundamental physics (in part but not only in part because we have no idea how to do this), but it is plausible that neurobiological variables are relevant to many psychological phenomena even conditional on psychological variables. If so, an explanation that incorporates neurobiological variables as well as psychological variables in a mixed level explanation will be superior to one that appeals only to psychological variables. This seems to me the correct assessment. We shouldn’t be looking for an account that implies otherwise.

6 Weslake on non-pragmatic superiority

I noted above that Weslake advances a positive proposal about what the non-pragmatic superiority of “upper level” explanations consists in. Very briefly, he holds this is a matter of there being “physically impossible [but metaphysically possible] systems to which the macroscopic explanation applies but to which the microscopic explanation does not” (2010, p. 287) so that in this sense the former is “more abstract” and applies to a wider range of possibilities than the former. The presence of these features can make the upper level explanation non-pragmatically better. For example, Weslake claims that the ideal gas law would hold in a world in which the underlying mechanics is Newtonian in addition to the actual world which is quantum–mechanical (2010, p. 291). According to Weslake, this makes an explanation in terms of the ideal gas law non-pragmatically better than a quantum–mechanical explanation.

I do not find this convincing. To begin with, the fact that there are metaphysically (or logically) possible but physically impossible scenarios in which the ideal gas law holds seems in itself irrelevant to its explanatory goodness. Consider a contra-nomic but arguably metaphysically possible scenario in which little men move molecules around in a way that conforms to the ideal gas law.Footnote 29 Why should this possibility contribute anything to the “depth” of the explanations the ideal gas law provides? Or suppose that there are (in the relevant sense) possible but non-existent systems (e.g., composed of silicon) for which the generalizations of folk psychology hold exactly. Why should this fact be relevant to the assessment of the explanatory depth of folk psychology as applied to human beings?

The considerations to which Weslake appeals in connection with the ideal gas law are better captured in the following way, which appeals to the ideas about conditional irrelevance described above. It is a mathematical fact that any underlying micro-theory having certain generic features will lead to the ideal gas law. The quantum mechanical theory that correctly characterizes the actual world has these generic features. As it happens, these generic features or something close to them are also shared by certain quasi-Newtonian models of the gas. Put in terms of the framework described above, we can thus appeal to these generic features rather than more specific details of quantum mechanics to explain the behavior of gases—given the generic features, the ideal gas law is conditionally independent of these more specific details. However—and this is the crucial point—it is the fact that these generic features hold for the actual quantum mechanical laws governing our world that establishes their explanatory relevance to gas behavior. The Newtonian models “inherit” their explanatory relevance from the features they possess that are shared with actual quantum mechanical laws and not because there is a general preference for explanations whose scope covers contra-nomic worlds. Weslake mistakenly interprets the correct idea that whether the ideal gas law holds is conditionally independent of certain details of the underlying physics (and that this has implications for its explanatory status) as a claim about the explanatory relevance of the law’s holding in physically impossible situations. As I have tried to illustrate, conditional independence claims need not be understood in this way.

If, as I have argued, Weslake’s attempt to show that “upper level” explanations are sometimes non-pragmatically better is unsuccessful, then unless there is some other reason for accepting this claim about non-pragmatic superiority, this provides an additional reason for not demanding that accounts of upper-level explanation establish this claim.

I can further clarify the relationship between my own proposal and Weslake’s by commenting briefly on Weslake’s discussion of Woodward (2003) on the relationship between macro-level and micro-level explanations of the behavior of an ideal gas. Weslake takes Woodward to be claiming that there is modal information in the former that is not provided by the latter and that, even abstracting away from our limitations, the former answers w-questions not answered by the latter. He rejects these claims, writing

If we assume a reasonable form of physicalism, then there are no questions that can be formulated in terms of any other variables that do not correspond to one of these questions [about the values of microscopic variables]. So there are no physically possible counterfactuals on which the fundamental physical explanation is silent. The fundamental physical explanation provides the resources to answer any possible w- question. …there is no missing modal information of the kind claimed. (281)

I agree with Weslake that there is no missing modal information. However, I don’t think this observation has the significance that Weslake takes it to have. In particular, it does not undermine the arguments advanced above about role of conditional independence in licensing more upper-level explanations of the behavior of the gas. In particular, insofar as it is possible to extract claims about the macroscopic behavior of the gas from the underlying microphysics, these will themselves show that certain variations in the underlying microstates of the gas are irrelevant to its macroscopic behavior and hence can be absorbed into a macroscopic representation with fewer degrees of freedom. Put differently, the claim that the macroscopic behavior of the gas is “implicit” in the underlying microphysics does not distinguish between two very different possibilities: that (1) the underlying physics shows that variations in those micro-details are relevant to macroscopic behavior in a way that cannot be captured by a few macro-variables and that (2) the underlying physics shows that the macroscopic behavior can be so captured by such macro-variables. In the former case, but not the latter, we have Goldenfeld’s and Kadanoff’s model chaos, in which to model the macroscopic behavior we must model the miro-constituents in full detail. Both possibilities (1) and (2) are consistent with Weslake’s implicitness claim and his claim about what physicalism requires, but have very different implications for the possibility of upper-level explanation that neglects micro-details.

7 Conclusion

I have, in effect, urged that we should ask a different question about upper-level explanations than the question that animates Weslake and Franklin-Hall. Rather than asking, as they do, whether, under interventionist assumptions, abstracting away from epistemic and calculational limitations, upper-level explanations would still be superior to fundamental explanations, a better question is this: what features of the world and what conception of explanation make it possible, given our limitations, to sometimes formulate successful upper-level explanations?