1 Introduction

As is familiar from examples involving banks, airports, and painted mules, changes in conversationally salient error-possibilities, and/or changes in how much is at stake, appear to generate shifts in our judgments regarding the correct application of ‘know’.Footnote 1Contextualists have argued that these judgments motivate the semantic thesis that the contents of ‘knowledge’-ascribing and denying sentences vary with context. Impurists have argued instead that these judgments motivate the metaphysical thesis that knowledge is constitutively tied to practical or conversational factors.Footnote 2 But more conservative theorists have rejected the claim that the shifts in our judgments about ‘know’ warrant our embrace of surprising metaphysical or semantic theses about knowledge (or ‘know’). Among these more conservative theorists are those who suggest that our judgments are an upshot of the fact that shifts in stakes and/or salient error possibilities lead to shifts in the presence of belief (see esp. Bach 2005: §V and Nagel 2008, 2010a, b, 2011).Footnote 3

Such belief-centric (or ‘doxastic’) proposals face a number of familiar challenges and difficulties. Some prominent obstacles include accounting for examples where belief is stipulated to be present and for examples where knowledge is ascribed (or denied) to a subject in a different practical or conversational situation from the speaker. In a series of papers, Nagel (2008, 2010a, b, 2011) has attempted to address these challenges for doxastic proposals. Although Nagel’s proposals have not garnered much serious attention in the literature, her attempt to revive the doxastic approach seems especially worthy of consideration in light of recent criticism of other attempts by conservatively-minded theorists to explain the aforementioned shifts in our judgments about ‘know’ (see e.g. Nagel 2010a: pp. 286–301; Blome-Tillmann 2013; Dimmock and Huvenes 2014; Dinges forthcoming a). It also seems worthy of consideration in light of an interesting and innovative aspect of Nagel’s proposals—her appeal to egocentric bias. As we shall see later on, appeal to egocentric bias would seem to have the potential to address problems both for various epistemological theories, and also for theories in a range of other philosophical arenas, including accounts of epistemic modals, predicates of personal taste, and moral claims. A more careful look at how egocentric bias functions would therefore seem to be of some significant and general philosophical interest.

2 A doxastic approach

Consider the following examples (adapted from Nagel 2010a: p. 287; Cohen 2002: pp. 312–313):

Table A John is in a store looking at what appears to be a bright red table a few yards ahead of him. John’s young son asks John, ‘Do you know that the table is red?’. John replies, ‘Yes, I know it’s red’.

Table B John is in a store looking at what appears to be a bright red table a few yards ahead of him. John’s young son asks John, ‘Do you know that the table is red?’, and remarks that the table would appear just the same to John if it was white but illuminated by red lights. John replies, ‘No, I don’t know that it’s red’.

Assume that in both Table A and Table B the table is indeed red, and the lighting conditions normal. The standard story, applied to our examples, is that John’s knowledge ascription seems true in Table A and his knowledge denial seems true in Table B.Footnote 4Non-sceptical epistemologists concede that John’s knowledge ascription in examples like Table A is true. These theorists therefore have no trouble accommodating the judgment that his knowledge ascription in Table A seems true. But John’s grounds for believing that the table is red seem to be the same in Table A and Table B. It might therefore seem that if John knows in Table A, he knows in Table B. The challenge for non-sceptical epistemologists is therefore to explain why it nevertheless seems true for John to deny that he knows in examples like Table B.

As noted in the Introduction, some prominent non-sceptical epistemologists have argued that our judgments about examples like Table B can be explained if we accept either contextualism or impurism. But more conservative non-sceptical epistemologists—classical invariantists—reject these contentious theories, and maintain that ‘know’ is not semantically context sensitive and that knowledge is not constitutively tied to ‘non-epistemic’ factors, such as salient error possibilities or stakes.Footnote 5 In order to defend their position, classical invariantists therefore require some alternative explanation for our judgments regarding examples like Table B.

One natural classical invariantist suggestion is that the reason John’s knowledge denial seems true in Table B is that considering the possibility that the table is white but illuminated by red lights causes John to lose his belief that the table is red. If that’s right, then—assuming belief is required for knowledge—it follows that John’s utterance of ‘I don’t know that the table is red’ in Table B is in fact true. It is therefore unsurprising that it seems true.

There seem to be three broad mechanisms via which John might lose his belief in Table B:Footnote 6

Lowers credence Some empirical research suggests that considering additional error possibilities leads to lower levels of subjective confidence (Kelley 1972). It might therefore be proposed that considering the possibility of tricky lighting causes John to lower his credence (degree of confidence) in the proposition that the table is red such that his credence is no longer high enough for him to count as believing that it is red (cf. Nagel 2010b: p. 422).

Raises threshold Some theorists have suggested that the level of credence required for belief is determined (in part) by the practical or conversational situation of the subject (Weatherson 2005). It might therefore be proposed that John’s consideration of the possibility of tricky lighting raises the credence threshold John must meet to count as believing that the table is red such that, even if John does not lower his credence upon considering the possibility of tricky lighting, he nevertheless ceases to believe that the table is red.Footnote 7

Other factors Having a belief might not simply be a matter of having a credence above a certain threshold—it might require something else in addition (or instead). It might therefore be proposed that considering the possibility of tricky lighting removes John’s belief that the table is red by impacting that ‘something else’.

For example, it might be proposed that believing that P requires—perhaps in addition to having a certain credence in P—the kind of psychological conviction associated with ’taking it to be settled’ that P or ‘having one’s mind made up’ that P (cf. Nagel 2010b: pp. 416–421). It might then be suggested that in situations where we are considering ways we might be mistaken with respect to P, we often require additional evidence before being psychologically able to (e.g.) take it to be settled that P.Footnote 8 In that case, considering the possibility that the table is white but illuminated by red lights might remove John’s belief that the table is red not because it lowers his credence that the table is red, and not because the credence threshold for belief goes up, but rather because considering that possibility causes John to no longer take it to be settled that the table is red. Call this view removespsychological conviction.Footnote 9

The particular mechanism via which John loses his belief is not central to the discussion to come, so I shall remain neutral on that issue in what follows.Footnote 10 In addition, I shall continue to focus on examples, like Table A and Table B, that concern how shifts in salient error possibilities impact our judgments about the correct application of ‘know’. Doxastic approaches to examples involving shifts in practical factors, such as stakes, meet with parallel problems, and the discussion to follow can be fairly straightforwardly extended to cover such examples. To keep things manageable, I shall therefore largely ignore stakes-based cases in what follows.

3 Challenges

An immediate concern for those sympathetic to the kind of doxastic approach sketched in the previous section is that we exhibit similar judgments about the correct application of ‘know’ even when belief is stipulated to be present. For example, we can imagine a case, Table B*, which is just like Table B except that Table B* contains the additional stipulation that John believes, on the basis of how the table looks, that the table is red. The concern is that even with such an additional stipulation in place, it seems true for John to respond ‘I don’t know that the table is red’ (cf. Nagel 2010a: pp. 287–288, DeRose 2009: pp. 1–2). How is this to be squared with a doxastic approach to explaining the shifts in our judgments about ‘know’?

Nagel (2008, 2010a, b, 2011) puts forward a novel strategy for responding to such ‘stipulated-belief’ cases. In regard to a case similar to Table B*, she writes:

People who are actively thinking about the influence of lighting conditions on colour judgements can still go ahead and make their colour judgements without checking the lighting, but would typically do so only under conditions of compromised or motivated belief formation. But these conditions – haste, distraction, wishful thinking – are the sort of conditions that tend to lower accuracy of judgement. When the accuracy of one’s judgement appears to be compromised, one seems to be a mere believer, rather than a knower. (Nagel 2010a: p. 303)

Roughly, then, the strategy is that if it is stipulated that John believes that the table is red despite actively considering the possibility that it is white but illuminated by red lights (and despite not checking for red lights), he will appear to have formed his belief via the influence of some epistemically problematic factor, such as wishful thinking. The presence of such a factor would seem to us to render John’s belief formation insufficiently accurate (reliable), and so we judge that John lacks knowledge.Footnote 11

A great deal more could be said about this strategy for handling stipulated-belief cases, but for the purposes of the present paper, I shall just grant that it is successful.Footnote 12 In what follows, I wish to focus on a further set of problematic cases for the doxastic approach, concerning situations where the subject of a knowledge claim is in a different conversational or practical situation from the speaker. Various cases illustrate the problem. The following are two representative examples:

Table (3rdPerson) Suppose that John and his son are in a situation like Table B—John’s son has just raised the possibility that the table is white but illuminated by red lights, and John utters ‘I don’t know that the table is red’. But now suppose that John’s wife, Alice, is a few feet away from John, with a similar clear view of the table as him. Alice has clearly not overheard John and her son’s conversation, and is not considering ways she might be mistaken, such as that the table is white but illuminated by red lights. Alice believes, on the basis of how the table looks, that the table is red.

Suppose that after John issues his knowledge denial, John’s son notices his mother looking at the table. John’s son says to John that the table would also appear just the same to his mother if it was white but illuminated by red lights, and asks John if his mother knows that the table is red. John responds ‘No, she has the same evidence as me. She doesn’t know that it’s red either.’

In regard to similar cases, it is standardly reported that John’s response seems true (see e.g. Nagel 2010a: pp. 287–288; Vogel 1990: pp. 15–16; Cohen 1999; DeRose 2009: pp. 3–6).Footnote 13

Table (Modal Contrast) Suppose that John and his son are in a situation like Table B—John’s son has just raised the possibility that the table is white but illuminated by red lights, and John utters ‘I don’t know that the table is red’.

Two shop assistants, Rick and Mona, are standing next to the table, and have overheard John and his son’s conversation. Rick and Mona look up and check that there is no red lighting. Rick then asks Mona if she agrees with John that he doesn’t know that the table is red, reiterating John’s son’s observation that the table would appear just the same to John if it was white but illuminated by red lights. Mona replies, ‘I agree—he doesn’t know the table is red. But if his son hadn’t raised the possibility that the table is white but illuminated by red lights, he would know that it’s red’.

In regard to similar cases, it is standardly reported that Mona’s response seems strange (see e.g. Hawthorne 2004: p. 177n; Nagel 2010b: p. 426; Blome-Tillmann 2009: p. 320).Footnote 14

A doxastic approach, even one supplemented by Nagel’s strategy for handling stipulated-belief cases, seems ill-suited to account for our judgments in these cases. Consider Table (3rdPerson). In that example, John’s son remarks that the table would appear just the same to Alice if it was white but illuminated by red lights, and John utters ‘[Alice] doesn’t know that it’s red’. John’s utterance seems true, but since Alice is not considering error possibilities, the doxastic approach does not seem to supply any obstacle to Alice knowing that the table is red. Indeed, as Alice resembles the John-character in Table A, it seems that advocates of the doxastic approach (as non-sceptics) should accept that Alice knows that the table is red. So why does John’s utterance seem true?

Or consider Table (Modal Contrast). In that example, Rick remarks that the table would seem just the same to John if it was white but illuminated by red lights, and Mona utters ‘[John] doesn’t know that the table is red. But if his son hadn’t mentioned the possibility that the table is white but illuminated by red lights, he would know that it’s red’. Mona’s utterance seems strange, but on the doxastic approach, her utterance is plausibly true. As things actually stand, the possibility of tricky lighting is salient to John. Given the doxastic approach, it should therefore be natural to suppose that John does not believe, and so does not know, that the table is red. However, if John’s son had not mentioned the possibility of tricky lighting, the obstacle to John believing that the table is red would presumably be absent, and John would know that the table is red. So why does Mona’s utterance seem strange?Footnote 15

The general problem underlying cases like Table (3rdPerson) and Table (Modal Contrast) can be usefully stated in terms of the familiar language of ‘ruling out’ error possibilities. Once an error possibility becomes suitably salient, we are prone to judge as though subjects must be able to rule out that possibility in order to be truly said to ‘know’ (cf. Lewis 1996). For example, once the possibility that the table is white but illuminated by red lights becomes suitably salient, we are prone to judge as though John and Alice must be able to rule out that possibility—i.e. possess something like the evidence acquired by explicitly looking up and checking the lighting—in order to be truly said to ‘know’ that the table is red.Footnote 16 A positive aspect of the doxastic approach is that it seems suited to explaining why we judge in this way in regard to subjects, like John in Table B, who are considering the error possibilities at issue: these subjects need to gather additional evidence in order to naturally form the relevant belief (i.e. to form the relevant belief without the influence of epistemically problematic factors, like wishful thinking). However, we also judge as though those subjects who are not considering the error possibilities at issue, such as Alice in Table (3rdPerson), must be able to rule out those possibilities in order to be truly said to ‘know’ (cf. Bach 2005: §V). But at least insofar as the error possibilities at issue are distant or improbable ones—such as the possibility that the table is white but illuminated by red lights—subjects who are not considering those error possibilities presumably do not need to be able to rule them out in order to naturally form beliefs.Footnote 17 Thus, in regard to examples involving these kinds of subjects, a doxastic approach seems ill-suited to explaining our judgments.Footnote 18

4 Egocentric bias

The preceding cases, involving attributions/denials of knowledge to subjects in different conversational situations from the speaker, can seem extremely problematic for advocates of the doxastic approach. But Nagel (2010a: pp. 301–306, 2010b: pp. 425–426) provides an ingenious way to exploit the doxastic approach to explain our judgments even in the examples sketched above—an appeal to egocentric bias.Footnote 19

Empirical research indicates that we have trouble making accurate judgments about those in more naïve positions than ourselves. In particular, a significant body of research has shown that, when making judgments about those more ignorant than ourselves, we tend to mistakenly treat them as though they share our knowledge (see e.g. Nickerson 1999; Birch and Bloom 2004). This tendency is often termed ‘epistemic egocentrism’ or ‘the curse of knowledge’. Nagel does not propose that a tendency to share our knowledge might explain our judgments in the cases presented in the previous section, but she does suggest that a similar bias could be in play, since the broader problem of epistemic egocentrism concerns not just our knowledge but also our “beliefs, attitudes and concerns” (2010a: p. 302).Footnote 20 Nagel (2010a: pp. 301–306; see also 2010b: pp. 425–426) proposes that our judgments about examples like Table (3rdPerson) and Table (Modal Contrast) might be a reflection of an egocentric tendency to treat others as sharing our concerns about error possibilities. In regard to a case similar to Table (3rdPerson), Nagel writes:

Once concerns about the possibility of tricky lighting have been raised for me, I illegitimately evaluate [the subject’s] predicament as if he shared those concerns. Ordinarily, a person who is actively concerned about the lighting would glance up to check it prior to making a judgement about the colour of the table. People who are actively thinking about the influence of lighting conditions on colour judgements can still go ahead and make their colour judgements without checking the lighting, but would typically do so only under conditions of compromised or motivated belief formation. (Nagel 2010a: p. 303)

By reading the description in Table (3rdPerson), the possibility that the table is white but illuminated by red lights becomes salient to us. Due to egocentric bias, we mistakenly treat Alice as though she is also considering that possibility. John’s knowledge denial to Alice therefore seems true because it seems to us that Alice could only hold the belief that the table is red if she were under the influence of some epistemically problematic factor, such as wishful thinking.Footnote 21

An initial concern. Is it plausible that we would treat Alice as though she is considering the possibility that the table is white but illuminated by red lights even though it is apparent from the case description that she is not considering such error possibilities? An interesting feature of epistemic egocentrism is that it is surprisingly robust. We tend to treat others as sharing our knowledge (at least to some degree) even when it is apparent that they lack this knowledge (see e.g. Fischoff 1975; Camerer et al. 1989; Birch and Bloom 2004). Nagel (2010a: pp. 301–306) proposes that our tendency to treat others as considering the error possibilities that we are considering might exhibit similar robustness—that we exhibit the relevant bias even if it is stipulated that the subject is not considering such possibilities. For the present, let us grant that this response is successful; the issues here will come to the fore in Sects. 46.

A similar egocentric explanation can be offered in response to examples like Table (Modal Contrast). In regard to that case, the proposal would be that when trying to imagine counterfactual situations in which John is not considering the possibility that the table is white but illuminated by red lights, we still mistakenly treat John as though he is considering that possibility. As a result, it seems to us that even if John were not considering the possibility that the table is white but illuminated by red lights, he would still fail to know that the table is red, since he would only be able to form the belief that it is red if he were under the influence of some epistemically problematic factor, like wishful thinking. Thus, Mona’s utterance of ‘He doesn’t know that the table is red. But if his son hadn’t raised the possibility that the table is white but illuminated lights, he would know it’s red’ seems false because her utterance of the conditional seems false (cf. Nagel 2010b: p. 426).Footnote 22

The proposal also seems able to account for the more general observation associated with our problem cases—namely, that when the possibility that the table is white but illuminated by red lights becomes suitably salient, we are prone to treat both subjects who are and subjects who are not considering that possibility as needing to be able to rule out that possibility in order to know. If, due to egocentric bias, we treat subjects who it is apparent are not considering the error possibilities that we are as though they are considering those possibilities, appeal to doxastic effects can presumably explain why we treat those subjects as also needing to rule out the error possibilities we are considering in order to possess knowledge.Footnote 23

Interestingly, appeal to egocentric bias would also seem to have the potential for much broader philosophical application. For one thing, it seems that structurally similar appeals to egocentric bias could be put to work defending rival positions in the debates surrounding the metaphysics/semantics of knowledge. For example, impurist invariantists (commonly called ‘subject-sensitive invariantists’) claim that knowledge is necessarily connected to the subject’s conversational or practical concerns. On such accounts, the ignorance-inducing effects of considering error possibilities, or focusing on particular practical matters, are limited to those subjects who are in fact considering those error possibilities or focusing on those practical matters. As a result, impurist invariantists also face trouble with examples like Table (3rdPerson) and Table (Modal Contrast) (see e.g. Stanley 2005: pp. 98–99 & 106). But drawing on Nagel’s suggestions, impurist invariantists might propose that our judgments regarding such cases are not in fact a threat to their theories: they are merely the upshot of an egocentric tendency to treat other subjects as though they share our conversational and practical concerns.Footnote 24

There is also potential for application in other domains. For example, a lot has been made in the literature on epistemic modals (epistemic uses of expressions like ‘might’ and ‘possible’) of so-called ‘eavesdropper cases’. These are examples in which an eavesdropper, who knows that P, overhears a speaker, who does not know that P, make a claim of the form ‘It might be that not-P’. It is oft-reported that it seems true for the eavesdropper to say ‘That’s false’ in reference to the speaker’s ‘might’-claim (see e.g. Hawthorne 2007). Such cases pose a problem for some standard contextualist accounts of epistemic modals on which epistemic uses of ‘It might be that not-P’ express, roughly, that not-P is compatible with what the speaker knows (or with what the speaker and her intended audience knows). But if the eavesdropper (and us) treat the speaker as sharing our knowledge that P, that could explain why the eavesdropper (and us) judge that it seems correct for the eavesdropper to say ‘That’s false’.

Egocentric bias could also be central to explaining some cases of moral disagreement and of disagreement about personal taste that pose trouble for contextualist theories of moral terms like ‘good’ and ‘wrong’ (see e.g. Khoo and Knobe forthcoming) and of taste expressions like ‘fun’ and ‘disgusting’ (see e.g. MacFarlane 2014: ch. 6). If we treat others as sharing our moral codes and personal tastes, then even supposing that terms like ‘good’ and ‘disgusting’ express different properties when used by different those with different moral codes or personal tastes, we may mistakenly treat others as expressing the same properties that we express when using those terms—and erroneously judge that we are in disagreement as a result.Footnote 25 Given its potentially broad diagnostic potential, then, it seems that a better understanding of egocentric bias would be beneficial.

5 A partial bias

Nagel suggests that our tendency to treat others as considering the error possibilities that we are considering is a facet of epistemic egocentrism (Nagel 2010a: pp. 301–306, 2010b: p. 425). This is important for Nagel’s project because epistemic egocentrism is a surprisingly robust bias. Remarkably, subjects in the relevant empirical studies continue to be biased even if they are told about epistemic egocentrism and its effects on judgment (Fischoff 1975), and even if they are given financial incentives to avoid the bias (Camerer et al. 1989). But although the bias is robust, it is not as strong as our previous discussion (and Nagel’s own discussion) might have seemed to indicate.

In regard to very young children, Birch and Bloom (2004, 2007) suggest that epistemic egocentrism is very powerful. Consider an example familiar from debates in developmental psychology:

Displacement An experimental participant is told a story about ‘Sally’ who places her candy in basket A and then leaves the room. In the story, another character then moves Sally’s candy to basket B. The participant is asked where Sally will first look for her candy when she comes back into the room. Children under four typically respond that Sally will first look in basket B (Wimmer and Perner 1983; Baron-Cohen et al. 1985).

Birch and Bloom (2004, 2007) (see also Birch and Bernstein 2007) suggest that these results be understood in terms of epistemic egocentrism. Birch and Bloom propose that the child responds that Sally will first look for her candy in basket B because the child treats Sally as sharing his knowledge that the candy is in basket B, and that he does this despite being apprised of the information (from the story) that Sally was out of the room when her candy was moved.Footnote 26

This clearly resembles the kind of bias that Nagel alleges is present in examples like Table (3rdPerson) and Table (Modal Contrast). We treat the relevant subject as sharing our concern with the possibility that the table is white but illuminated by red lights, despite being apprised of the information that the subject is not considering such error possibilities. The problem is that epistemic egocentrism does not manifest in this kind of strong way in adults.

A central feature of epistemic egocentrism as it manifests in adults is that it is “a partial bias” (Birch and Bloom 2004: p. 258, Box 2; see also Camerer et al. 1989). We (adults) do not straightforwardly treat others as though they know what we do, at least not when it is apparent that those others do not share our knowledge. Rather, our knowledge impairs our ability to make accurate judgments about those more ignorant than ourselves. A common theme in the literature on understanding egocentric bias is the idea that our judgments about how another person will act or judge are shaped by our thinking about how we would act or judge in that person’s situation (see e.g. Nickerson 1999; cf. Nagel 2010a: p. 302). On this picture, epistemic egocentrism with respect to knowledge emerges because we are unable to fully suppress the effects of our own knowledge when making judgments about those more ignorant than ourselves. The result is various partial errors.

Consider the displacement task. Adults (and older children) do not judge that Sally will first look in basket B. We thus do not treat Sally as though she straightforwardly shares our knowledge of her candy’s location when making a judgment about how Sally will act. But research does suggest that our own knowledge that the candy is in basket B might lead us to commit various partial errors.

For example, one potential partial effect is that our own knowledge that the candy is in basket B will lead us to overestimate the likelihood that Sally will first look in basket B. In this regard, Birch and Bloom (2007) conducted a study that indicates that (adult) subjects who are told that Sally’s candy is removed from basket A and then placed in basket B will judge that it is more likely that Sally will first look in basket B than will those subjects who are told merely that the candy has been removed from basket A and then returned to a basket (but not told which basket).Footnote 27 Such results indicate that the more knowledgeable subjects are biased in their judgments about the likelihood of where Sally will first look for her candy. But any bias in their judgments about how Sally will act is decidedly partial: Birch and Bloom’s study suggests that the more knowledgeable subjects will still assign a low probability to the claim that Sally will first look in basket B, even if it is higher than the probability assigned by the more ignorant subjects.

A couple of prominent examples serves to further demonstrate the partial nature of epistemic egocentrism. An influential study by Baron and Hershey (1988), one Nagel (2010a: pp. 302–303) emphasises when introducing her own proposal, comprised an investigation into the effects of outcome knowledge on our evaluation of various medical and monetary decisions. As part of the study, subjects were given the following medical case:

A 55-year old man had a heart condition. He had to stop working because of chest pain. He enjoyed his work and did not want to stop. His pain also interfered with other things, such as travel and recreation. A type of bypass operation would relieve his pain and increase his life expectancy from age 65 to age 70. However, 8% of the people who have this operation die from the operation itself.

His physician decided to go ahead with the operation. The operation succeeded. Evaluate the physician’s decision to go ahead with the operation. (Baron and Hershey 1988: p. 57)

Participants were also supplied with a case that was the same except that it involved a negative outcome—in that case, it was stated that the operation was unsuccessful and the man died.Footnote 28 The participants were told that the man’s physician had no further information on which to ground her decision to go ahead with the operation other than what is given in the first paragraph of the quoted text, and were asked to evaluate the physician’s decision, “the decision itself, the quality of the thinking that went into it”, on the following scale: 3—clearly correct, and the opposite decision would be inexcusable; 2—correct, all-things-considered; 1—correct, but the opposite would be reasonable too; 0—the decision and its opposite are equally good; \(-1\)—incorrect, but not inexcusable; \(-2\)—incorrect, all-things-considered; \(-3\)—incorrect and inexcusable.

If participants had straightforwardly treated the physician as though she shared their knowledge of the actual outcome when making her decision, we would presumably expect a preponderance of 3 s in the case where the operation succeeded, and −3 s in the case where the man died. The reason is simple: if (e.g.) the physician knew the man was going to die, then her decision to undertake the operation would presumably be incorrect and inexcusable. But this is not what happened. Instead, what was found was a tendency to evaluate the physician’s decision as slightly higher up the aforementioned scale in the case where the operation succeeded than in the case where the man died. (Baron and Hershey calculated the mean decision evaluation in the positive outcome (operation succeeded) case and then subtracted it from the mean decision evaluation in the negative outcome (patient died) case. They report that, across a range of similar positive/negative-outcome case pairs, the mean difference was 0.7—i.e. a positive outcome as opposed to a negative outcome produced, on average, an increase of slightly more than half an increment on the ranking scale described above.) This suggests that our evaluation of decisions is indeed biased by our knowledge of the relevant outcome, but that once again, the effect is partial. We do not straightforwardly treat the physician as though she shares our knowledge of the outcome.

Consider a final example. As part of a well-known study by Fischoff (1975), a group of participants (‘before-subjects’) were given some information concerning a particular event (e.g. a conflict between British and Gurka forces in the 1800 s), and asked to estimate the likelihood of each of a list of four possible outcomes, given that information. Another group of participants were given the same information, but were also told that a particular outcome in fact came to pass (e.g. that the Gurka forces triumphed). This second group of participants (‘after-subjects’) were asked to judge the likelihood of the four possible outcomes ‘as they would have, had they not known what happened’. After-subjects gave estimates of how likely they would have judged the reported outcome to be that were significantly higher than the estimates produced for that outcome by before-subjects. Nevertheless, when asked to judge ‘as they would have, had they not known what happened’, after-subjects did not treat their counterfactually-imagined more ignorant selves as though they knew the relevant outcome. If they had, they would presumably have issued probabilities close to 100% for the outcome that they were told came to pass.Footnote 29 Instead, after-subjects’ estimates of how likely they would have judged that outcome to be were merely inflated. (Across a range of historical events, Fischoff reports that the mean increase in probability estimate for the reported outcome among after-subjects vs. before-subjects was 9.2%.)Footnote 30

The upshot of such studies is that although we are often ‘cursed’ by our own knowledge when making judgments about others, we do not straightforwardly treat others as though they know what we know, at least not in situations where it is apparent that they do not possess that knowledge. The claim that we treat such subjects as ‘sharing’ our knowledge is thus a rather misleading oversimplification (cf. Birch and Bloom 2004: p. 258, Box 2). What really happens is that our own privileged knowledge leads us to make various partial judgmental errors—errors in judgment about how others will judge and act, or about the appropriateness of their so doing, that are plausibly the result of a failure to fully suppress our own knowledge. A more accurate characterisation of the phenomenon is thus that we treat those who it is apparent are more ignorant than ourselves as though they share our knowledge to some degree.Footnote 31 Or drawing on the familiar language of ‘epistemic position’ (DeRose 1995), it seems that our own knowledge leads us to treat more ignorant others as though they occupy a better epistemic position (one more like our own) than they really do. Consider the displacement task. We do not straightforwardly treat Sally as sharing our knowledge that her candy is in basket B. But insofar as we overestimate the likelihood that Sally will first look in basket B, we do seem to be treating Sally as though her epistemic position with respect to the proposition that her candy in in basket B is better than it really is.

6 Partial bias and salient error possibilities

Let’s turn now to consider the other facet of epistemic egocentrism (broadly understood) that Nagel proposes accounts for our judgments in the problem cases encountered in Sect. 2—that we will treat other subjects as though they are considering the error possibilities that we are. The preceding reflections on how epistemic egocentrism manifests with respect to knowledge suggest that it is implausible that we will straightforwardly treat others as though they are considering the same unlikely error possibilities that we are considering, at least not when it is apparent that they are not considering those possibilities.Footnote 32

Consider Alice in Table (3rdPerson). It is relatively uncommon for a subject to be considering unlikely error possibilities, and there is no apparent reason to think that Alice would be inclined to do so. It is also explicitly stipulated in the case description that Alice is not considering error possibilities (ways she might be mistaken), such as the possibility that the table is white but illuminated by red lights. The preceding reflections on epistemic egocentrism thus suggest that we will not treat Alice as though she is straightforwardly considering that error possibility. (At least not so long as we are over 4 years old.) Instead, we will exhibit (at most) partial bias. Mirroring how epistemic egocentrism manifests with respect to knowledge, we might expect that we will treat Alice as though she shares our concerns with error possibilities to some degree. (We will explore how exactly this partial bias might manifest shortly.)

Similar remarks apply to Table (Modal Contrast). In regard to this case, the Fischoff (1975) study provides a useful comparison. Just as we do not treat our counterfactually-imagined selves as straightforwardly knowing the outcome when asked to judge ‘as we would have, had we not known the outcome’, so it seems implausible to allege that when asked to consider Mona’s utterance of ‘If his son had not mentioned the possibility that the table is white but illuminated be red lights, [John] would know that it’s red’, we will straightforwardly treat ‘counterfactual-John’ as though he is considering that error possibility. Once again, the most it seems plausible to expect is partial bias: that we will treat counterfactual-John as though he shares our concerns with error possibilities to some degree.

At this point, those sympathetic to the appeal to egocentric bias might insist that the original proposal was merely an oversimplification—one, perhaps not coincidentally, that is often found in the relevant psychological literature—and that partial bias effects are sufficient to explain our judgments. In the remaining discussion, I shall argue on various grounds that this is not the case, and that Nagel’s proposed defence of the doxastic approach is consequently unsuccessful.

On initial inspection, it might seem that partial egocentric effects are suited to explaining our judgments. Consider Table (3rdPerson). As noted above, our reflections on epistemic egocentrism suggest that we will not treat Alice as though she is straightforwardly considering the possibility that the table is white but illuminated by red lights. But we may nevertheless treat Alice as though she has some greater degree of concern with, or awareness of, error possibilities than she really does. Given the proposals put forward in Sect. 1, this seems liable to have an important impact on our assessment of Alice’s doxastic condition.

For example, recall lowers credence, the view that the consideration of error possibilities tends to result in lower levels of subjective confidence (Sect. 1). Given lowers credence, if we treat Alice as though she has greater awareness of error possibilities than she really does, we may judge that it would be natural for her to exhibit a lower credence in the proposition that the table is red than would in fact be natural for her to exhibit. As a result, we may be prone to mistakenly treat her as though she requires more evidence than she really does in order to naturally maintain a credence above the threshold required for belief. Or consider removes psychological conviction (Sect. 1), the view that the consideration of error possibilities impacts how much evidence a subject requires in order to take the question at issue ‘to be settled’. Given removes psychological conviction, if we treat Alice as though she has a greater awareness of error possibilities than she really does, we may be prone to mistakenly treat Alice as requiring more evidence than she really does in order to naturally take it to be settled (and so believe) that the table is red.

More broadly, then, it seems that an egocentric tendency to treat Alice as though she shares our concern about error possibilities to some degree may still lead us to treat Alice as requiring more evidence (or a stronger epistemic position) than she really does in order to naturally form the belief that the table is red. Perhaps this could explain our judgments regarding Table (3rdPerson). If we overestimate the evidence that Alice requires in order to naturally form the belief that the table is red, we may mistakenly judge that Alice would not naturally form the belief that the table is red on the evidence she possesses. As a result, it may seem to us that Alice would only form that belief if she were under the influence of some epistemically problematic factor, like wishful thinking. This could explain why John’s utterance of ‘[Alice] doesn’t know that it’s red either’ seems true. Similar remarks apply to Table (Modal Contrast).Footnote 33

7 Undermining the proposal

There are two broad problems with the kind of appeal to partial bias effects sketched at the close of the previous section. The first problem is that there is plausibly an egocentric bias working in the opposing direction to the one Nagel postulates, and that may well cancel out the bias effect that Nagel is proposing.

It is a standard feature of examples like our various ‘Table’ cases to stipulate that the relevant proposition is indeed true—e.g. to stipulate that the table is indeed red (as was done in Sect. 1). But we are being asked to make judgments about subjects who have not been supplied with this assurance. As our earlier review of the literature on epistemic egocentrism indicated, being explicitly assured of a particular outcome—e.g. that the table is red, that the patient died, that the Gurkas triumphed in their conflict with the British—is itself plausibly going to lead to partial egocentric bias when making judgments about subjects who lack that assurance. In the general terms introduced above, the relevant research would seem to indicate that our own ‘God’s-eye-view’ will lead us to treat the subjects in the various Table cases as though their epistemic position with respect to the proposition that the table is red is better (i.e. more like ours) than it really is.

For example, consider Table (3rdPerson). When being asked to evaluate a subject like Alice, we are being asked to envisage a subject who both lacks our God’s-eye-view assurance that the table is indeed red and is not considering the error possibilities that we are considering. The preceding reflections suggest that this situation will result in egocentric bias effects acting in two opposing directions. The first egocentric bias effect—tied to our Gods-eye-view—should, broadly speaking, lead us to treat Alice as though her epistemic position is better than it really is. The second effect—tied to our consideration of the possibility that the table is white but illuminated by red lights—should, broadly speaking, lead us to overestimate the strength of epistemic position required for Alice to naturally form the belief that the table is red. The concern is that the first egocentric effect may simply cancel the second one out.

Is a response to this concern available? Nagel does address a similar concern, crediting it to John MacFarlane (Nagel’s text has been amended to fit the present example):

Objection: to the extent that we overshare our privileged information, why wouldn’t we also project onto [the subject] our knowledge that [the table is red]? When we are evaluating [the subject’s] state of mind with respect to this very proposition, we need to represent her as making some transition from her resources to this target. We may naturally misrepresent these resources, but it is unlikely that we would do so by including among them the stipulation about the target proposition that we have been given in our “God’s eye” view of the case: a stipulation of the truth of the proposition cannot be represented as included in her resources for making a judgment on that very proposition. (Nagel 2010b: p. 425n)

One might question whether we really do represent Alice as making some ‘transition’ from her resources (presumably something like her evidence) to her belief that the table is red when judging whether John’s utterance of ‘She doesn’t know the table is red’ is true. But even setting that issue aside, the more important problem with this response is that it does nothing to challenge the claim that a partial egocentric effect associated with our God’s-eye-view will be present.

Consider the Fischoff (1975) study mentioned above. In that study, ‘after-subjects’ were given a God’s-eye-view assurance of the actual outcome, and asked to judge ‘as they would have, had they not known the outcome’. Paralleling Nagel’s contention in the quoted passage, it is presumably unlikely that these after-subjects will treat their counterfactually-imagined more ignorant selves as though the stipulation given in their God’s-eye-view forms part of their ‘resources’. And indeed that is just what Fischoff found. As discussed above, his results did not indicate that after-subjects treat their counterfactually-imagined selves as straightforwardly sharing their God’s-eye-view of the outcome. But Fischoff’s results nevertheless indicated that the after-subjects’ God’s-eye-view gave rise to partial egocentric bias effects. In the present terms, after-subjects’ responses indicated that they were judging as though their counterfactually-imagined more ignorant selves possessed resources (evidence) that were better than they really were.

In a similar vein, then, although it seems reasonable for Nagel to allege that we will not treat Alice as straightforwardly sharing our God’s-eye-view assurance that the table is red, research on epistemic egocentrism nevertheless indicates that our possession of that assurance will give rise to partial egocentric bias. In the present terms, to a tendency to treat Alice as though her resources (evidence) are better than they really are. Nagel’s response in the quoted passage therefore does nothing to undermine the claim that a partial bias associated with our God’s-eye-view will be present. Given that Nagel’s proposed egocentric bias effect centred around the consideration of error possibilities is—as argued above—also at most merely partial, the concern that the two partial effects may simply cancel one another out remains in force.Footnote 34

Let’s now turn to look at a second problem: that even if Nagel’s proposed egocentric effect is not cancelled out, partial bias effects are in any case insufficient to explain our judgments. An initial worry here concerns just how ‘partial’ the relevant biases appear to be. Recall the extent of the egocentric bias in the studies from the ‘curse of knowledge’ literature. In the medical decision case (Baron and Hershey 1988), the subjects told that the patient died, or that the operation succeeded, don’t treat the physician as though she knew what they know—they do not evaluate her decision as a − 3 (incorrect and inexcusable) or a + 3 (correct and the opposite would be inexcusable) on the scale introduced earlier. The effects of being supplied different outcome information, although significant, were much more modest. On average, the difference between knowledge of a positive vs. a negative outcome corresponded to a 0.7 difference on the decision evaluation scale—i.e. a difference of slightly more than 1/2 a point, far less than the 6 point difference between + 3 and − 3. Fischoff’s (1975) study on the effects of outcome knowledge on people’s estimates of how ‘they would have judged, had they not known the outcome’, and Birch and Bloom’s (2007) study on adults’ judgments in displacement tasks, revealed similar egocentric effects: clearly significant, but nothing like the effects that would be expected if we were treating the more ignorant subjects as though they straightforwardly shared our knowledge.

These results raise the concern that any egocentric tendency to treat others as ‘sharing’ our concerns about error possibilities is likely to be similarly modest. In terms of the effect on our evaluation of Alice’s doxastic condition, such results suggest that even if we are prone to overestimate how much evidence Alice requires to naturally form the belief that the table is red, our overestimate is likely to be relatively modest. However, the evidence that Alice in fact possesses—the evidence acquired by virtue of her clear view of the table from a few feet away in apparently normal lighting conditions—is plausibly more than sufficient for a subject (like Alice) who is not considering error possibilities to naturally form the belief that the table is red. Weaker evidence, such as the evidence acquired from perception at a greater distance, or from testimony from a generally reliable source, would presumably be sufficient.Footnote 35 This raises the concern that if our overestimate of the evidence that Alice requires in order to naturally form the belief that the table is red is relatively modest, it may well still strike us that the evidence Alice in fact possesses is sufficient for her to naturally form the belief that the table is red. We should therefore remain sceptical that appeal to partial egocentric bias is sufficient to explain our judgments regarding examples like Table (3rdPerson). (A similar concern arises in regard to examples like Table (Modal Contrast).)

There is however a more pressing problem. Consider John in Table (3rdPerson), a subject who is straightforwardly considering the possibility that the table is white but illuminated by red lights. Plausibly, the minimum additional evidence that strikes us as required for John to be truly said to ‘know’ that the table is red is the minimum evidence sufficient to rule out that error possibility—i.e. something like the evidence acquired by looking up and checking the lighting (see Sect. 2). To account for this, advocates of the doxastic approach should presumably allege that the minimum additional evidence that we take John to require to naturally form the belief that the table is red is the minimum evidence sufficient to rule out the possibility that the table is white but illuminated by red lights (cf. Nagel 2010a: p. 303, 2011: pp. 13–15).

As argued above, an upshot of the partial nature of egocentric bias is that we won’t treat Alice as just like John—we won’t treat her as though she is straightforwardly considering the possibility that the table is white but illuminated by red lights. In terms of the effect on our assessment of Alice’s doxastic condition, this suggests that whatever additional evidence (if any) we take Alice to require to naturally form the belief that the table is red is going to be weaker than whatever additional evidence we take John to require to naturally form that belief. Assuming the minimum additional evidence that we take John to require is the minimum evidence sufficient to rule out the possibility that the table is white but illuminated by red lights, it should therefore seem to us that Alice is able to naturally form the belief that the table is red on weaker evidence than is required to rule out that possibility.

This raises the concern (inter alia) that even if appeal to egocentric bias can explain our judgments regarding the original Table (3rdPerson) case, it is not going to be able to explain judgments regarding a range of cases almost identical to Table (3rdPerson). In particular, appeal to egocentric bias seems ill-suited to explain our judgments regarding a range of cases that are the same except that both John and Alice have better evidence that the table is red—but in which they still do not have the kind of evidence intuitively required to rule out that the table is white but illuminated by red lights. In such cases, John’s son will raise the possibility that the table is white but illuminated by red lights, and it is presumably still going to seem true both for John to utter ‘I don’t know that the table is red’ and for him to go on to claim ‘Alice has the same evidence as me. She doesn’t know either’ (cf. Bach 2005: §V). However, appeal to egocentric bias is not going to be sufficient to explain why we would take Alice to lack knowledge in all of these cases, since, across a range of these cases, Alice presumably ought to strike us as having evidence sufficient to naturally form the belief that the table is red. (Once again, parallel concerns arise regarding Table (Modal Contrast).)

Our reflections on the partial nature of epistemic egocentrism thus bring out a number of pressing concerns. The first main concern is that there will be an egocentric bias acting in the opposing direction to the bias that Nagel postulates, and that may well cancel that bias out. A further set of problems centres on the explanatory adequacy of partial bias effects. There is an initial worry: that due to the decidedly partial nature of egocentric bias, it is far from guaranteed that Nagel’s proposed egocentric effect is going to be sufficient to explain our judgments in the original cases. But there is also a more pressing problem: that even if the effect is suited to explain our judgments in the original cases, there are likely to be structurally similar cases that appeal to egocentric bias cannot explain.Footnote 36

8 Closing remarks

The preceding reflections on egocentric bias suggest that examples involving subjects in different conversational situations from the speaker remain problematic for doxastic approaches. Although it has not been argued for explicitly here, it also seems likely that structurally similar concerns will undermine attempts to appeal to egocentric bias to explain similar problematic cases involving subjects in different practical situations from the speaker (see fn. 22). In the absence of a satisfying alternative explanation of these various cases, the doxastic approach looks to be in trouble.Footnote 37 Given the problems facing other classical invariantist explanations of the shifts in our judgments about ‘know’, this is surely unwelcome news for those seeking to preserve a more conservative epistemological outlook.

Our discussion of egocentric bias also holds some more general philosophical lessons. The partial nature of egocentric bias suggests that its potential for broader diagnostic application may be more limited than indicated in our earlier discussion (Sect. 3). For example, consider the eavesdropper cases mentioned above, which have proven central to the literature on epistemic modals. Recall that in those cases, we judge that an eavesdropper who knows that P speaks truly when she utters ‘That’s false’ in reference to a more ignorant speaker’s utterance of ‘It might be that not-P’. If we straightforwardly treated the speaker (and her intended audience) as sharing our knowledge that P, that could explain why we judge the eavesdropper’s claim to be true, even supposing that what the speaker’s ‘might’ claim expresses is that not-P is compatible with what the speaker (and her intended audience) knows. But our discussion indicates that so long as it is clear from the case description that the speaker and her intended audience do not share our knowledge that P, we will not treat them as though they do share that knowledge. (Instead, we will at most treat the more ignorant speaker (and her intended audience) as though they share our knowledge to some degree, or exhibit a better epistemic position with respect to P than they really do.) This suggests that there will be a broad range of eavesdropper cases that appeal to egocentric bias cannot assist in explaining.

Nevertheless, our reflections should not be taken to imply that egocentric bias is entirely unimportant to philosophical practice and theorising. For one thing, it seems like a kind of biasing effect that philosophers should keep in mind when constructing hypothetical cases and testing them on ordinary speakers. And appeal to egocentric bias may well still have crucial roles to play in explaining our judgments regarding hypothetical cases—we just need to be sure to take account of the partial ways in which the bias manifests when making such explanatory moves.