Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Within logic, abductive reasoning has been studied mainly from a purely syntactic perspective. Definitions of an abductive problem and its solution(s) are given in terms of a theory and a formula, and therefore most of the formal logical work on the subject has focused on:

  1. 1.

    Discussing what a theory and a formula should satisfy in order to constitute an abductive problem, and what a formula should satisfy in order to be an abductive solution [13.1]; see also Chap. 10

  2. 2.

    Proposing algorithms to find abductive solutions [13.2, 13.3, 13.4, 13.5, 13.6]

  3. 3.

    Analyzing the structural properties of abductive consequence relations [13.7, 13.8, 13.9].

In all these studies, which follow the so-called Aliseda–Kakas/Kowalski–Magnani/Meheus (GlossaryTerm

AKM

)-schema of abduction Chap. 10, explanationism and consequentialism are considered, but the epistemic character of abductive reasoning seems to have been pushed into the background. Such character is considered crucial in this chapter, as it will be discussed.

This chapter’s main proposal is an epistemic and dynamic approach to abductive reasoning. The proposal is close to the ideas of [13.10, 13.11, 13.12, 13.13] in that it stresses the key role that agents play within the abductive reasoning scenario; after all, at the heart, abduction deals with agents and their (individual or collective) information. In this sense, this collaboration is closer to the Gabbay–Woods (GlossaryTerm

GW

)-schema [13.14, 13.15], see also Chap. 14, which is based on the concept of ignorance problem that arises when a cognitive agent has a cognitive target that cannot be attained from what she currently knows, and thus highlights the distinctive epistemic feature of abduction that is key to this chapter’s considerations. Even so, this presentation goes one step further, as it fully adopts a dynamic perspective by making explicit the actions involved in the abductive process; after all, abduction studies the way agents react epistemically (as individuals or groupwise) to new observations.

More precisely, this proposal argues (Sect. 13.2) that abductive reasoning can be better understood as a process that involves an agent’s information. To this end, it presents definitions of an abductive problem and an abductive solution in terms of an agent’s knowledge and her beliefs as well as a subjective criteria for selecting the agent’s best explanation, and outlines a policy through which the chosen abductive solution can be integrated into the agent’s information. Then, the discussed ideas and definitions are formalized using tools from dynamic epistemic logic (GlossaryTerm

DEL

). This choice is not accidental: classical epistemic logic (GlossaryTerm

EL

 [13.16, 13.17]) with its possible worlds semantic model is a powerful framework that allows to represent an agent’s knowledge and beliefs not only about propositional facts but also about her own information. Its dynamic extension, DEL [13.18, 13.19], allows the representation of diverse epistemic actions (as diverse forms of announcements and different policies for belief revision) that make such information change. Section 13.3 introduces the needed tools, and then the ideas and definitions discussed in Sect. 13.2 are formalized in Sects. 13.4,  13.513.6, and 13.7. The chapter closes with a brief exploration (Sect. 13.8) of the epistemic and dynamic aspects of abductive reasoning that are brought to light when nonideal agents are considered.

Abductive reasoning The concept of abductive reasoning has been discussed in various fields, and this has led to different ideas of what abduction should consist of (see [13.20], among others). For example, while certain authors claim that there is an abductive problem only when neither the observed χ nor its negation follows from a theory [13.2], others say that there is also an abductive problem when, though χ does not follow, its negation does [13.1], a situation that has been typically called a belief revision problem. There are also several opinions of what an abductive solution is. Most of the work on strategies for finding abductive solutions focuses on formulas that are already part of the system (the aforementioned [13.2, 13.3, 13.4, 13.5, 13.6]), while some others take a broader view, allowing not only changes in the underlying logical consequence relation [13.21] but also the creation and modification of concepts [13.22].

The present proposal focuses on a simple account: Abductive reasoning will be understood as a reasoning process that goes from a single unjustified fact to its abductive explanations, where an explanation is a formula of the system that satisfies certain properties. Still, similar epistemic and dynamic approaches can be made to other interpretations of abduction, as those that involve the creation of new concepts or changes in awareness [13.23, 13.24].

Abductive reasoning in dynamic epistemic logic This contribution is a revised version of a proposal whose different parts have been presented in diverse venues. While Sects. 13.2, 13.4, and 13.6 are based on [13.25], Sects. 13.5 and 13.7 are based on [13.26] and Sect. 13.8 is based on [13.27].

1 Classical Abduction

After Peirce’s formulation of abductive reasoning (see [13.28] and Chap. 10), he immediately adds [13.29, p. 231] that:

“ [The abductive solution] cannot be abductively inferred, or if you prefer the expression, cannot be abductively conjectured, until its entire content is already present in the premises, If [the abductive solution] were true, [the abductive problem] would be a matter of course.”

According to these ideas, abduction is a process that is triggered when a surprising fact is observed by an epistemic agent. Although the process returns an explicative hypothesis, the genuine result of an abductive inference is the plausibility of such hypothesis. The truth of the obtained hypothesis is thereby conjectured as plausible, which makes abduction an inferential process of a nonmonotonic character whose conclusion is rather a provisional proposal that could be revised in the light of new information.

When formalized within logical frameworks, the key concepts in abductive reasoning have traditionally taken the following form (Chap. 10). First, it is said that an abductive problem arises when there is a formula that does not follow from the current theory.

Definition 13.1 Abductive problem

Let Φ and χ be a theory and a formula, respectively, in some language \(\mathcal{L}\). Let \(\vdash\) be a consequence relation on \(\mathcal{L}\):

  • The pair \((\Phi,\chi)\) constitutes a (novel) abductive problem when neither χ nor \(\lnot\chi\) are consequences of Φ, that is, when

    $$\Phi\ \not\vdash\ \chi\qquad\text{and}\qquad\Phi\ \not\vdash\ \lnot\chi\;.$$
  • The pair \((\Phi,\chi)\) constitutes an anomalous abductive problem when, though χ is not a consequence of Φ, \(\lnot\chi\) is, that is, when

    $$\Phi\ \not\vdash\ \chi\qquad\text{and}\qquad\Phi\ \vdash\ \lnot\chi\;.$$

It is typically assumed that the theory Φ is a set of formulas closed under logical consequence, and that \(\vdash\) is a truth-preserving consequence relation.

Consider a novel abductive problem. The observation of a χ about which the theory Φ does not have any opinion shows that Φ is incomplete. Further information that completes Φ making χ a consequence of it solves the problem, as now the theory is strong enough to explain χ. Consider now an anomalous abductive problem. The observation of a χ whose negation is entailed by the theory shows that the theory contains a mistake. Now two steps are needed. First, perform a theory revision that stops \(\lnot\chi\) from being a consequence of Φ; this turns the anomalous problem into a novel one, and now the search for further information that completes the theory, making χ a consequence of it, can be performed. Here are the formal definitions.

Definition 13.2 Abductive solution

  • Given a novel abductive problem \((\Phi,\chi)\), the formula η is said to be an abductive solution when

    $$\Phi\cup\{\eta\}\ \vdash\ \chi\;.$$
  • Given an anomalous abductive problem \((\Phi,\chi)\), the formula η is an abductive solution when it is possible to perform a theory revision to get a novel problem \((\Phi^{\prime},\chi)\) for which η is a solution.

This definition of an abductive solution is often considered as too weak: η can take many trivial forms, as anything that contradicts Φ (then everything, including χ, follows from \(\Phi\cup\{\eta\}\)) and even χ itself (clearly, \(\Phi\cup\{\chi\}\vdash\chi\)). Further conditions can be imposed in order to define more satisfactory solutions; here are some of them [13.1] (Chap. 10).

Definition 13.3 Classification of abductive solutions

Let \((\Phi,\chi)\) be an abductive problem. An abductive solution η is

$$\begin{array}[]{lll}\text{consistent}&\text{ iff }&\Phi,\eta\ \not\vdash\ \bot\\ \text{explanatory}&\text{ iff }&\eta\ \not\vdash\ \chi\\ \text{minimal}&\text{ iff }&\text{for every other solution }\zeta,\\ &&\eta\vdash\zeta\text{ implies }\zeta\ \vdash\ \eta\end{array}$$

The consistency requirement discards solutions that are inconsistent with the theory, something a reasonable explanation should not do. In a similar way, the explanatory requirement discards those explanations that would justify the problem by themselves, since it is preferred that the explanation only complements the current theory. Finally, the minimality requirement works as Occam’s razor, looking for the simplest explanation: A solution is minimal when it is in fact logically equivalent to any other solution it implies. For further details on these definitions, the reader is referred to Chap. 10.

2 A Dynamic Epistemic Perspective

The present contribution proposes an approach to abductive reasoning from an epistemic and dynamic perspective. Instead of understanding abductive reasoning as a process that modifies a theory whenever there is a formula that is not entailed by the theory under some particular consequence relation, as the traditional definition of an abductive problem does, the proposed approach understands abductive reasoning as a process that changes an agent’s information whenever, due to some epistemic action, the agent has come to know or believe a fact that she could not have predicted otherwise.

Such an epistemic and dynamic approach is natural. First, abduction, as other forms of nonmonotonic reasoning (e. g., belief revision, default reasoning), is classified as form of a common-sense reasoning rather than a mathematical one, and most of its classic examples involve real agents and their information (e. g., Mary observes that the light does not go on; Karen observes that the lawn is wet; Holmes observes that Mr. Wilson’s right cuff is very shiny). Thus, even though abductive reasoning has been linked to scientific theories (as interpreted in philosophy of science), in its most basic forms it deals with an agent’s (or a set of agents’) information. Second, abductive reasoning implies a change in the agent’s information (Mary assumes that the electricity supply has failed; Karen assumes it has rained; Holmes assumes Mr. Wilson has done a lot of writing lately), and thus it is essential to distinguish the different stages during the abductive process: the stage before the observation, the stage after the observation has raised the abductive problem (and thus the one when the agent starts looking for an explanation), and the stage in which the explanation that has been chosen is incorporated into the agent’s information. This describes, of course, a dynamic process.

There is a final issue that is crucial for an epistemic approach to abductive reasoning. From this contribution’s perspective, abductive reasoning involves not one epistemic attitude (as is typically assumed in most approaches) but rather (at least) two: that of those propositions about which the agent has full certainty; and that of those propositions that she considers very likely but she still cannot be certain about. The reason is that an agent typically tries to explain facts she has come to know due to some observation, but the chosen solution, being a hypothesis that might be dropped in the light of further observations, should not attain the full certainty status. The use of different epistemic notions also gives more flexibility to deal with a wider variety of abductive problems and abductive solutions, making the analysis closer, from the authors’ perspective, to Peirce’s original formulation.

All in all, the abductive process can be studied by asking four questions:

  1. 1.

    What is an abductive problem?

  2. 2.

    What is an abductive solution?

  3. 3.

    How is the best solution(s) selected?

  4. 4.

    How does the agent assimilate the chosen solution(s)?

In the following, answers to these questions are discussed.

2.1 What Is an Abductive Problem?

There are, from an epistemic and dynamic perspective, two important concepts in the definition of an abductive problem. The first is what a formula χ should satisfy in order to become an abductive problem. The second is the action that triggers the abductive problem, that is, the action that turns a formula χ into an abductive problem.

For the former concept, a formula is typically said to be an abductive problem when it is surprising. There are different ways to define a surprising observation of χ (some of them in a DEL setting [13.30]). Most of the approaches that define this notion in terms of what the agent knows (believes) understand a surprise as something that does not follow from such knowledge (beliefs). In other words, it is said that a given χ is surprising whenever the agent does not know (believe) it, or, more radically, whenever the agent knows (believes) \(\lnot\chi\).

Now, note how in the context of abductive reasoning it is not reasonable to define a surprising observation in terms of what the agent knows (believes) after such epistemic action. The reason is that, after observing χ, an agent would typically come to know (believe) it. Thus, if the mentioned definitions are followed focusing on the agent’s information after the observation, no χ would be surprising and there would be no abductive problems at all! It is more reasonable to define a surprising observation not in terms of what the agent knows (believes) as a result of the observation, but rather in terms of what she knew (believed) before it. More precisely, it will be said that a known (believed) χ is surprising with respect to an agent whenever she could not have come to know (believe) it.

Of course, the meaning of the sentence the agent could have come to know (believe) χ still needs to be clarified. This is a crucial notion, as it will indicate not only when a formula χ is an abductive problem (the agent could not have come to know (believe) χ), but also what a formula η needs in order to be an abductive solution (with the help of η, the agent could have come to know (believe) χ). Here the ability to come to know (believe) a given formula will be understood as the ability to infer it, and the simplest way to state this idea is the following: An agent could have come to know (believe) χ if and only if there is an implication \(\eta\rightarrow\chi\) such that the agent knew both the implication and its antecedent. Other formulations that do not use the material implication \(\rightarrow\) are also possible (e. g., the agent may know both \(\lnot\eta\lor\chi\) and η to come to know χ), but in the semantic model this contribution uses (Sect. 13.3), they are logically equivalent to the proposed one.

With respect to the action that triggers an abductive problem χ, this action is typically assumed to be the observation of χ itself. Here a more general idea will be considered: The action that triggers the abductive problem will be simply the observation of some formula ψ. Thus, though ψ should indeed be related to χ (after all, χ is an abductive problem because the agent comes to know χ by observing ψ), the agent will not be restricted to look for explanations of the formula that has been observed: She will also be able to look for explanations of any formula χ she has come to know (believe) through the observation but could not have come to know (believe) by herself before. Note how other actions are also reasonable, as the agent might want to explain a belief she attained after a belief revision (Sect. 13.4.1).

Here is the intuitive definition of an abductive problem in full detail:

“ Let s1 represent the epistemic state of an agent, and let s2 be the epistemic state that results from the agent observing some given ψ. A formula χ constitutes an abductive problem for the agent at s 2 whenever χ is known and there is no implication \(\eta\rightarrow\chi\) such that the agent knew both the implication and its antecedent at s 1 .”

It is important to emphasize how an abductive problem has been defined with respect to an agent and stage (i. e., some epistemic situation). Thus, whether a formula is an abductive problem depends on the formula but also on the information of that given agent at that given stage. The definition is given purely in terms of the agent’s knowledge, but it can also be given purely in terms of her beliefs, or even in terms of both, as it will be seen later.

The presented definition could seem very restrictive. Even if the reader agrees with the basic idea (χ is an abductive problem for a given agent whenever she knows χ but she could not have come to know (believe) it), she/he does not need to agree with the way key parts of it are understood. Nevertheless, as stated in the introduction, this contribution does not intend on providing a full account of the many different understandings of what abductive reasoning does. Rather, its aim is to show how an epistemic and dynamic perspective can shed a new light on the way abductive reasoning is understood, even when assuming its simplest interpretation.

2.2 What Is an Abductive Solution?

In this proposal’s setting, an abductive solution for a given χ will be defined in terms of what the agent could have been able to infer before the observation that raised the problem. As mentioned before, it will be said that η is a solution for the abductive problem χ when the agent could have come to know (believe) χ with the help of η. In this simple case in which the ability to come to know (believe) a given formula is understood as the ability to infer the formula by means of a simple modus ponens step, the following definition is obtained:

“ A formula η constitutes an abductive solution for the abductive problem χ at some given state s2 if the agent knew \(\eta\rightarrow\chi\) at the previous state s1. Thus, the set of solutions for an abductive problem χ is the set of antecedents of implications which have χ as consequent and were known before the observation that triggered the abductive problem.”

Note how abductive solutions are looked for not when the agent has come to know (believe) χ, but rather at the stage immediately before it. Thus, η is a solution when, had it been known (believed) before, would have allowed the agent to come to know (believe) (to predict/expect) χ.

2.3 How is the Best Explanation Selected?

Although there are several notions of explanation for modeling the behavior of why-questions in scientific contexts (e. g., the law model, the statistical relevance model, or the genetic model), most of these consider a consequence (entailment) relation; explanation and consequence go typically hand in hand. However, finding suitable and reasonable criteria for selecting the best explanation has constituted a fundamental problem in abductive reasoning [13.31, 13.32, 13.33], and in fact many authors consider it to be the heart of the subject. Many approaches are based on logical criteria, but beyond requisites to avoid triviality and certain restrictions to the syntactic form, the definition of suitable criteria is still an open problem. Some approaches have suggested the use of contextual aspects, such as an ordering among formulas or among full theories. In particular, for the latter, a typical option is the use of preferential models based on qualitative properties that are beyond the pure causal or deductive relationship between the abductive problem and its abductive solution. However, these preference criteria are seen as an external device which works on top of the deductive part of the explanatory mechanism, and as such they have been criticized because they seem to fall outside the logical framework.

Approaching abductive reasoning from an epistemic point of view provides a different perspective. It has been discussed already how the explanation an agent will choose for a given abductive problem does not depend on how the problematic formula could have been predicted, but rather on how the agent could have predicted it. In general, different agents have different information, and thus they might disagree in what each one calls the best explanation (and even in what each one calls explanation at all). This suggests that, instead of looking for criteria to select the best explanation, the goal should be a criterion to select the agent’s best explanation. Now, once the agent has a set of formulas that explain the abductive problem from her point of view, how can she choose the best? This proposal’s answer makes use of the fact that the considered agents have not only knowledge but also beliefs: Among all these explanations, some are more plausible than others from her point of view. These are precisely the ones the agent will choose when trying to explain a surprising observation: The best explanation can be defined in terms of a preference ordering among the agent’s epistemic possibilities. It could be argued that this criterion is not logical in the classic sense because it is not based exclusively on the deductive relationship between the observed fact and different ways in which it could have been derived. Nevertheless, it is logical in a broader sense since it does depend on the agent’s information: her knowledge and, crucially, her beliefs.

2.4 How is the Best Explanation Incorporated Into the Agent’s Information?

Once the best explanation has been selected, it has to be incorporated into the agent’s information. One of the features that distinguishes abductive reasoning from deductive reasoning is its nonmonotonic nature: The chosen explanation does not need to be true, and in fact can be discarded in the light of further information. This indicates that an abductive solution cannot be assimilated as knowledge. Nevertheless, an epistemic agent has not only this hard form of information which is not subjected to modifications; she also has a soft form that can be revised as many times as it is needed: beliefs. Therefore, once the best abductive solution η has been chosen, the agent’s information can be changed, leading her to believe that η is the case.

2.5 Abduction in a Picture

It is interesting to notice how the stated definitions of abductive problem and abductive solution rely on some form of counterfactivity, as in Peirce’s original formulation (and also as discussed in [13.15]): A given η is a solution of a problem χ if it would have allowed the agent to predict χ. This can be better described with the following diagram.

figure dfigure d

The upper path is the real one: By means of an observation, the agent goes from the epistemic state s1 to the epistemic state s2 in which she knows χ, and by accepting the abductive solution η she goes further to s3. The existence of this path, the fact that χ is an abductive problem and η is one of its abductive solutions, indicates that, at s1, the lower path would have been possible: Incorporating η to the agent’s information would have taken her to an epistemic state \(s^{\prime}_{2}\) where she would have been able to infer χ. Of course, \(s^{\prime}_{3}\) is not identical to s3: In \(s^{\prime}_{3}\) both η and χ are equally reliable because the second is inferred from the first, but in s3, η is less reliable than χ since although the second is obtained via an observation, the first is just a hypothesis that is subject to revision in the light of further information.

3 Representing Knowledge and Beliefs

As mentioned, the most natural framework for formalizing the discussed ideas is that of DEL, the dynamic extension of epistemic logic. In particular, with the plausibility models of [13.34] it is possible to represent an agent’s knowledge and beliefs as well as acts of observation and belief revision, all of which are crucial to the stated understanding of the abductive process. This section introduces these needed tools; the discussed definitions will be formalized in Sect. 13.4.

3.1 Language and Models

Definition 13.4 Language

Given a set of atomic propositions P, formulas \(\varphi\) of the language \(\mathcal{L}\) are given by

$$\varphi::=p\mid\lnot\varphi\mid\varphi\vee\varphi\mid\langle\leq\rangle\varphi\mid\langle\sim\rangle\varphi\;,$$

where \(p\in\mathrm{P}\). Formulas of the form \(\langle\leq\rangle\varphi\) are read as there is a world at least as plausible (as the current one) where \(\varphi\) holds, and those of the form \(\langle\sim\rangle\varphi\) are read as there is a world epistemically indistinguishable (from the current one) where \(\varphi\) holds. Other Boolean connectives (\(\wedge,\rightarrow,\leftrightarrow\)) as well as the universal modalities, \([\leq]\) and \([\sim]\), are defined as usual (\([\leq]\varphi:=\neg\langle\leq\rangle\neg\varphi\) and \([\sim]\varphi:=\neg\langle\sim\rangle\neg\varphi\) for the latter).

The modalities \(\langle\leq\rangle\) and \(\langle\sim\rangle\), respectively, make it possible to define the notions of belief and knowledge within \(\mathcal{L}\). The language’s semantic model, a plausibility model, is defined as follows.

Definition 13.5 Plausibility model

Let P be a set of atomic propositions. A plausibility model is a tuple \(M=\langle W,\leq,V\rangle\), where:

  1. 1.

    W is a nonempty set of possible worlds

  2. 2.

    \({\leq}\subseteq(W\times W)\) is a locally connected and conversely well-founded preorder, the agent’s plausibility relation, representing the plausibility order of the worlds from her point of view (w ≤ u is read as u is at least as plausible as w)

  3. 3.

    \(V:W\rightarrow{\wp({\mathrm{P}})}\) is an atomic valuation function, indicating the atoms in P that are true at each possible world.

pointed plausibility model \((M,w)\) is a plausibility model with a distinguished world \(w\in W\).

Before proceeding, recall that a relation \(R\subseteq(W\times W)\) is locally connected when every two elements that are R-comparable to a third are also R-comparable. It is conversely well-founded when there is no infinite \(\overline{R}\)-ascending chain of elements in W, where \(\overline{R}\), the strict version of R, is defined as \(\overline{R}wu\) iff Rwu and not Ruw. Finally, it is a preorder when it is reflexive and transitive.

The key idea behind plausibility models is that an agent’s beliefs can be defined as what is true in the most plausible worlds from the agent’s perspective, and modalities for the plausibility relation ≤ will allow this definition to be formed. In order to define the agent’s knowledge, the approach is to assume that two worlds are epistemically indistinguishable for the agent if and only if she considers one of them at least as plausible as the other (if and only if they are comparable via ≤). The epistemic indistinguishability relation ∼ can therefore be defined as the union of ≤ and its converse, that is, as \(\sim\,:=\,\leq\cup\geq\). Thus, ∼ is the symmetric closure of ≤ and hence \(\leq\;\subseteq\;\sim\). Moreover, since ≤ is reflexive and transitive, ∼ is an equivalence relation. This epistemic indistinguishability relation ∼ should not be confused with the equal plausibility relation, denoted by ∼eq, and defined as the intersection of ≤ and ≥, that is, \(\simeq\,:=\,\leq\cap\geq\). For further details and discussion on these models, their requirements and their properties, the reader is referred to [13.34, 13.35].

Example 13.1

The following diagram represents a plausibility model M based on the atomic propositions \({\mathrm{P}}:=\{l,e,b\}\). Circles represent possible worlds (named w1 up to w5), and each one of them includes exactly the atomic propositions that are true at that world (e. g., at w2, the atomic propositions l and e are true, but b is false). Arrows represent the plausibility relation, with transitive arcs omitted (so \(w_{4}\leq w_{5}\leq w_{2}\leq w_{1}\leq w_{3}\), but also \(w_{4}\leq w_{2}\), \(w_{4}\leq w_{1}\), \(w_{4}\leq w_{3}\) and so on). Moreover, ∼ is then the full Cartesian product, that is, for every worlds u and v in the model, u ∼ v.

figure efigure e

For the semantic interpretation, the two modalities \(\langle\leq\rangle\) and \(\langle\sim\rangle\) are interpreted with the help of their respective relations in the standard modal way.

Definition 13.6 Semantic interpretation

Let \((M,w)\) be a pointed plausibility model with \(M=\langle W,\leq,V\rangle\). Atomic propositions and Boolean operators are interpreted as usual. For the remaining cases,

$$\begin{aligned}\displaystyle&\displaystyle(M,w)\Vdash\langle\leq\rangle\varphi\text{ iff }\ \exists u\in W\text{ s.t. }w\leq u\text{ \& }(M,u)\Vdash\varphi\\ \displaystyle&\displaystyle(M,w)\Vdash\langle\sim\rangle\varphi\text{ iff }\ \exists u\in W\text{ s.t. }w\sim u\text{ \& }(M,u)\Vdash\varphi\;.\end{aligned}$$

3.1.1 Defining Knowledge and Beliefs

The notion of knowledge in plausibility models is defined by means of the epistemic indistinguishability relation in the standard way: The agent knows \(\varphi\) at some world w if and only if \(\varphi\) is the case in every world she considers to be epistemically possible from w. (This makes knowledge a very strong notion, corresponding to an “absolutely unrevisable belief” [13.34]). The modality \([\sim]\) can be used to this end. For the notion of beliefs, the idea is, as stated before, that the agent believes \(\varphi\) at a given w if and only if \(\varphi\) is the case in the most plausible worlds from w. Thanks to the properties of the plausibility relation (a locally connected and conversely well-founded preorder), \(\varphi\) is true in the most plausible (i. e., the ≤-maximal) worlds from w if and only if, in accordance with the plausibility order, from some moment onward there are only \(\varphi\)-worlds (see [13.34, 13.36, 13.37] for the technical details). The modalities \(\langle\leq\rangle\) and \([\leq]\) can be used to this end. Summarizing,

$$\begin{array}[]{ll}\text{The agent}\ \textit{knows}\ \varphi&K{\varphi}:=[\sim]\varphi\\ \text{The agent}\ \textit{believes}\ \varphi&B{\varphi}:={\langle\leq\rangle}{[\leq]}\varphi\end{array}$$

Observe how, since \(\leq\subseteq\sim\), the formula \(K{\varphi}\rightarrow B{\varphi}\) is valid (but its converse is not).

The dual of these notions, epistemic possibility and most likely possibility, can be defined as the correspondent modal duals

$$\hat{K}{\varphi}:=\langle\sim\rangle\varphi{\quad\quad}\hat{B}{\varphi}:={[\leq]}{\langle\leq\rangle}\varphi\;.$$

Example 13.2

Consider the plausibility model M of Example 13.1, and take w2 as the evaluation point. Since \(w_{2}\sim u\) holds for every possible world u in the model, every world is epistemically possible from w2’s perspective. But every world in the model satisfies \(b\rightarrow l\) (the implication is true at w2, w1, and w3 because the antecedent b is false, and true at w4 and w5 because the consequent l is true), so \([\sim](b\rightarrow l)\), that is, \(K(b\rightarrow l)\) is true at w2: The agent knows \(b\rightarrow l\) at w 2 . On the other hand, \(\neg l\) is not true in every world, but it is true in w3, the most plausible one from w2’s perspective, so \({\langle\leq\rangle}{[\leq]}\neg l\), that is, \(B{\neg l}\), is true at w2: The agent believes \(\neg l\) at w 2 . Moreover, observe how b is neither known (it is not true in every epistemically indistinguishable world) nor believed (it is not true in the most plausible worlds) at w2. Still, it is true in some epistemic possibilities from w2 (e. g., w5); hence, \(\langle\sim\rangle{b}\) (i. e., \(\hat{K}{b}\)) holds at w2: At that world, the agent considers b possible.

A more detailed description of this framework, a number of the epistemic notions that can be defined within it, its technical details and its axiom system can be found in [13.34].

Following the DEL idea, actions that modify an agent’s information can be represented as operations that transform the underlying semantic model. In the rest of this section, operations that can be applied over plausibility models will be recalled, and extensions of the language that allow to describe the changes such operations bring about will be provided. These will be used in Sect. 13.4 to represent and describe abductive reasoning.

3.2 Operations on Models

3.2.1 Update, Also Known as Observation

The most natural operation over Kripke-like semantic models is that of update. This operation reduces the domain of the model, and is typically given in terms of the formula the worlds should satisfy in order to survive the operation.

Definition 13.7 Update operation

Let the tuple \(M=\langle W,\leq,V\rangle\) be a plausibility model and let ψ be a formula in \(\mathcal{L}\). The update operation yields the plausibility model \(M_{{\psi}{!}}=\langle W^{\prime},\leq^{\prime},V^{\prime}\rangle\) where \(W^{\prime}:=\{w\in W|(M,w)\Vdash\psi\}\), \(\leq^{\prime}:={\leq}\cap{(W^{\prime}\times W^{\prime})}\) and, for every \(w\in W^{\prime}\), \(V^{\prime}(w):=V(w)\).

This operation reduces the domain of the model (preserving only those worlds that satisfy the given ψ) and restricts the plausibility relation and the atomic valuation function accordingly. Since a submodel is obtained, the operation preserves the (universal) properties of the plausibility relation and hence it preserves plausibility models: If M is a plausibility model, then so is \(M_{{\psi}{!}}\).

In order to describe the effects of an update within the language, existential modalities of the form \(\langle\psi!\rangle\) are used, for every formula ψ. Here is their semantic interpretation

$$\begin{aligned}\displaystyle&\displaystyle(M,w)\Vdash{\langle{\psi}!\rangle\varphi}\text{ iff }(M,w)\Vdash\psi\\ \displaystyle&\displaystyle\quad\text{ and }(M_{{\psi}{!}},w)\Vdash\varphi\;.\end{aligned}$$

In words, an update formula \({\langle{\psi}!\rangle}\varphi\) holds at \((M,w)\) if and only if ψ is the case (i. e., the evaluation point will survive the operation) and, after the update, \(\varphi\) is the case. The universal modality \([{\psi}!]\) is defined as the modal dual of \(\langle{\psi}!\rangle\), that is, \([{\psi}!]{\varphi}:=\neg\langle{\psi}!\rangle\neg\varphi\).

In addition to being the most natural operation over Kripke-like models, an update also has a straightforward epistemic interpretation: it works as an act of a public announcement [13.38, 13.39] or, as it will be called here, an act of observation. When the agent observes a given ψ, she can discard those epistemically possible worlds that fail to satisfy this formula, thereby obtaining a model with only worlds that satisfied ψ before the operation. More details on this operation and its modalities (including an axiom system) can be found in the papers [13.38, 13.39] or in the textbooks [13.18, 13.19].

Example 13.3

Consider the model M in Example 13.1 again. Suppose the agent observes l; this can be modeled as an update with l, which yields the following model \(M_{{l}!}\)

figure ffigure f

The most plausible world in M has been discarded in \(M_{l!}\). As explained in Example 13.2, the agent believes \(\neg l\) in M, but after the observation this is not the case anymore: \(\neg l\) does not hold in the unique most plausible world of the new model \(M_{l!}\). In fact, \(\neg l\) does not hold in any epistemically possible world, and thus after the observation the agent knows l; in symbols

$$(M_{l!},w_{2})\Vdash K{l},\text{ that is, }(M,w_{2})\Vdash[l!]K{l}\;.$$

3.2.2 Upgrade, Also Known as Belief Revision

Another natural operation over plausibility-like models is the rearrangement of worlds within an epistemic partition. Of course, there are several ways in which a new order can be defined. The following rearrangement, taken from [13.40], is one of the many possibilities.

Definition 13.8 Upgrade operation

Let the tuple \(M=\langle W,\leq,V\rangle\) be a plausibility model and let ψ be a formula in \(\mathcal{L}\). The upgrade operation produces the plausibility model \(M_{{\psi}\Uparrow}=\langle W,\leq^{\prime},V\rangle\), which differs from M just in the plausibility order, given now by

$$\begin{aligned}\displaystyle\leq^{\prime}:=&\displaystyle\{(w,u)|w\leq u\text{ and }(M,u)\Vdash\psi\}\cup\\ \displaystyle&\displaystyle\{(w,u)|w\leq u\text{ and }(M,w)\Vdash\neg\psi\}\cup\\ \displaystyle&\displaystyle\{(w,u)|w\sim u,(M,w)\Vdash\neg\psi\\ \displaystyle&\displaystyle\quad\text{ and }(M,u)\Vdash\psi\}\;.\end{aligned}$$

The new plausibility relation states that after an upgrade with ψ, all ψ-worlds become more plausible than all \(\neg\psi\)-worlds, and within the two zones the old ordering remains [13.40]. More precisely, a world u will be at least as plausible as a world w, \(w\leq^{\prime}u\), if and only if they already are of that order and u satisfies ψ, or they already are of that order and w satisfies \(\neg\psi\), or they are comparable, w satisfies \(\neg\psi\) and u satisfies ψ. This operation preserves the properties of the plausibility relation and hence preserves plausibility models, as shown in [13.35].

In order to describe effects of this operation within the language, an existential modality \(\langle\psi\Uparrow\rangle\) is introduced for every formula ψ,

$$(M,w)\Vdash{\langle\psi\Uparrow\rangle}\varphi\ \text{ iff }\ (M_{{\psi}\Uparrow,w})\Vdash\varphi\;.$$

In words, an upgrade formula \(\langle{\psi}\Uparrow\rangle{\varphi}\) holds at \((M,w)\) if and only if \(\varphi\) is the case after an upgrade with ψ. The universal modality \([{\psi}\Uparrow]\) is defined as the modal dual of \(\langle{\psi}\Uparrow\rangle\), as in the update case.

This operation also has a very natural epistemic interpretation. The plausibility relation defines the agent’s beliefs, and hence any changes in the relation can be interpreted as changes in the agent’s beliefs [13.34, 13.40, 13.41]. In particular, an act of revising beliefs after a reliable and yet fallible source has suggested ψ can be represented by an operation that puts ψ-worlds at the top of the plausibility order. Moreover, each one of the different methods to obtain a relation with former ψ-worlds at the top can be seen as a different policy for revising beliefs. Details on the operation and its modalities (including an axiom system) can be found in the papers [13.34, 13.40] or in the textbook [13.19].

Example 13.4

Consider the model \(M_{l!}\) in Example 13.3, that is, the model that results from the agent observing l at the initial model M in Example 13.1. Suppose the agent performs a belief revision toward b; this can be modeled as an upgrade with b, which yields the following model \((M_{l!})_{b\Uparrow}\):

figure gfigure g

The ordering of the worlds has changed, making those worlds that satisfy b (w4 and w5) more plausible than those that do not (w2 and w1), keeping the old ordering with these two zones (w5 strictly above w4 and w1 strictly above w2). In \(M_{l!}\) the agent believed \(\neg b\wedge\neg e\), as such formula was the case in the model’s unique most plausible world w1, but this is not the case anymore in \((M_{l!})_{b\Uparrow}\): The unique most plausible world, w5, satisfies \(b\wedge e\), and thus the formula is part of the agent’s beliefs. In symbols,

$$\begin{aligned}\displaystyle&\displaystyle((M_{l!})_{b\Uparrow},w_{2})\Vdash B(b\wedge e),\\ \displaystyle&\displaystyle\text{that is, }(M_{l!},w_{2})\Vdash[l\Uparrow]B{(b\wedge e)}\;.\end{aligned}$$

4 Abductive Problem and Solution

Given the intuitive definitions discussed and the formal tools introduced, it is now time to formalize the ideas.

4.1 Abductive Problem

First, the definition of what an abductive problem is.

Definition 13.9 Abductive problem

Let \((M,w)\) be a pointed plausibility model, and consider \((M_{\psi!},w)\), the pointed plausibility model that results from observing a given ψ at \((M,w)\).

A formula χ is an abductive problem at \((M_{\psi!},w)\) if and only if it is known at such stage but it was not known before, that is, if and only if

$$(M_{\psi!},w)\Vdash K{\chi}\quad\text{ and }\quad(M,w)\Vdash\neg K{\chi}\;.$$

Equivalently, a formula χ can become an abductive problem at \((M,w)\) if and only if it is not known at such stage but will be known after observing ψ, that is, if and only if

$$(M,w)\Vdash\neg K{\chi}\wedge[{\psi!}]K{\chi}\;.$$

Note again how the definition of an abductive problem is relative to an agent’s information at some given stage (the one represented by the pointed model).

There are two points worth emphasizing. First, note again how the definition distinguishes between the formula that becomes the abductive problem, χ, and the formula whose observation triggers the abductive problem, ψ. Although these two formulas are typically understood to be the same (χ becomes an abductive problem after being observed), the choice in this contribution is to distinguish between them. One reason for this is technical: Here the idea is that the agent will look for explanations of formulas that she could not have known before the observation but knows afterward. However, stating this as the agent knows χ after observing it is restrictive in the DEL setting as not every formula satisfies this condition. This is because the underlying EL framework is powerful enough to talk about the knowledge an agent has not only about facts but also about her own knowledge, and so there are formulas expressing situations such as it is raining and you do not know it (\(r\wedge\neg Kr\)), which can be observed but are not known afterward (now you know that it is raining!). Another reason is, as stated earlier, generality: The described agent will be able to look for explanations not only of the formulas she can observe, but also of those that she can come to know through an observation. Still, this choice does not imply that the observed formula and the one that becomes an abductive problem are unrelated: In order for the agent to know χ after observing ψ, she must have known \(\psi\rightarrow[{\psi!}]\chi\) before the action. This is nothing but the reduction axiom for the knowledge modality in public announcement logic

$$[{\psi!}]K{\chi}\leftrightarrow(\psi\rightarrow K(\psi\rightarrow[{\psi!}]{\chi}))\;.$$

Second, the requirements Definition 13.9 asks for χ to be an abductive problem are not exactly the ones stated in the previous section: The sentence there is no implication \(\eta\rightarrow\chi\) such that, before χ became an abductive problem, the agent knew both the implication and its antecedent has been replaced by the agent did not know χ before χ became an abductive problem. The reason is that, in DEL, the agent’s knowledge and beliefs are closed under logical consequence (still, small variations of the EL framework allows the representation of nonideal agents and their abductive reasoning; see Sect. 13.8), and in such setting the two statements are equivalent: If there is an η such that the agent knew \(\eta\rightarrow\chi\) and η before χ became an abductive problem, then clearly she knew χ too, and if she knew χ, then there was a η such that \(\eta\rightarrow\chi\) and η were both known, namely χ itself. In fact, the restatement of the requirement emphasizes that it is the observation of ψ what causes the agent to know χ and hence what creates the abductive problem.

It is worthwhile to highlight how, although the definition of an abductive problem was given in terms of the agent’s knowledge, it can also be given in terms of her beliefs: It also makes sense for her to look for explanations of what she has come to believe!

“ A formula χ is said to be an abductive problem at \((M_{\psi\Uparrow},w)\) if and only if \((M_{\psi\Uparrow},w)\Vdash B\chi\) and \((M,w)\Vdash\neg B\chi\).”

With this definition, a formula is a problem if it is believed now but was not believe before a belief revision with ψ. But not only that. The agent can also face abductive problems that combine knowledge and beliefs. For example, she can face an abductive problem with χ if she does not know the formula at some stage but believes it after a belief revision with ψ:

“ A formula χ is said to be an abductive problem at \((M_{\psi\Uparrow},w)\) if and only if \((M_{\psi\Uparrow},w)\Vdash B\chi\) and \((M,w)\Vdash\neg K{\chi}\).”

The stated definition allows to describe several forms of abductive problems, all of which differ in the strength of the attachment of the agent to the problematic χ (known, strongly believed, safely believed, believed, etc.) after the epistemic action (update, upgrade) and the strength of her attachment to the formula before the action.

4.2 Classifying Problems

As mentioned, some approaches classify an abductive problem χ according to whether χ or \(\neg\chi\) follows from the theory: If neither χ nor \(\neg\chi\) follows, then χ is called a novel abductive problem; if χ does not follow but \(\neg\chi\) does, then χ is called an anomalous abductive problem. Given the requirement the agent did not know χ before χ became an abductive problem (\(\neg K{\chi}\)) in Definition 13.9, one could suggest the agent knew \(\neg\chi\) (\(K{\neg\chi}\)) as an alternative, but since the definition also asks for χ to be known after the observation in order to be an abductive problem, such suggestion turns out to be too strong for propositional formulas: If \(\neg\chi\) is propositional and the agent knows it at some stage, then every epistemic possibility satisfies \(\neg\chi\). Thus, since no epistemic action can change the (propositional) formula’s truth value, the only way for the agent to know χ afterward is for the action to eliminate every epistemic possibility, making \(K{\varphi}\) true for every formula \(\varphi\) and thus turning the agent inconsistent. But even though it is not possible to classify abductive problems in terms of the knowledge the agent had about the formula before the observation, it is still possible (and more reasonable) to classify them by using weaker notions, such as beliefs. Here is one possibility.

Definition 13.10 Expected, novel and anomalous problems

Suppose χ is an abductive problem at \((M_{\psi!},w)\). Then χ is said to be:

  • An expected abductive problem if and only if \((M,w)\Vdash B\chi\)

  • An novel abductive problem if and only if \((M,w)\Vdash\neg B\chi\wedge\neg B{\neg\chi}\)

  • An anomalous abductive problem if and only if \((M,w)\Vdash B{\neg\chi}\).

Many people would not call the first case an abductive problem: The observation is a confirmation rather than a surprise, and thus it does not need to trigger any further epistemic action. Nevertheless, the case shows how this proposal allows for such situations to be considered. In fact, the classification can be refined by considering further attitudes, such as the safe beliefs of [13.34] or the strong beliefs of [13.42] (both definable in \(\mathcal{L}\)).

4.3 Abductive Solutions

An abductive solution is now to be defined. Here is a version that uses only the notion of knowledge.

Definition 13.11 Abductive solution

Let \((M,w)\) be a pointed plausibility model, and consider \((M_{\psi!},w)\), the pointed plausibility model that results from observing ψ at \((M,w)\).

If at \((M_{\psi!},w)\) the formula χ is an abductive problem, then η is an abductive solution if and only if the agent knew that η implied χ before the observation, that is, if and only if

$$(M,w)\Vdash K{(\eta\rightarrow\chi)}\;.$$

Equivalently, if at \((M,w)\) the formula χ can become an abductive problem, then η will be an abductive solution if and only if the agent knows that η implies χ, that is, if and only if

$$(M,w)\Vdash K(\eta\rightarrow\chi)\;.$$

Just as in the case of abductive problem, it is also possible to define an abductive solution in terms of weaker notions as beliefs. For example, while a very strict agent would accept η as explanation only when \(\eta\rightarrow\chi\) was known, a less strict agent could accept it when such implication was only believed.

It is worth emphasizing that, in the stated definition, a solution for a problem χ (at some \(M_{\psi!}\)) is a formula η such that \(\eta\rightarrow\chi\) is known not when the abductive problem has arisen (at \(M_{\psi!}\)) but rather at the stage immediately before (at M). This is because an explanation is a piece of information that would have allowed the agent to predict the surprising observation. In fact, if an abductive solution for a problem χ were defined as a formula η such that \(\eta\rightarrow\chi\) is known once χ is an abductive problem (at \(M_{\psi!}\)), then every formula \(\varphi\) would be a solution since (at \(M_{\psi!}\)) \(K{\chi}\) would be the case (because χ is an abductive problem) and hence so would be \(K(\varphi\rightarrow\chi)\) for every formula \(\varphi\).

Observe also how, again in the stated definition, if η is a solution for the abductive problem χ (at some \(M_{\psi!}\)), then η could not be known before the observation that triggered the problem (at M). Otherwise, both \(K(\eta\rightarrow\chi)\) and \(K{\eta}\) would be the case at such stage (M) and hence, by the closure under logical consequence of knowledge in EL, so would be \(K{\chi}\), contradicting the fact that χ is an abductive problem.

Proposition 13.1

Let χ be an abductive problem and η be one of its abductive solutions, both at \((M_{\psi!},w)\). Then, \((M,w)\Vdash\neg K{\eta}\).

4.4 Classifying Solutions

It is common in the literature to classify abductive solutions according to their properties (Chap. 10). For example (Definitions 13.2 and 13.3; again, see Chap. 10), given a surprising observation χ, an abductive solution η is said to be:

  • Plain when it is a solution

  • Consistent when it does not contradict the agent’s information

  • Explanatory when it does not explain χ by itself.

Similar properties can be described in the present setting. To begin with, the plain property simply states that η is an abductive solution; a definition that has been already provided (Definition 13.11).

For the consistency property, the intuitive idea is for the solution to be compatible with the agent’s information. To this end, consider the following definition.

Definition 13.12 Consistent solution

Let χ be an abductive problem and η be one of its abductive solutions, both at \((M_{\psi!},w)\). It is said that η is a consistent solution if and only if the agent considers it possible at \((M_{\psi!},w)\), that is, if and only if

$$(M_{\psi!},w)\ \Vdash\ \hat{K}{\eta}\;.$$

Thus, a solution is consistent when it is epistemically possible. Note how this requirement is given in terms of the agent’s information after the epistemic action that triggered the abductive problem, and not before it. In fact, there are formulas that, in a given situation, are solutions according to the stated definition, and yet not epistemically possible once the abductive problem has been raised.

Discussion 13.1

Not every abductive solution is consistent.

Proof:

Let η and χ be propositional formulas, and take a model M in which the agent considers at least one \((\neg\eta\wedge\neg\chi)\)-world to be epistemically possible, with the rest of the epistemic possibilities being \((\neg\eta\wedge\chi)\)-worlds. After observing χ, \(\neg\chi\)-worlds will be discarded and there will be only \((\neg\eta\wedge\chi)\)-worlds left, thus making χ itself an abductive problem (it is not known at M but it will be known at \(M_{\chi!}\)) and η an abductive solution (every epistemic possibility at M satisfies \(\eta\rightarrow\chi\), so the agent knows this implication). Nevertheless, there are no η-worlds at \(M_{\chi!}\), and therefore \(\hat{K}{\eta}\) is false at such stage.

The explanatory property is interesting. The idea in the classic setting is to avoid solutions that imply the problematic χ per se, such as χ itself or any formula logically equivalent to it. In the current epistemic setting, this idea can be understood in a different way: A solution η is explanatory when the acceptance of η (which, as discussed, will be modeled via belief revision; see Sect. 13.6) changes the agent’s information, that is, when the agent’s information is different from \((M_{\psi!},w)\) to \(((M_{\psi!})_{\eta\Uparrow},w)\) (the model that results after integrating the solution η). This assertion could be formalized by stating that the agent’s information is the same in two pointed models if and only if the agent has the same knowledge in both, but this would be insufficient: The model operation representing an act of belief revision (the upgrade of Definition 13.8) is devised to change only the agent’s beliefs (although certain knowledge, such as knowledge about beliefs, might also change). A second attempt would be to state that the agent’s information is the same in two pointed models if and only if they coincide in the agent’s knowledge and beliefs, but the mentioned operation can change a model without changing the agent’s beliefs.

Within the current modal epistemic logic framework, a more natural way of specifying the idea of an agent having the same information in two models is via the notion of bisimulation.

Definition 13.13 Bisimulation

Let P be a set of atomic propositions and let \(M=\langle W\), \(\leq,V\rangle\) and \(M^{\prime}=\langle W^{\prime},\leq^{\prime},V^{\prime}\rangle\) be two plausibility models based on this set. A nonempty relation \(Z\subseteq(W\times W^{\prime})\) is called a bisimulation between M and \(M^{\prime}\) (notation: \(M\underline{\leftrightarrow}_{Z}M^{\prime}\)) if and only if, for every \((w,w^{\prime})\in Z\):

  • \(V(w)=V^{\prime}(w^{\prime})\), that is, w and \(w^{\prime}\) satisfy the same atomic propositions

  • If there is a \(u\in W\) such that w ≤ u, then there is a \(u^{\prime}\in W^{\prime}\) such that \(w^{\prime}\leq^{\prime}u^{\prime}\) and \(Zuu^{\prime}\)

  • If there is a \(u^{\prime}\in W^{\prime}\) such that \(w^{\prime}\leq^{\prime}u^{\prime}\), then there is a \(u\in W\) such that w ≤ u and \(Zuu^{\prime}\).

Two models M and \(M^{\prime}\) are bisimilar (notation: \(M\underline{\leftrightarrow}M^{\prime}\)) when there is a bisimulation between them, and two pointed models \((M,w)\) and \((M^{\prime},w^{\prime})\) are bisimilar (notation: \((M,w)\underline{\leftrightarrow}(M^{\prime},w^{\prime})\)) when there is a bisimulation between M and \(M^{\prime}\) containing the pair \((w,w^{\prime})\).

This notion is significant because, under image-finiteness (a plausibility model is image-finite if and only if every world can ≤-see only a finite number of worlds), it characterizes modal equivalence, that is, it characterises models that satisfy exactly the same formulas in the modal language.

Theorem 13.1

Let P be a set of atomic propositions and let \(M=\langle W\), \(\leq,V\rangle\) and \(M^{\prime}=\langle W^{\prime},\leq^{\prime},V^{\prime}\rangle\) be two image-finite plausibility models. Then \((M,w)\underline{\leftrightarrow}(M^{\prime},w^{\prime})\) if and only if, for every formula \(\varphi\in\mathcal{L}\), \((M,w)\Vdash\varphi\) iff \((M^{\prime},w^{\prime})\Vdash\varphi\).

Now it is possible to state a formal definition of what it means for a solution to be explanatory.

Definition 13.14 Explanatory solution

Let χ be an abductive problem and η be one of its abductive solutions, both at \((M_{\psi!},w)\). It is said that η is an explanatory solution if and only if its acceptance changes the agent’s information, that is, if and only if there is no bisimulation between \((M_{\psi!},w)\) and \(((M_{\psi!})_{\eta\Uparrow},w)\).

This definition, devised in order to avoid solutions that explain the abductive problem per se, has pleasant side effects. In the abductive reasoning literature, a solution is called trivial when it is logically equivalent to the abductive problem χ (i. e., when it is not explanatory) or when it is a contradiction (to the agent’s knowledge, or a logical contradiction). Under the given definition, every trivial solution is not explanatory: Accepting any such solution will not change the agent’s information. The reason is that, in both cases, the upgrade operation will not make any change in the model: In the first case because, after the observation, the agent knows the abductive problem formula, and hence every epistemically possible world satisfies it (as well as every formula logically equivalent to the problem); in the second case because no epistemically possible world satisfies it. In this way, this framework characterizes trivial solutions not in terms of their form, as is typically done, but rather in terms of their effect: Accepting them will not give the agent any new information.

In particular, this shows how the act of incorporating a contradictory explanation will not make the agent collapse and turn into someone that knows and believes everything, as happens in traditional approaches; thus, a logic of formal inconsistency (e. g., [13.43]; see also Chap. 15) is not strictly necessary. This is a consequence of two simple but powerful ideas:

  1. 1.

    Distinguishing an agent’s different epistemic attitudes

  2. 2.

    Assimilating an abductive solution not as knowledge, but rather as a belief.

5 Selecting the Best Explanation

Finding suitable and reasonable criteria for selecting the best explanation is a fundamental problem in abductive reasoning [13.32, 13.33], and in fact many authors consider this to be the heart of the subject. The so-called thesis of purpose, stated in [13.33], establishes that the aim of scientific abduction is:

  1. 1.

    To generate new hypotheses

  2. 2.

    To select hypotheses for further examination and testing.

Hence a central issue in scientific abduction is to provide methods for selecting. Because the true state of the world is unknown, selecting the best explanation requires more than just consistency with the available information, and there are many proposals of what these extra criteria should be.

Some approaches are based on probabilistic measurements [13.44, 13.45, 13.46]. Even Sherlock Holmes advised that, in order to evaluate explanations, one should “balance probabilities and choose the most likely” (The Hound of the Baskervilles), but unfortunately explanations rarely come equipped with probabilities.

In abductive logic programming, a common strategy is to look for abductive solutions at the dead ends of prolog proofs [13.47]. Sound and complete procedures can be defined also by using stable models and answer sets [13.48, 13.49]. Apart from selection criteria based on consistency and integrity constraints, it is common to start with a set of abducible predicates and select explanations built only from ground atoms using them (see Chap. 10 for more details on abductive logic programming).

There are also approaches that use logical criteria, but beyond the already mentioned requisites to avoid triviality, the definition of suitable criteria is still an open problem. One of the most pursued ideas is that of minimality, a concept that can be understood syntactically (e. g., [13.3] and [13.5] look for literals), semantically (a minimal explanations is equivalent to any other explanation it implies [13.1]), with respect to the set of possible explanations (the best explanation is the weakest, i. e., the one that is implied by the rest of them), and even with respect to the current information (the best explanation is the one that disrupt less the current information).

In fact, most logical criteria are based on restrictions on the logical form of the solutions but, as mentioned in [13.1], finer criteria to select between two equally valid solutions require contextual aspects. With this idea in mind some approaches have proposed to use an ordering among formulas [13.10, 13.50, 13.51] or among full theories (i. e., possible worlds [13.52, 13.53]). In particular, for the latter, a common option is the use of preferential models (e. g., [13.54]) in which preferential criteria for selecting the best explanation are regarded as qualitative properties that are beyond the pure causal or deductive relationship between the abductive problem and its abductive solution. But these preference criteria are normally treated as an external device, which works on top of the logical or deductive part of the explanatory mechanism, and thus it has been criticized because it seems to fall outside a logical framework.

The epistemic approach of this proposal provides with an interesting alternative. The concepts of an abductive problem and an abductive solution have been defined in terms of the agent’s epistemic attitudes, so it is natural to use such attitudes as a criterion for selecting the best explanation. Consider, for instance, the following elaboration of an example presented in Chap. 10.

“ Mary and Gaby arrive late to Mary’s apartment; the light switch is pressed but the light does not turn on. Knowing that the apartment is old, Mary assumes a failure in the electric line as the explanation for the light not turning on. Gaby, on the other hand, does not have any information about the apartment, so she explains the light not turning on by assuming that the bulb is burned out.”

After pressing the switch, both Mary and Gaby observe that the light does not turn on. There are several explanations for this: It is possible that the electric line failed, as Mary assumed, but it can also be the case that the bulb is burned out, as Gaby thinks, and it is even possible that the switch is faulty. Then, why do they choose a different explanation? The reason is that, though they both observe that the light does not turn on, they have different background information: Mary knows that the apartment is old, and hence she considers a failure in the electric line more likely than any other explanation, but Gaby does not have that piece of information, so for her a burned out bulb explains the lack of light better.

The example shows that, even when facing the same surprising observation (the light does not turn on), agents with different knowledge and beliefs may choose a different best explanation: While Mary assumes that the electric line has failed, Gaby thinks that the bulb is burned out. Both explanations are equally logical since either a failure on the electric line or else a burned out bulb is enough to explain why the light does not turn on. What makes Mary to choose the first and Gaby the second is that they have different knowledge and different beliefs. This suggest first, that, instead of looking for criteria to select the best explanation, the goal should be a criteria to select the agent’s best explanation.

But there is more. The explanation an agent will choose for a given abductive problem depends not only on how the problematic formula could have been predicted, but also on what the agent herself knows and what she considers more likely to be the case. It could be argued that this criterion is not logical in the classical sense because it is not based exclusively on the deductive relationship between the observed fact and the different ways in which it could have been derived. Nevertheless, it is logical in a broader sense since it does depend on the agent’s information: her knowledge and her beliefs. In particular, in the plausibility models framework, the agent’s knowledge and beliefs are defined in terms of a plausibility relation among epistemic possibilities, so it is natural to use precisely this relation as a criterion for selecting each agent’s best explanation(s).

This section presents a straightforward use of this idea. It discusses how the plausibility order among epistemic possibilities can be lifted to a plausibility order among formulas, thus providing a natural criterion to select the agent’s best explanation. A generalization of this idea that works instead with all explanations will be discussed later (Sect. 13.7).

5.1 Ordering Explanations

A plausibility model provides an ordering among possible worlds. This order can be lifted to get an ordering among set of worlds, that is, an ordering among formulas of the language (with each formula seen as the set of those worlds that make it true). The different ways in which such ordering can be defined has been studied in preference logic (see [13.55, 13.56, 13.57] or, for a more detailed exposition [13.58, Chap. 3.3]); this section recalls the main ideas, showing how they can be applied to the task of selecting the best explanation in abductive reasoning.

In general, an ordering among objects can be lifted to an ordering among sets of such objects in different ways. For example, one can say that the lifted ordering puts the set of objects satisfying the property ψ (the set of ψ-objects) over the set of objects satisfying the property \(\varphi\) (the set of \(\varphi\)-objects) when there is a ψ-object that the original ordering among objects places above some \(\varphi\)-object (a \(\exists\exists\) preference of ψ over \(\varphi\); see below). But one can be more drastic and say that the set of ψ-objects is above the set of \(\varphi\)-ones when the original ordering places every ψ-object above every \(\varphi\)-one (a \(\forall\forall\) preference of ψ over \(\varphi\)). This quantification combination gives raise to the following possibilities

$$\begin{array}[]{lll}\varphi\leq_{\exists\exists}\psi&\text{ iff }&\textit{there is}\ \text{a}\ \varphi\text{-object }w\text{ and}\ \textit{there}\\ &&\textit{is}\ \text{a }\psi\text{-object }u\text{ such that }w\leq u\\ \varphi\leq_{\forall\exists}\psi&\text{ iff }&\text{for every}\ \varphi\text{-object }w\ \textit{there is}\ \text{a }\\ &&\psi\text{-object }u\text{ such that }w\leq u\\ \varphi\leq_{\forall\forall}\psi&\text{ iff }&w\leq u\ \textit{for every}\ \varphi\text{-object }w\text{ and}\\ &&\textit{every}\ \psi\text{-object }u\\ \varphi\leq_{\exists\forall}\psi&\text{ iff }&\textit{there is a}\ \varphi\text{-object }w\text{ such that}\\ &&w\leq u\ \textit{for every}\ \psi\text{-object }u\end{array}$$

The first two orderings can be defined within the language \(\mathcal{L}\)

$$\begin{aligned}\displaystyle&\displaystyle\varphi\leq_{\exists\exists}\psi:=\langle\sim\rangle(\varphi\wedge\langle\leq\rangle\psi)\\ \displaystyle&\displaystyle\varphi\leq_{\forall\exists}\psi:=[\sim](\varphi\rightarrow\langle\leq\rangle\psi)\;.\end{aligned}$$

The first formula indicates that there is a ψ-world that is at least as plausible as a \(\varphi\)-one, \(\varphi\leq_{\exists\exists}\psi\), exactly when there is an epistemic possibility that satisfies \(\varphi\) and that can see an at least as plausible ψ-world. The second one only changes the first quantification (turning, accordingly, the conjunction into an implication): For every \(\varphi\)-world there is a ψ-world that is at least as plausible.

The last two orderings are not immediate. Given the formulas for the previous two orderings, one could propose \([\sim](\varphi\rightarrow[\leq]\psi)\) for the \(\forall\forall\) case, but this formula is not correct: It states that every world that is at least as plausible as any \(\varphi\)-world satisfies ψ, but it does not guarantee that every ψ-world is indeed above every \(\varphi\)-world:

  1. 1.

    There might be a ψ-world incomparable to some \(\varphi\)-one, and even if all worlds are comparable

  2. 2.

    There might be a ψ-world strictly below a \(\varphi\)-one (<, the strict version of ≤, is defined as w < u if and only if w ≤ u and not u ≤ w).

The plausibility order is locally connected (i. e., inside each epistemic partition, every world is comparable to each other) so (1) cannot occur. Thus, a formula defining \(\leq_{\forall\forall}\) only needs to guarantee that no ψ-world is strictly below a \(\varphi\)-one; in other words, it needs to express that, given any ψ-world, every world that is strictly more plausible satisfies \(\neg\varphi\). Such formula can be easily stated in a language that extends \(\mathcal{L}\) with a standard modality for the relation <

$$\varphi\leq_{\forall\forall}\psi\ :=\ [\sim](\psi\rightarrow[<]{\neg\varphi})\;.$$

Finally, the \(\exists\forall\) ordering presents a similar situation. Following the first two cases one could propose \(\langle\sim\rangle(\varphi\wedge[<]\psi)\), but such formula is not appropriate, even in the current full-comparability case: It holds even when there are ψ-worlds below the chosen \(\varphi\)-one. In order to guarantee the existence of a \(\varphi\)-world that is at most as plausible as every ψ-world, the formula should state that every world that is strictly less plausible than the \(\varphi\)-world satisfies \(\neg\psi\). Extending the language again, this time with a modality for >, makes such formula straightforward

$$\varphi\leq_{\exists\forall}\ \psi:=\ \langle\sim\rangle(\varphi\wedge[> ]\neg\psi)\;.$$

All in all, the important fact is that among these four orderings on sets of worlds (i. e., formulas), two are definable within \(\mathcal{L}\) and the other two only need simple extensions. This shows how the plausibility order among worlds that defines the agent’s knowledge and beliefs (Sect. 13.3.1) also defines plausibility orderings among formulas (sets of worlds), and hence provides a criterion for selecting the best abductive solution for a given agent. It will now be shown how this criterion can be used, and how it leads to situations in which agents with different knowledge and beliefs choose different best explanations.

Example 13.5

Recall Mary and Gaby’s example. Both observe that after pressing the switch the light does not turn on, but each one of them chooses a different explanation: While Mary assumes that the electric line failed, Gaby thinks that the bulb is burned out. As it has been argued, the reason why they choose different explanations is that they have different knowledge and beliefs. Here is a formalization of the situation.

The following plausibility models show Mary and Gaby’s knowledge and beliefs before pressing the switch. They involve tree atomic propositions: l standing for lack of light, e standing for a failure in the electric line and b standing for a burned out bulb. Again, each possible world has indicated within it exactly those atomic propositions that are true in each one of them, and the arrows represent the plausibility relation (transitive arrows are omitted).

figure hfigure h

Observe how both Mary and Gaby know that both a failure on the electric line and a burned out bulb imply lack of light (both \(e\rightarrow l\) and \(b\rightarrow l\) hold in every world). In fact, the only difference in the models is the plausibility order between worlds w2 and w4. Mary knows that the apartment is old so she considers a failure on the line (e) more likely than a burned out bulb (b), and hence the situation where the electric line fails but the bulb is not burned out (w2) is more likely than its opposite (w4). Gaby, on the other hand, does not know anything about the apartment, and hence for her a burned out bulb with a working electric line (w4) is more plausible than a working bulb and a failing electric line (w2). It is also assumed that, for both of them, the most likely possibility is the one in which everything works correctly (w1) and the least plausible case is the one in which everything fails (w5).

After they both observe that pressing the switch does not turn on the light, the unique world where l is not the case, w3, is eliminated, thus producing the following models.

figure ifigure i

As a result of the observation, Mary and Gaby know that there is no light (\(K{l}\) holds in both models), something that they did not know before. Thus, following Definition 13.9, both have an abductive problem with l.

According to Definition 13.11, both e and b are abductive solutions for the abductive problem l for both Mary and Gaby: Both formulas are the antecedent of implications that have l as a consequent and that were known before the observation. So, how can each girl choose her own best explanation? For Mary, the unique ordering that puts b above e is the weakest one, \(\exists\exists\) (there is a b-world, w4, at least as plausible as a e-one, w5). Nevertheless, from her point of view, e is above b not only in the weak \(\exists\exists\) way (w2 is at least as plausible as w4) but also in the stronger \(\forall\exists\) way (every b-world has a e-world that is at least as plausible as it). Thus, one can say that e is a more plausible explanation from Mary’s perspective. In Gaby’s case something analogous happens: b is above e not only in the weak \(\exists\exists\) way (w4 is at least as plausible as w2) but also in the strong \(\forall\exists\) way. Hence, it can be said that, for Gaby, b is the best explanation.

6 Integrating the Best Solution

Once the agent has selected the best explanation for her, she can incorporate it into her information. As discussed in Sect. 13.2, even though the nonmonotonic nature of abductive reasoning indicates that an abductive solution should not be assimilated as knowledge, the richness of the present framework allows the possibility to integrate it as a part of the agent’s beliefs. Here is a modality describing such action.

Definition 13.15 Modality for abductive reasoning

Let \((M,w)\) be a pointed plausibility model and consider again \((M_{\psi!},w)\), the pointed plausibility model that results from observing ψ at \((M,w)\). Every pair of formulas η and χ in \(\mathcal{L}\) define an existential modality \(\langle\text{Abd}_{\eta}^{\chi}\rangle\varphi\), read as the agent can perform an abductive step for χ with η after which \(\varphi\) is the case, and whose semantic interpretation is as follows

$$\begin{aligned}\displaystyle&\displaystyle(M_{\psi!},w)\Vdash\langle\text{Abd}_{\eta}^{\chi}\rangle\varphi\\ \displaystyle&\displaystyle\quad\text{iff}\\ \displaystyle&\displaystyle(1)(M_{\psi!},w)\Vdash K{\chi}\text{ and }(M,w)\Vdash\neg K{\chi}\;,\\ \displaystyle&\displaystyle(2)(M,w)\Vdash K(\eta\rightarrow\chi)\;,\text{ and }\\ \displaystyle&\displaystyle(3)((M_{\psi!})_{\eta\Uparrow},w)\Vdash\varphi\;.\end{aligned}$$

Equivalently, \(\langle\text{Abd}_{\eta}^{\chi}\rangle\varphi\)’s semantic interpretation can be defined as

$$\begin{aligned}\displaystyle&\displaystyle(M_{\psi!},w)\Vdash\langle\text{Abd}_{\eta}^{\chi}\rangle\varphi\\ \displaystyle&\displaystyle\quad\text{iff}\\ \displaystyle&\displaystyle(M,w)\Vdash\neg K{\chi}\wedge K(\eta\rightarrow\chi)\wedge[\psi!](K{\chi}\wedge[\eta\Uparrow]\varphi)\;.\end{aligned}$$

The definition states that \(\langle\text{Abd}_{\eta}^{\chi}\rangle\varphi\) is true at \((M_{\psi!},w)\) if and only if:

  1. 1.

    χ is an abductive problem at \((M_{\psi!},w)\)

  2. 2.

    η is an abductive solution also at \((M_{\psi!},w)\)

  3. 3.

    An upgrade (Definition 13.8) with η will make \(\varphi\) true.

The last part makes precise the idea of how an agent should incorporate the selected explanation: It cannot be incorporated as knowledge, but it can be incorporated as a belief.

Example 13.6

Returning to Example 13.5, once Mary and Gaby have selected their respective best explanation, they can perform an abductive step. In Mary’s case, worlds satisfying e (w5 and w2) will become more plausible than worlds that do not satisfy it (w4 and w1); in Gaby’s case, worlds satisfying b (w5 and w4) will become more plausible than worlds that do not satisfy it (w2 and w1). Applying these upgrades to the models Mary\({}_{l!}\) and Gaby\({}_{l!}\) in produces the following models.

figure jfigure j

As a result of the abductive step, each agent believes her own explanation: Mary believes that the electric line has failed (e is true in her unique most plausible world w2), and Gaby believes that the bulb is burned out (b is true in her unique most plausible world w4). That is, for every \(w\in\{w_{1},w_{2},w_{4},w_{5}\}\),

$$\begin{aligned}\displaystyle&\displaystyle(\text{Mary}_{l!},w)\quad\Vdash\quad\langle\text{Abd}_{e}^{l}\rangle Be\\ \displaystyle&\displaystyle(\text{Gaby}_{l!},w)\quad\Vdash\quad\langle\text{Abd}_{b}^{l}\rangle Bb\;.\end{aligned}$$

6.1 Abduction in a Picture, Once Again

The definitions that have been provided allow more precision in the diagram of abductive reasoning presented in Sect. 13.2.5. Here is the updated version for the case in which the definitions are given just in terms of the agent’s knowledge. Note how the inferring χ step has been dropped, as it is not needed in an omniscient setting such as DEL. Again, circles represent the agent’s epistemic states (i. e., full plausibility models) and arrows are labeled with the operations that modify the agent’s information.

figure kfigure k

Again, the upper path represents what really happened. After observing ψ, the agent reaches the epistemic state s2 in which she knows χ. But before the observation, at s1, she did not know χ, and thus this formula is an abductive problem at s2. Observe how \(\eta\rightarrow\chi\) was known at s1: hence, η is an abductive solution at s2 and the agent can perform an abductive step with it to reach state s3. This abductive solution η would have helped the agent to infer (and hence to come to know) χ, and the lower path represents this alternative situation. In general, it cannot be guaranteed that the agent would have known χ (or even η) at state \(s^{\prime}_{2}\): these formulas could have had epistemic modalities, and hence the observation could have changed their truth value. However, if both formulas are propositional, \(K\chi\) and \(K\eta\) hold at \(s^{\prime}_{2}\).

6.2 Further Classification

Section 13.4.4 presented an epistemic version of the the common classification of abductive solutions. But the current DEL setting allows further possibilities and hence a finer classification. For example, here are two straightforward ideas. First, a solution η has been defined as the antecedent of an implication that has χ as a consequent and that was known before the epistemic action that triggered the problem. Nevertheless, given that both formulas might contain epistemic operators, the agent can go from knowing the implication to not knowing it. Second, it has been stated that the agent incorporates the selected explanation via a belief revision (i. e., an upgrade). Nevertheless, since the solution might contain epistemic operators, the upgrade does not guarantee that the agent will believe the solution after the operation.

Definition 13.16 Adequate solution and successful solution

Let the formula η be an abductive solution for the abductive problem χ at \((M_{\psi!},w)\). Then:

  • η is an adequate solution if and only if the agent still knows \(\eta\rightarrow\chi\) at \((M_{\psi!},w)\), that is, if and only if

    $$(M_{\psi!},w)\Vdash K(\eta\rightarrow\chi)\;.$$
  • η is a successful solution if and only if it is believed after the abductive step, that is, if and only if

    $$(M_{\psi!},w)\Vdash\langle\text{Abd}_{\eta}^{\chi}\rangle B{\eta}\;.$$

Here it is a result about the adequacy property.

Proposition 13.2

Every abductive solution is adequate.

Proof:

More precisely, suppose that at \((M_{\psi!},w)\) the formula χ is an abductive problem and η is one of its abductive solutions. Since χ is an abductive problem, \((M_{\psi!},w)\Vdash K{\chi}\) and hence \((M_{\psi!},w)\Vdash K(\eta\rightarrow\chi)\).

Given this result, this property is of little interest in the current setting. However, it becomes interesting in settings with nonomniscient agents. In such frameworks, it is possible for the agent not to know \(\eta\rightarrow\chi\) even when she knows χ and she knew \(\eta\rightarrow\chi\) before.

Here it is another result, now about the property of being a successful solution.

Discussion 13.2

Not every abductive solution is successful.

Proof:

Let \(\mathrm{P}=\{p,q\}\) be the set of atomic propositions, and consider the pointed plausibility models below (reflexive and transitive arrows omitted) in which the evaluation points are double circled.

figure lfigure l

Observe how q is an abductive problem at \(M_{q!}\) since it is not known at M (there is an epistemically possible world where q fails, namely, w3) but it is known at \(M_{q!}\). Observe also how \(p\wedge\neg B{p}\) is an abductive solution since \(K((p\wedge\neg B{p})\rightarrow q)\) holds at M (it is true at w1 and w2 because q is true in those worlds, and also true at w3 because \(p\wedge\neg B{p}\) fails in this world). Furthermore, \(p\wedge\neg B{p}\) is a consistent solution since it is epistemically possible in \(M_{q!}\) (p and \(\neg B{p}\) are both true at w1, the latter because there is a most plausible world, w2, where p is not the case, and hence the agent does not believe p). Nevertheless, after an upgrade with \(p\wedge\neg B{p}\) this very formula is not believed. It fails at the unique most plausible world w1 because \(\neg B{p}\) fails at it: the most plausible world (w1 itself) satisfies p and hence the agent now believes p, that is, \(B{p}\) is the case.

Nevertheless, if a propositional solution η is also consistent, then it is successful.

Proposition 13.3

Suppose that at \((M_{\psi!},w)\) the formula η is an abductive solution for the abductive problem χ. If η is a propositional and consistent solution, then it is successful.

Proof:

If η is a consistent solution, then at \((M_{\psi!},w)\) there is at least one epistemically possible η-world. Therefore, an upgrade with η will put worlds that satisfied η in \((M_{\psi!},w)\) on top of the plausibility order. Now, η is propositional, and hence its truth value depends only on the valuation of each possible world; since the upgrade operation does not affect the valuation, then any world satisfying η in \(M_{\psi!}\) will still satisfy it in \((M_{\psi!})_{\eta\Uparrow}\). Hence, after the operation, the most plausible worlds will satisfy η, and thus \(((M_{\psi!})_{\eta\Uparrow},w)\Vdash B{\eta}\) will be the case. This, together with the fact that at \(M_{\psi!}\) the formula χ is an abductive problem and the formula η is an abductive solution, yield \((M_{\psi!},w)\Vdash\langle\text{Abd}_{\eta}^{\chi}\rangle B{\eta}\).

It has been already stated that a solution is explanatory when it changes the agent’s information. A further classification of abductive solutions can be provided according to how much they change the agent’s information, that is, according to the attitude of the agent toward the solution before it was incorporated.

Definition 13.17

Suppose that χ is an abductive problem at \((M_{\psi!},w)\). An explanatory abductive solution η is:

  • Weakly explanatory when \((M_{\psi!},w)\Vdash B{\eta}\)

  • Neutral when \((M_{\psi!},w)\Vdash\neg B{\eta}\wedge\neg B\neg\eta\)

  • Strongly explanatory when \((M{\psi!},w)\Vdash B\neg\eta\).

Again, there are more possibilities if further epistemic attitudes are considered.

6.3 Properties in a Picture

Consider an anomalous abductive problem χ (i. e., \(B{\neg\chi}\) holds at s1) whose abductive solution η is consistent (\(\hat{K}{\eta}\) holds at s2) and successful (\(B{\eta}\) holds at s3), recalling also that every solution is adequate (so \(K(\eta\rightarrow\chi)\) holds at s2). This extends the diagram of Sect. 13.6.1 in the following way.

figure mfigure m

Moreover, consider the case in which both χ and η are propositional, the typical case in abductive reasoning in which the agent looks for explanations of facts, and not of her own (or, in a multiagent setting, of other agents’) epistemic state. First, in such case, η should be an epistemic possibility not only at s2 but also at s1. But not only that; it is possible now to state the effects of the abductive step at s2 (the agent will believe η and will still know \(\eta\rightarrow\chi\)) and of the hypothetical announcement of η at s1 (she would have known both η and χ, and she would have still known \(\eta\rightarrow\chi\)). Therefore,

figure nfigure n

This diagram beautifully illustrates what lies behind this proposal’s understanding of abductive reasoning. In the propositional case, if η is a consistent and successful abductive solution for the abductive problem χ, then, after abductive reasoning, the agent will know χ and will believe η. In fact, when the observed formula ψ is actually the same χ that becomes an abductive problem, the epistemic effect of abductive reasoning, from knowledge to beliefs, can be described with the following validity [13.59],

$$K(\eta\rightarrow\chi)\rightarrow[\chi!](K{\chi}\rightarrow\langle\text{Abd}_{\eta}^{\chi}\rangle B{\eta})\;.$$

What makes η a reasonable solution is the existence of an alternative reality in which she observed η and, thanks to that, came to know χ. Similar diagrams can be obtained for the cases in which the definitions of an abductive problem and an abductive solution are given in terms of epistemic attitudes other than knowledge.

7 Working with the Explanations

The reason why abductive solutions are incorporated as beliefs and not as knowledge is because the selected explanation (in fact, any explanation) is just a hypothesis, subject to change in light of further information. Consider the following continuation of the Mary and Gaby’s situation.

Example 13.7

After their respective abductive steps (models (Mary\({}_{!l}\))\({}_{e\Uparrow}\) and (Gaby\({}_{l!}\))\({}_{b\Uparrow}\) of Example 13.6), Mary and Gaby take a closer look at the bulb and observe that it is not burned out (\(\neg b\)). Semantically this is simply an observation operation that eliminates w4 and w5, exactly those epistemic possibilities where the bulb is burned out (i. e., where b holds). The resulting models are the following.

figure ofigure o

This observation does not affect Mary’s explanation: She still believes that the electric line has failed (e is true in her unique most plausible world w2). But Gaby’s case is different: She does not have an explanation for l anymore. Although she knows it (\(K{l}\) holds at the model on the bottom, that is, l is true in every epistemic possibility), she neither knows nor believes the antecedent of a known implication with l as a consequent (besides, of course, the trivial ones); she needs to perform a further abductive step in order to explain it.

There is, nevertheless, a way to avoid the extra abductive reasoning step. Recall that after applying the defined upgrade operation (Definition 13.8), all the worlds satisfying the given formula become more plausible than the ones that do not satisfy it, and within the two zones the old ordering remains. If the lifted worlds are not those that satisfy the agent’s most plausible explanation but rather those that satisfy at least one of her explanations, the resulting model will have two layers: the lower one with worlds that do not satisfy any explanation, and the upper one with worlds that satisfy at least one. But inside the upper layer the old ordering will remain. In other words, the most plausible worlds in the resulting model (i. e., the most plausible ones in the upper layer) will be the ones that satisfy at least one explanation and that were already more plausible than the rest. Thus, with respect to the most plausible explanation, the same result is achieved: After such upgrade, roughly, the agent will believe the explanation that was the most plausible for her.

The difference with respect to the approach of the previous section is that the worlds that appear below the most plausible ones are not arbitrary. Worlds on the second best layer satisfy already some explanation; an explanation that was not chosen because it was not the most plausible one. Then, if further observations make the original best explanation obsolete, once that the correspondent (and now also obsolete) worlds have been discarded, the ones that will be at the top of the plausibility ordering will be the previously second best. Thus, an explanation will be already present and no further abductive steps will be needed.

7.1 A Modality

The idea just described is formalized now by introducing a modality that, given an abductive problem χ, upgrades those worlds that satisfy at least one of its abductive explanations.

Definition 13.18 Modality for formula-based abduction

Let \((M,w)\) be a pointed plausibility model and consider again \((M_{\psi!},w)\), the pointed plausibility model that results from observing ψ at \((M,w)\). Every formula χ in \(\mathcal{L}\) defines an existential modality of the form \(\langle\text{Abd}\,{\chi}\rangle\varphi\), read as the agent can perform a complete abductive step for χ after which \(\varphi\) is the case, and whose semantic interpretation is as follows

$$\begin{aligned}\displaystyle&\displaystyle(M_{\psi!},w)\Vdash\langle\text{Abd}\,{\chi}\rangle\varphi\\ \displaystyle&\displaystyle\quad\text{ iff }\\ \displaystyle&\displaystyle(1)(M_{\psi!},w)\Vdash K{\chi}\text{ and }(M,w)\Vdash\neg K{\chi}\;,\\ \displaystyle&\displaystyle(2)((M_{\psi!})_{(\bigvee\Sigma_{\chi})\Uparrow},w)\Vdash\varphi\;,\end{aligned}$$

where Σχ is the set of abductive solutions for χ, that is,

$$\Sigma_{\chi}:=\{\eta\mid(M,w)\Vdash K(\eta\rightarrow\chi)\}\;.$$

Equivalently, \(\langle\text{Abd}\,{\chi}\rangle\varphi\)’s semantic interpretation can be defined as

$$\begin{aligned}\displaystyle&\displaystyle(M_{\psi!},w)\Vdash\langle\text{Abd}\,{\chi}\rangle\varphi\\ \displaystyle&\displaystyle\quad\text{ iff }\\ \displaystyle&\displaystyle(M,w)\Vdash\neg K{\chi}\wedge[{\psi!}](K{\chi}\wedge[\vee\Sigma_{\chi}\Uparrow]\varphi)\;.\end{aligned}$$

The correspondent universal modality, \([\text{Abd}\,{\chi}]\), is defined as usual.

The definition states that \(\langle\text{Abd}\,{\chi}\rangle\varphi\) is true at \((M_{\psi!},w)\) if and only if (1) χ is an abductive problem at \((M_{\psi!},w)\), and (2) an upgrade with \(\bigvee\Sigma_{\chi}\) will make \(\varphi\) true. The last part makes precise the idea of working with all the solutions: Σχ contains all abductive solutions for χ, so \(\bigvee\Sigma_{\chi}\) is a disjunction characterising those worlds that satisfy at least one of them and hence an upgrade with it will move such worlds to the topmost layer. But inside this layer, the former plausibility order will persist, and hence worlds at the top of it will be precisely those that satisfy at least one solution for χ and, among them, were already the most plausible ones.

Remark 13.1

The set Σχ contains, among others, χ, \(\chi\wedge\chi\), and so on, and hence \(\bigvee\Sigma_{\chi}\) is an infinite disjunction. Syntactic restrictions can be imposed in order to avoid such situations (e. g., asking for solutions that are also minimal conjunctions of literals). Another possibility, closer to the semantic spirit of this approach, is to work with finite plausibility models, and then look for solutions among the formulas that characterize each possible world.

The following example shows how this new operation allows the agent to have ready another explanation in case the initially best one turns out to be incorrect.

Example 13.8

Let us go back to Mary and Gaby’s example all the way to the stage after which they have observed that the light does not turn on (models Mary\({}_{l!}\) and Gaby\({}_{l!}\) of Example 13.5, repeated here).

figure pfigure p

Suppose that, instead of selecting their respective most plausible explanation and assimilating it (as they did in Example 13.5), Mary and Gaby work with all their explanations: Instead of an upgrade with e for Mary and an upgrade with b for Gaby, both of them perform an upgrade with \(e\vee b\). This produces the following models.

figure qfigure q

The worlds satisfying \(e\vee b\) (w2, w4, and w5) have been upgraded. As a result of this, both Mary and Gaby have an explanation for l, but each one of them has her own explanation: While Mary believes that the electric line has failed (e is the case in the most plausible world at the model on the top), Gaby believes that the bulb is burned out (b holds in the most plausible world at the model on the bottom).

So far the result of the upgrade is, with respect to Mary and Gaby beliefs, exactly the same as with the previous proposal where only worlds that satisfy the most plausible explanation are upgraded (in both cases, w2 and w4 are Mary’s and Gaby’s most plausible worlds, respectively). But note what happens now when they both observe that the bulb is in fact not burned out (\(\neg b\)): Such action produces the following situation.

figure rfigure r

Again, the observation does not affect Mary’s explanation (e still holds in the most plausible world at model on the top), but it does change Gaby’s since her previous explanation b is not possible anymore. The difference is that now she does not need to perform an extra abductive step because she has already another explanation: She now believes that the electric line has failed (e holds in the most plausible world at model on the bottom).

Thus, after an upgrade with all explanations, what the agent will be lead to believe depends on her plausibility order, just as with the first proposal. Nevertheless, if further information invalidates such best explanation, the agent will believe the next to best one without the need of further abductive steps.

8 A Brief Exploration to Nonideal Agents

As most (if not all) proposals for representing a given phenomena, the presented epistemic and dynamic approach to abduction has made some assumptions for the sake of simplicity. One of the most important of these is the fact that agents whose information is represented within the plausibility framework are ideal: Their knowledge and beliefs are closed under logical consequence. This supposition is not exclusive of this approach; the classic logical definitions of abductive reasoning assume not only that the given set of formulas Φ, the theory, is closed under logical consequence, but also that \(\vdash\) is the logical consequence relation.

The present proposal highlights the epistemic nature of abductive reasoning, and so it is natural to ask how such reasoning process works for a different kind of agents, in particular, for those whose information does not need to have ideal properties and thus are, in that sense, closer to real computational agents with limited resources (and also closer to us human beings). This final section briefly discusses some ideas; further developments in this direction can be found in [13.60].

8.1 Considering Inference

Suppose Karl is in his dining room and sees smoke coming out of the kitchen. This seems unjustified at first, but then he realises that the chicken he placed on the fire has been there for a long time. Initially Karl did not have any explanation for the smoke, but after a moment he realized that such event was actually not surprising at all.

This case is different from the discussed ones because Karl is not an ideal agent: He does not have at hand all logical consequences of his information, and therefore he did not realize that the information he had before seeing the smoke was enough to predict it (i. e., to infer that there would be smoke). Described in more technical terms, seeing the smoke raised an abductive problem for Karl, but such problem arose because he did not have, at the time of the observation, all the logical consequences of the information he actually had (otherwise there would have been no abductive problem at all). Accordingly, in such case the abductive solution is not necessarily a piece of information that would have allowed Karl to predict the smoke; it might be a simple inference step that made explicit what was only implicit before.

This shows not only how agents whose information is not closed under logical consequence can face at least a new kind of abductive problem, but also how such problems give rise to a different kind of solutions.

8.2 Different Reasoning Abilities

In the previous example, the abductive solution was a simple inference step because Karl had the needed reasoning tools to infer there is smoke in the kitchen from the chicken has been on the fire for a long time. But what if that was not the case? That is, what if, besides not having at hand all the logical consequences of his information, Karl did not have the required reasoning tools to infer some of them?

In such new situation, Karl faces again an abductive problem, but this time of a different nature. The surprising observation could have been predicted in the sense that it is a logical consequence of Karl’s information the chicken has been on the fire for a long time, just as in the initial version of this example. The difference is that such observation is not something that Karl could have predicted by himself: He did not have the needed tools. One can say that, even though there is smoke in the kitchen is objectively derivable from the initial information, it is not subjectively derivable in the sense that Karl could not have done it. To put it in other words, besides not having at hand all the logical consequences of her actual information, Karl might not even be able to reach them.

Accordingly, the simple inference step of before cannot be a solution to the problem now, as Karl does not have the needed tools to perform it. One possible solution is, as in the traditional case, a piece of information that would have allowed Karl to predict the smoke from some other previously known fact, but a more interesting one is some reasoning tool that would have helped him to predict the fire from the known fact the chicken has been on the fire for a long time.

New cases arise when further kinds of agents are considered. A systematic study of such cases can be found in [13.61].

9 Conclusions

This chapter has proposed an epistemic and dynamic approach to abductive reasoning, understanding this form of reasoning as a process that:

  1. 1.

    Is triggered by an epistemic action through which the agent comes to know or believe certain χ that otherwise she could not have been able to know or believe

  2. 2.

    Looks for explanations for χ in the set of formulas that could have helped the agent to come to know or believe χ

  3. 3.

    Incorporates the chosen explanation as a part of the agent’s beliefs.

Besides providing formal definitions of what an abductive problem and an abductive solution are in terms of an agent’s knowledge and beliefs, the present proposal has discussed:

  1. 1.

    A classification of abductive problems in terms of both how convinced the agent is of the problematic formula after the observation (she knows it, or just believes it) and how plausible the formula was before the epistemic action that triggered the problem

  2. 2.

    A classification of abductive solutions based not only on their deductive relation with the abductive problem or their syntactic form, but also in terms of both their plausibility before the problem was raised and the way it will affect the agent’s information once they are incorporated

  3. 3.

    A new perspective that looks not for the best explanation but rather for the agent’s best explanation, and the possibility to carry out this search in terms of which explanations are more likely from the agent’s point of view, that is, in terms of the agent’s beliefs

  4. 4.

    The possibility of integrating the chosen solution into the agent’s information as part of her beliefs, which allows not only to identify trivial solutions because of their effect rather than their form, but also to revise and eventually discard solutions that become obsolete in the light of further information.

Crucial for all these contributions has been the use of plausibility models and, in general, the DEL guidelines, which puts emphasis in the representation of both epistemic attitudes and the actions that affect them.

It is worthwhile to compare, albeit briefly, the present proposal to other epistemic approaches to abductive reasoning. Besides immediate differences in the respective semantic models (while other approaches follow the Alchourrón–Gärdenfors–Makinson (GlossaryTerm

AGM

) belief revision, using a set of formulas for representing the agent’s information, here possible worlds are used), there are two main points that distinguish the presented ideas from other proposals. First, here several epistemic attitudes are taken into account, thus making a clear difference between what the agent holds with full certainty (knowledge) and what she considers very likely but still cannot guarantee (beliefs); this allows to distinguish between the certainty of both the previous information and the surprising observation, and the mere plausibility of the chosen solution (recall the validity \(K(\eta\rightarrow\chi)\rightarrow[\chi!](K{\chi}\rightarrow\langle\text{Abd}_{\eta}^{\chi}\rangle B{\eta})\), briefly discussed at the end of Sect. 13.6). Second, this approach goes one step further by making explicit the different stages of the abductive process, thus making also explicit the epistemic actions involved. This highlights the importance of actions such as belief revision, commonly understood in epistemic approaches to abduction as the one triggered by the abductive problem [13.12, 13.62], and also such as observation, understood here as the one that triggers the abductive process.

This chapter presents only the first steps toward a proper study of abductive reasoning from an epistemic and dynamic perspective, and several of the current proposals can be refined. For example, the specific definition of an abductive problem (Definition 13.9) relies on the fact that, within the DEL framework, agents are logically omniscient. As it has been hinted at in Sect.13.8, in a nonomniscient DEL setting [13.35, 13.63] the ideas discussed in Sect. 13.2 would produce a different formal definition (which, incidentally, would allow to classify abductive problems and abductive solutions according to some derivability criteria). Moreover, it would be possible to analyze the full abductive picture presented in Sect. 13.2.1, which requires inference steps in the alternative reality path. These extensions are relevant: They would allow a better understanding of the abductive process as performed by real agents.

But it is also possible to do more than just follow the traditional research lines in abductive reasoning, and here are two interesting possibilities (whose development exceeds the limits of this chapter). First, the DEL framework allows multiagent scenarios in which abductive problems would arise in the context of a community of agents. In such setting, further to the public observation and revision used here, actions that affect the knowledge and beliefs of different agents in different ways are possible. For example, an agent may be privately informed about ψ: If this raises an abductive problem χ for her and another agent has private information about \(\eta\rightarrow\chi\), they can interact to obtain the abductive solution η. Second, the DEL framework deals with high-order knowledge, thus allowing to study cases in which an agent, instead of looking for an explanation of a fact, looks for an explanation of her own epistemic state. Interestingly, explanations might involve epistemic actions as well as the lack of them.

According to those considerations, this logical approach takes into account the dynamics aspects of logical information processing, and one of them is abductive inference, one of the most important forms of inference in scientific practices. The aforementioned multiagent scenarios allow to model concrete practices, particularly those that develop a methodology based on observation, verification, and systematic formulation of provisional hypotheses, such as in empirical sciences, social sciences, and clinical diagnosis. The epistemological repercussions of this DEL approach is given by the conceptual resources that it offers, useful to model several aspects of explanatory processes. If known theories of belief revision, at the last resort, say nothing about context of discovery, by means of DEL the accessibility of this context to rational epistemological and logical analysis is extended, further on classical logical treatment of abduction. From the perspective of game theoretic semantics, for example, now it is easier to determine what rules are strategic and what are operatories when abductive steps were given. But applications should also be considered to tackle certain philosophical problems. For example, abductive scenarios within multiagent settings can be used to study the implications of different forms of communication within scientific communities.