1 Introduction

Evans et al. [15] carried out a psychological study about deductive reasoning, which demonstrated possibly conflicting processes in human reasoning. Participants were presented different syllogisms for which they had to decide whether they accepted these syllogisms as valid. Consider \(\textsc {S}_{ vit }\):

figure a

The Conclusion necessarily follows from the premises under classical logic. However, about half of the participants said that the syllogism was not valid. They were explicitly asked to logically validate or invalidate various syllogisms, but did not seem to have the intellectual capability to do so. Even worse, they were not even aware about their inabilities. Participants reflectively read the instructions and understood well that they were required to reason logically from the premises to the conclusion. However, the results show that their intuitions were stronger and delivered a tendency to say ‘yes’ or ‘no’ depending on whether the syllogism was believable [14]. The responses of participants for various syllogisms, which differed with respect to their validity and whether they were believable in the contextual setting, were evaluated in [15]. Four of them are depicted in Table 1. The first two premises of all four cases are of the same logical form, namely No A are B. Some C are B. The first two cases differ from the last two cases with respect to the conclusions: In the first two cases the conclusions correspond to the logical form Some C are A and in the last two cases the conclusions correspond to the logical form Some A are C. The first two syllogisms are indeed valid under classical logic, whereas the last two are not. However, as the last column shows, the percentage of the participants that validated the syllogism, does not necessarily comply with the results under classical logic. Evans, Barston and Pollard asserted that the participants were influenced by their own beliefs, their so-called belief bias.

Table 1. Four types of syllogisms taken from [11]. The percentages of the participants that accepted the syllogism as being valid are shown in the last column.

Khemlani and Johnson-Laird [24] have compared the predictions of 12 cognitive theories to participants’ responses in syllogistic reasoning. The Verbal Model Theory [28] performed best with an accurate prediction of 84%, closely followed by the Mental Model Theory [21], which achieved 83%. Recently, [3] developed a logical form for the representation of syllogisms under the logic programming approach, the Weak Completion Semantics, and predicted even 89% of the participants’ responses.

The Weak Completion Semantics is a new cognitive theory, which originates from [30], but is mathematically sound [19], and has been successfully applied – among others – to the suppression task [8], the selection task [9] to reasoning about conditionals [5, 7] and to spatial reasoning [6]. As the Weak Completion Semantics aims at modeling human reasoning adequately and predicted well the participants’ responses in syllogistic reasoning, a natural question to ask, is whether the belief-bias effect in syllogistic reasoning can be modeled within this approach.

After briefly discussing the belief-bias effect, we introduce the Weak Completion Semantics. Taking [3, 4] as starting point, Sects. 4 and 5 present six principles for modeling quantified statements in human reasoning and show their representations in logic programs. Finally, Sect. 6 presents how the belief-bias effect can be modeled under the Weak Completion Semantics by discussing the four cases in Table 1.

2 The Belief-Bias Effect

Evans et al. [13] distinguish between the negative and the positive belief bias: The negative belief bias describes the case when the support for an unbelievable conclusion is suppressed. On the other hand, the positive belief bias describes the case when the acceptance for a believable conclusion is raised. Consider again Table 1: The negative belief bias happens for 46% of the participants in the case of \(\textsc {S}_{ vit }\) and the positive belief bias happens for 92% of the participants in the case of \(\textsc {S}_{ cig }\).

As pointed out in [16], Wilkins [31] already observed that syllogisms which conflict with our beliefs are more difficult to solve. Since then, various theories have tried to explain why humans deviate from the classical logically valid answers. Some conclusions can be explained by converting the premises as proposed in [2] or by assuming that the type of the premises creates an atmosphere which influences the acceptance for the conclusion [16, 32]. Johnson-Laird and Byrne [22] proposed the mental model theory [21], which additionally supposes the search for counterexamples when validating the conclusion. Later, Stenning and van Lambalgen [30] explain why certain aspects influence the interpretations made by humans when evaluating syllogisms and discuss this in the context of mental models. Evans et al. [10, 15] proposed a theory, which in the literature is sometimes referred to as the selective scrutiny model [1, 16]. First, humans heuristically accept any syllogism having a believable conclusion, and only proceed with a logical evaluation if the conclusion contradicts their belief. Adler and Rips [1] claim that this behavior is rational in the sense of efficient belief maintenance. Yet another approach, the selective processing model [12], accounts only for a single preferred model: If the conclusion is neutral or believable, humans attempt to construct a model that supports it. Otherwise, they attempt to construct a model that rejects it.

According to Garnham and Oakhill [16] the belief-bias effect can take place at several stages: First, beliefs can influence our understanding of the premises. Second, in case a statement contradicts our belief, we might search for alternative models and check whether the conclusion is plausible. This seems to comply with Stenning and van Lambalgen’s proposal to model human reasoning by a two step procedure [30]. The first step, the representational part, determines how our beliefs influence the understanding of the premises. The second step, the procedural part, determines whether we search for alternative models based on the plausibility of the conclusion.

In this paper we will follow up on this distinction when modeling the belief-bias effect.

3 Weak Completion Semantics

The general notation, which we will use in the paper, is based on [18, 25].

3.1 Logic Programs

We restrict ourselves to datalog programs, i.e., the set of terms consists only of constants and variables.

$$\begin{aligned} A\leftarrow & {} L_1\wedge \ldots \ldots \wedge L_n. \end{aligned}$$
(1)
$$\begin{aligned} A\leftarrow & {} \top . \end{aligned}$$
(2)
$$\begin{aligned} A\leftarrow & {} \bot . \end{aligned}$$
(3)

A is an atom and the \(L_i\) with \(1 \le i \le n\) are literals. The atom A is called head of the clause and the subformula to the right of the implication symbol is called body of the clause. If the clause contains variables, then they are implicitly universally quantified within the scope of the entire clause. A ground clause is a clause not containing variables. Clauses of the form (2) and (3) are called facts and assumptions, respectively. The notion of falsehood appears counterintuitive at first sight, but programs will be interpreted under their (weak) completion where we replace the implication by the equivalence sign. We assume a fixed set of constants, denoted by \(\mathcal{C}\), which is nonempty and finite. \(\mathsf {constants}(\mathcal{P})\) denotes the set of all constants occurring in \(\mathcal{P}\). If not stated otherwise, we assume that \(\mathcal{C}= \mathsf {constants}(\mathcal{P})\). \( g \mathcal{P}\) denotes ground \(\mathcal{P}\), which means that \(\mathcal{P}\) contains exactly all the ground clauses with respect to the alphabet. \(\mathsf {atoms}(\mathcal{P})\) denotes the set of all atoms occurring in \( g \mathcal{P}\). If atom A is not the head of any clause in \(\mathcal{P}\), then A is undefined in \( g \mathcal{P}\). The set of all atoms that are undefined in \( g \mathcal{P}\) is \(\mathsf {undef}(\mathcal{P})\).

Table 2. \(\top \), \(\bot \), and U denote true, false, and unknown, respectively.

3.2 Three-Valued Łukasiewicz Logic

We consider the three-valued Łukasiewicz logic [26], for which the corresponding truth values are \(\top \), \(\bot \) and U, which mean true, false and unknown, respectively. A three-valued interpretation I is a mapping from formulas to the set of truth values \(\{ \top , \bot , \text{ U }\}\). The truth value of a given formula under I is determined according to the truth tables in Table 2. We represent an interpretation as a pair \(I = \langle I^\top ,I^\bot \rangle \) of disjoint sets of atoms where \(I^\top \) is the set of all atoms that are mapped to \(\top \) by I, and \(I^\bot \) is the set of all atoms that are mapped to \(\bot \) by I. Atoms, which do not occur in \(I^\top \cup I^\bot \), are mapped to U. Let \(I = \langle I^\top , I^\bot \rangle \) and \(J = \langle J^\top , J^\bot \rangle \) be two interpretations: \(I \subseteq J \) iff \( I^\top \subseteq J^\top \) and \( I^\bot \subseteq J^\bot . \) \(I(F) = \top \) means that a formula F is mapped to true under I. \(\mathcal{M}\) is a model of \( g \mathcal{P}\) if it is an interpretation, which maps each clause occurring in \( g \mathcal{P}\) to \(\top \). I is the least model of \( g \mathcal{P}\) iff for any other model J of \( g \mathcal{P}\) it holds that \(I \subseteq J\).

3.3 Reasoning with Respect to Least Models

Consider the following transformation for \(\mathcal{P}\): 1. Replace all clauses in \( g \mathcal{P}\) with the same head \(A \leftarrow body _1, A \leftarrow body _2, \dots \) by the single expression \(A \leftarrow body _1 \vee body _2, \vee \dots \). 2. Replace all occurrences of \(\leftarrow \) by \(\leftrightarrow \). The resulting set of equivalences is called the weak completion of \(\mathcal{P}\) (\(\mathsf {wc}\,\mathcal{P}\)). The model intersection property holds for weakly completed programs, which guarantees the existence of a least model for every \(\mathcal{P}\) [20]. Stenning and van Lambalgen [30] devised the following operator, which has been generalized for first-order programs in [19]: Let I be an interpretation in \(\varPhi _{\mathcal{P}}(I) = \langle J^\top , J^\bot \rangle \), where

$$ \begin{array}{@{}lc@{}l} J^\top &{} =&{} \{A \mid \!\text { there exists }A \leftarrow body \in g \mathcal{P}\text { and }I( body ) = \top \}, \\ J^\bot &{} =&{} \{A \mid \! A \not \in \mathsf {undef}(\mathcal{P})\text { and for all }A \leftarrow body \in g \mathcal{P}\text { we find that }I( body ) = \bot \}. \end{array} $$

As shown in [19] the least fixed point of \(\varPhi _{\mathcal{P}}\) is identical to the least model of the weak completion of \( g \mathcal{P}\) under three-valued Łukasiewicz logic (\(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}\)). Starting with \(I = \langle \emptyset , \emptyset \rangle \)\(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}\) is computed by iterating \(\varPhi _{\mathcal{P}}\). Given a program \(\mathcal{P}\) and a formula F\(\mathcal{P}\models _{ wcs }F\) iff \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}(F) = \mathbf {\top }\) for formula F.

3.4 Integrity Constraints

A set of integrity constraints \(\mathcal{I}\mathcal{C}\) comprises clauses of the form\(\text { U}\leftarrow body \), where \( body \) is a conjunction of literals. Given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\), \(\mathcal{P}\) satisfies \(\mathcal{I}\mathcal{C}\) iff for all\(\text { U}\leftarrow body \in \mathcal{I}\mathcal{C}\), we find that \(\mathcal{P}\models _{ wcs }\text{ U }\leftarrow body \) (i.e. \(\mathcal{P}\not \models _{ wcs } body \)).

3.5 Abduction

We extend two-valued abduction [23] for three-valued semantics. The set of abducibles \(\mathcal{A}_\mathcal{P}\) may not only contain facts but can also contain assumptions:

$$\begin{array}{@{}lll} \mathcal{A}_\mathcal{P}&= \{ A \leftarrow \top \mid A \in \mathsf {undef}(\mathcal{P}) \} \cup \{ A \leftarrow \bot \mid A \in \mathsf {undef}(\mathcal{P}) \}. \end{array} $$

Let \(\langle \mathcal{P}, \mathcal{A}_\mathcal{P}, \mathcal{I}\mathcal{C}, \models _{ wcs }\rangle \) be an abductive framework, \(\mathcal{E}\subset \mathcal{A}_\mathcal{P}\) and observation \(\mathcal{O}\) a non-empty set of literals.  

\(\mathcal{O}\) :

is explained by \(\mathcal{E}\) given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\) iff \(\mathcal{P}\cup \mathcal{E}\models _{ wcs }\mathcal{O}\) and \(\mathcal{P}\cup \mathcal{E}\models _{ wcs }\mathcal{I}\mathcal{C}\).

\(\mathcal{O}\) :

is explained given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\) iff there exists \(\mathcal{E}\) s.t. \(\mathcal{O}\) is explained by \(\mathcal{E}\) given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\).

  We assume that explanations are minimal, i.e. there is no other explanation \(\mathcal{E}^\prime \subset \mathcal{E}\) for \(\mathcal{O}\). We distinguish between skeptical and credulous reasoning in abduction as follows:  

F :

follows skeptically from \(\mathcal{P}\)\(\mathcal{I}\mathcal{C}\) and \(\mathcal{O}\) iff \(\mathcal{O}\) can be explained given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\), and for all minimal \(\mathcal{E}\) for \(\mathcal{O}\) given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\), it holds that \(\mathcal{P}\cup \mathcal{E}\models _{ wcs }F\).

F :

follows credulously from \(\mathcal{P}\),  \(\mathcal{I}\mathcal{C}\) and \(\mathcal{O}\) iff there exists a minimal \(\mathcal{E}\) for \(\mathcal{O}\) given \(\mathcal{P}\) and \(\mathcal{I}\mathcal{C}\), and it holds that \(\mathcal{P}\cup \mathcal{E}\models _{ wcs }F\).

In the following, we are interested in deriving skeptically entailed information. The entailment relation \(\models _{ wcs }^{s}\) is an abbreviation to express that a formula follows skeptically, i.e. \(\mathcal{P}, \mathcal{I}\mathcal{C}, \mathcal{O}\models _{ wcs }^{s}F\) denotes that F follows skeptically from \(\mathcal{P}, \mathcal{I}\mathcal{C}\) and \(\mathcal{O}\).

4 Six Principles on Quantified Statements

We introduce six principles for developing the representation of quantified statements and reasoning with respect to them, originally developed in [3]. Some are motivated by ideas from the area of Logic Programming and others are motivated by findings from Cognitive Science.

4.1 Licenses for Inferences (lice)

Stenning and van Lambalgen [30] propose to formalize conditionals in human reasoning not by inferences straight away, but rather by licenses for inferences. For instance, the conditional ‘if y(X) then z(X)’ is represented by the program, which consists of

$$\begin{array}{lll} z(X) \leftarrow y(X) \wedge \lnot ab _{yz}(X). \quad \quad ab _{yz}(X) \leftarrow \bot . \end{array} $$

The first clause states that ‘z(X) if y(X) and \(\lnot ab _{yz}(X)\)’. The second clause represents the closed-world assumption with respect to \( ab _{yz}(X)\), where \( ab _{yz}(X)\) is an abnormality predicate. We call this principle licenses for inferences (lice).

4.2 Negation by Transformation (trans)

The logic programs we consider under the Weak Completion Semantics do not allow heads of clauses to be negative literals. In order to represent a negative conclusion \(\lnot y(X)\), we introduce an auxiliary formula \(y'(X)\) together with the clause \( y(X) \leftarrow \lnot y'(X)\) and the integrity constraint\(\text { U}\leftarrow y(X) \wedge y'(X)\). This is a widely used technique in logic programming. Together with the principle introduced in Sect. 4.1 (lice), this additional clause is extended by the following two clauses:

$$ \begin{array}{lll} y(X) \leftarrow \lnot y'(X) \wedge \lnot ab _{nyy}(X). \quad \quad ab _{nyy}(X) \leftarrow \bot . \end{array} $$

Additionally, the integrity constraint\(\text { U}\leftarrow y(X) \wedge y'(X)\) states that an object cannot belong to both, y and \(y'\). We call this principle negation by transformation (trans).

4.3 Existential Import and Gricean Implicature (import)

Normally, we do not quantify over objects that do not exist. Accordingly, ‘all y are z’ implies ‘some y are z’, which is referred to as existential import and implied by Gricean implicature [17]. Existential import is assumed by the theory of mental models [21] or mental logic [29]. Likewise, humans require existential import for a conditional to be true [30]. Furthermore, the quantifier ‘some y are z’ often implies that ‘some y are not z’, which again is implied by the Gricean implicature [24]: Someone would not state ‘some y are z’ if that person knew that ‘all y are z’. As the person does not say ‘all y are z’ but ‘some y are z’, we assume that ‘not all y are z’, which in turn implies ‘some y are not z’. We call this principle existential import and Gricean implicature (import).

4.4 Unknown Generalization (unkGen)

Humans seem to distinguish between ‘some y are z’ and ‘some z are y’ [24]. However, if we would represent ‘some y are z’ by \(\exists X (y(X) \wedge z(X))\) then this is semantically equivalent to \(\exists X (z(X) \wedge y(X))\) because conjunction is commutative in first-order logic. Likewise, as we have discussed in Sect. 4.3, humans seem to distinguish between ‘some y are z’ and ‘all y are z’. Accordingly, if we only observe that an object belongs to y and z then we do not want to conclude both, ‘some y are z’ and ‘all y are z’. Therefore we introduce the following principle: If we know that ‘some y are z’ then there must not only be an object, which belongs to y and z (by Gricean implicature) but there must be another object, which belongs to y and for which it is unknown whether it belongs to z. We call this principle unknown generalization (unkGen).

4.5 No Derivation Through Double Negation (dNeg)

Under the Weak Completion Semantics, a positive conclusion can be derived from double negation within two conditionals. Consider the following two conditionals with each having a negative premise: If not x, then y. If not y then z. Additionally, assume that x is true. Let \(\mathcal{P}= \{ z \leftarrow \lnot y, \ y \leftarrow \lnot x, \ x \leftarrow \top \}\) be the program that encodes this information. The \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}\) is \(\langle \{ x, z \}, \{ y \} \rangle \): x is true because it is a fact and y is false because the negation of x is false. z is true by the negation of y. However, considering the results in [24], humans seem not to draw conclusions through double negatives. Accordingly, we block them with the abnormalities introduced by principle (import) in Sect. 4.1. We call this principle no derivation through double negation (dNeg).

4.6 Search for Alternative Models (searchAlt)

Consider again \(\textsc {S}_{ rich }\) and \(\textsc {S}_{ add }\): The premises are about things which contradict the conclusion. We assume that in case there seems no conclusion possible, humans might try to search for alternative models by explaining some part of the information that is presented. We call this principle Search for Alternative Models (searchAlt).

5 Representation of Quantified Statements as Programs

Based on the first five principles of the previous section, we encode the quantified statements in logic programs, where y and z will later be replaced by the properties of the corresponding objects. Note that, different to the principles in Sects. 4.1 to 4.5, principle (searchAlt) in Sect. 4.6 is not about the representation of the quantified statements but about the reasoning process, which will be discussed later. Note that the capital letters in brackets in the title of each of the following subsections, A, E, I and O are the classical abbreviations for the quantifiers All, No, Some and Some not.

5.1 All y are z (Ayz)

All y are z’ is represented by the program \(\mathcal{P}_{Ayz}\), which consists of the following clauses:

figure b

The least model of the weak completion of \(\mathcal{P}_{Ayz}\), \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{Iyz},\text { is }\langle \{y(o),z(o)\}, \{ ab _{yz}(o)\} \rangle \).

5.2 No y is z (Eyz)

Under FOL ‘No y is z’ is represented as \(\forall X (y(X) \rightarrow \lnot z(X))\), which is equivalent to \(\forall X (z(X) \rightarrow \lnot y(X))\). \(\mathcal{P}_{Eyz}\) consists of the following clauses:

figure c

Note that the last clause in \(\mathcal{P}_{Eyz}\) cannot be generalized to all X, because otherwise we allow conclusions by double negatives: principle \((\textsf {dNeg})\) states that we should block conclusions through double negatives. The least model of the weak completion of \(\mathcal{P}_{Eyz}\), \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{Eyz}\), is \(\langle \{y'(o),z(o)\}, \{ ab _{zny}(o), ab _{nyy}(o), y(o)\} \rangle \).

5.3 Some y are z (Iyz)

Some y are z’ is represented by the program \(\mathcal{P}_{Iyz}\):

figure d

\(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{Iyz}\) is \(\langle \{y(o_1), y(o_2), z(o_1)\}, \{ ab _{yz}(o_1)\} \rangle .\)

Nothing about \( ab _{yz}(o_2)\) is stated in \(\mathcal{P}_{Iyz}\). Accordingly, \(z(o_2)\) stays unknown in \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{Iyz}\).

5.4 Some y are Not z (Oyz)

Some y are not z’ is represented by the program \(\mathcal{P}_{Oyz}\):

figure e

The first four clauses as well as the integrity constraint are derived as in the program \(\mathcal{P}_{Eyz}\) except that object \(o_1\) is used instead of o and \( ab _{ynz}\) is restricted to \(o_1\) as in \(\mathcal{P}_{Iyz}\). The fifth clause of \(\mathcal{P}_{Oyz}\) is obtained by principle \((\textsf {unkGen})\). The last two clauses are not generalized to all objects for the same reason as discussed in Sect. 5.2: The generalization of \( ab _{nzz}\) to all objects would lead to conclusions through double negation in case there would be a second premise. The least model of the weak completion of \(\mathcal{P}_{Oyz}\), \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{Oyz}\), is \(\langle \{y(o_1), y(o_2), z'(o_1)\}, \{ ab _{ynz}(o_1), ab _{nzz}(o_1), ab _{nzz}(o_2), z(o_1) \} \rangle .\)

5.5 Entailment of the Quantified Statements

We specify when \(\textsf {A}yz\), \(\textsf {E}yz\), \(\textsf {I}yz\) or \(\textsf {O}yz\) are entailed by a model.

  • \(\mathcal{P}\,\models \, \textsf {A}yz\) iff there exists an object o such that \(\mathcal{P}\models _{ wcs }y(o)\) and for all objects o we find that if \(\mathcal{P}\models _{ wcs }y(o)\) then \(\mathcal{P}\models _{ wcs }z(o)\).

  • \(\mathcal{P}\,\models \, \textsf {E}yz\) iff there exists an object o such that \(\mathcal{P}\models _{ wcs }z(o)\) and for all objects o we find that if \(\mathcal{P}\models _{ wcs }z(o)\) then \(\mathcal{P}\models _{ wcs }\lnot y(o)\).

  • \(\mathcal{P}\,\models \, \textsf {I}yz\) iff there exists an object \(o_1\) such that \(\mathcal{P}\models _{ wcs }y(o_1) \wedge z(o_1)\) and there exists an object \(o_2\) such that \(\mathcal{P}\models _{ wcs }y(o_2)\) and \(\mathcal{P}\not \models _{ wcs }z(o_2)\).

  • \(\mathcal{P}\,\models \, \textsf {O}yz\) iff there exists an object \(o_1\) such that \(\mathcal{P}\models _{ wcs }y(o_1) \wedge \lnot z(o_1)\) and there exists an object \(o_2\) such that \(\mathcal{P}\models _{ wcs }y(o_2)\) and \(\mathcal{P}\not \models _{ wcs }\lnot z(o_2)\).

If nothing can be concluded, i.e. if \(\mathcal{P}\not \models \textsf {A}yz\), \(\mathcal{P}\not \models \textsf {E}yz\), \(\mathcal{P}\not \models \textsf {I}yz\) and \(\mathcal{P}\not \models \textsf {O}yz\), then principle (searchAlt) applies and we search for alternative models by trying to explain y. Later, y refers to the first property in the conclusion of the syllogism and z refers to the two properties left. For instance, consider \(\textsc {S}_{ dog }\): y refers to highly trained dogs (\( high\_trai \)) and z is either police dogs (\( pol\_dog \)) or vicious (\( vic \)). If nothing between either \( high\_trai \) and \( pol\_dog \) or between \( high\_trai \) and \( vic \) can be derived, we try to explain \( high\_trai \).

6 Modeling the Belief-Bias Effect

According to the observations made in Sect. 2, we model the belief-bias effect in two stages: (1) the belief can influence the representation, i.e. how the given information is understood, or (2) the belief can influence the reasoning, i.e. how new information is gained, if nothing can be derived. In the following, we model (1) with help of abnormalities, motivated by principle (lice). (2) is modeled by means of skeptical abduction, motivated by principle (searchAlt). The following four syllogisms are modeled according to the logic program representations proposed in Sects. 5.2 and 5.4.

6.1 No Belief-Bias Effect

\(\mathcal{P}_{ dog }\) represents the first two premises of \(\textsc {S}_{ dog }\) and consists of

figure f

\(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ dog } = \langle I^\top , I^\bot \rangle \), is as follows:

$$ \begin{array}{@{}lll} I^\top &{} = &{} \{ high\_trai (o_2), high\_trai (o_3), pol\_dog ^\prime (o_1), pol\_dog ^\prime (o_2), vic (o_1), vic (o_2) \}, \\ I^\bot &{} = &{} \{ pol\_dog (o_2), pol\_dog (o_1), ab _{ pol\_dog ^\prime }(o_1), ab _{ pol\_dog ^\prime }(o_2), ab _{ pol\_dog ^\prime }(o_3), \\ &{}&{} \;\; ab _{ pol\_dog }(o_1), ab _{ vic }(o_2) \}, \end{array} $$

Indeed, this model entails the Conclusion of \(\textsc {S}_{ dog }\), Some highly trained dogs are not police dogs: There exists an object, \(o_2\), such that \(\mathcal{P}_{ dog } \models _{ wcs } high\_trai (o_2) \wedge \lnot pol\_dog (o_2)\) and there exists another object, \(o_3\), such that \(\mathcal{P}_{ dog } \models _{ wcs } high\_trai (o_3) \text{ and } \mathcal{P}_{ dog } \not \models _{ wcs }\lnot pol\_dog (o_3).\) According to [15], \(\textsc {S}_{ dog }\) is logically valid and psychologically believable. No conflict arises either at the psychological or at the logical level. The majority validated the syllogism, which complies with what is entailed by \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ dog }\).

6.2 Belief-Bias Effect in Representation Stage

\(\mathcal{P}_{ vit }\) represents the first two premises of \(\textsc {S}_{ vit }\) and consists of

figure g

The corresponding \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ vit } = \langle I^\top , I^\bot \rangle \), is as follows:

$$ \begin{array}{@{}l@{}ll} I^\top &{} = &{} \{ vitamin (o_2), vitamin (o_3), inex (o_1), inex (o_2), nutri^\prime (o_1), nutri^\prime (o_2) \} \\ I^\bot &{} = &{} \{ nutri (o_1), nutri (o_2), ab _{ inex }(o_2), ab _{ nutri }(o_1), ab _{ nutri^\prime }(o_1), ab _{ nutri^\prime }(o_2),\\ &{}&{} ab _{ nutri^\prime }(o_3) \}, \end{array} $$

Indeed this model entails the Conclusion of \(\textsc {S}_{ vit }\), that Some vitamin tablets are not nutritional: There exists an object, \(o_2\), such that \(\mathcal{P}_{ vit } \models _{ wcs } vitamin (o_2) \wedge \lnot nutri (o_2)\) and there exists another object, \(o_3\), such that \(\mathcal{P}_{ vit } \models _{ wcs } vitamin (o_3) \text{ and } \mathcal{P}_{ vit } \not \models _{ wcs }\lnot nutri (o_3).\) The results of the psychological study in Table 1 indicate that there seemed to be two groups of participants: The group that validated the syllogism was not influenced by the bias with respect to nutritional things. Their understanding of the syllogism is reflected by \(\mathcal{P}_{ vit }\) and their conclusion complies with what is entailed by \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ vit }\). The participants who chose to invalidate the syllogism belong to the other group that has apparently been influenced by their belief. The belief bias occurred in the representation stage. Accordingly, we model this aspect with help of abnormality predicates as follows: Regarding both premises, it is commonly known that

$$\begin{aligned} \textit{The purpose of vitamin tablets is to aid nutrition.} \end{aligned}$$

This belief in the context of Premise 1 leads to

$$\begin{aligned}&{If\, something \,is\, a\, vitamin\, tablet,\, then \,it\, is\, abnormal.}\ ({regarding} \,\textsc {Premise} \,{1}\\&{of}\, S_{ vit }{)} \end{aligned}$$

We extend \(\mathcal{P}_{ vit }\) accordingly, which results in

$$ \mathcal{P}_{ vit }^{\mathsf {bias}} = \mathcal{P}_{ vit } \cup \{ ab_{nutri^\prime } (X) \leftarrow vitamin (X) \}.$$

Observe that \( ab _{ nutri^\prime }(X) \leftarrow vitamin (X)\) overrides \( ab _{ nutri^\prime }(X) \leftarrow \bot \) under the weak completion of \(\mathcal{P}_{ vit }^{\mathsf {bias}}\). \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_ vit ^\mathsf {bias}= \langle I^\top , I^\bot \rangle \) is

$$ \begin{array}{l@{}l@{}l} I^\top &{} = &{} \{ inex (o_1), inex (o_2), vitamin (o_2), vitamin (o_3), ab _{ nutri^\prime }(o_2), ab _{ nutri^\prime }(o_3) \}, \\ I^\bot &{} = &{} \{ nutri^\prime (o_2), nutri^\prime (o_3), ab _{ nutri }(o_1), ab _{ inex }(o_2) \}. \end{array} $$

In this case, the Conclusion of \(\textsc {S}_{ vit }\), that Some vitamin tablets are not nutritional, is not entailed. Actually, nothing is stated about the relation between vitamin tablets and them (not) being nutritional. However, as trivally Some vitamin tablets are inexpensive holds, principle (searchAlt) does not apply and we are done. According to [15], \(\textsc {S}_{ vit }\) is logically valid but psychologically unbelievable. There arises a conflict at the psychological level because we generally assume that the purpose of vitamin tablets is to aid nutrition. The participants who have been influenced by this belief did not validate the syllogism, which complies to the result above, as the Conclusion is not entailed by \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ vit }^{\mathsf {bias}}\) either.

6.3 Belief-Bias Effect in Reasoning Stage

\(\mathcal{P}_{ rich }\) represents the first two premises of \(\textsf {S}_{ rich }\) and consists of

figure h

Its least model of the weak completion, \(\langle I^\top , I^\bot \rangle \), is as follows:

$$ \begin{array}{l@{}l@{}l} I^\top &{} = &{} \{ hard\_wor (o_1), hard\_wor (o_2), mil^{\prime } (o_1), mil^{\prime } (o_2), rich (o_2), rich (o_3) \}, \\ I^\bot &{} = &{} \{ mil (o_1), mil (o_2), ab _{ hard\_wor }(o_2), ab _{ mil }(o_1), ab _{ mil^{\prime } }(o_1), ab _{ mil^{\prime } }(o_2),\\ &{}&{} ab _{ mil^{\prime } }(o_3) \}, \end{array} $$

and does not confirm the Conclusion of \(\textsc {S}_{ rich }\), that some millionaires are not rich people. Actually, the Conclusion in \(\textsc {S}_{ rich }\) states something which contradicts Premise 2, and cannot be about any of the previously introduced constant \(o_1\), \(o_2\) or \(o_3\). As nothing can be derived about the relation between \( mil \) and \( hard\_wor \) nor between \( mil \) and \( rich \), principle (searchAlt) applies: According to our background knowledge, we know that ‘normal’ millionaires exist, i.e. millionaires for whom we do not assume anything abnormal with respect to them being millionaires. Additionally, we cannot be sure that all millionaires are normal, i.e. we know that millionaires exist for whom we simply don’t know whether they are normal. We formulate this as an observation about two newly introduced constants, let’s say \(o_4\), representing a normal millionaire,Footnote 1 and \(o_5\), representing a millionaire for whom it is unknown whether he or she is normal:

$$ \mathcal{O}= \{ mil (o_4), \lnot ab _{ mil^{\prime } }(o_4), \lnot ab _{ mil }(o_4), mil (o_5)\}.$$

If we want to find an explanation for \(\mathcal{O}\) with respect to \(\mathcal{P}_{ mil }\), we can no longer assume that \(\mathcal{C}= \mathsf {constants}(\mathcal{P}_{ mil })\), because \(\mathcal{A}_{\mathcal{P}_{ mil }}\) does not contain any facts or assumptions about \(o_4\) and \(o_5\), as \(o_4\) and \(o_5\) do not occur in \(\mathcal{P}_{ mil }\). Therefore, we specify that the new set of constants under consideration is \(\mathcal{C}= \{o_1, o_2, o_3, o_4, o_5\}\). Given that \(\mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}_{ mil }) = \langle I^\top , I^\bot \rangle \) as defined above, \(\mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}_{ mil }^\mathcal{C}) = \langle I^\top , I^\bot \cup \{ ab _{ mil^{\prime } }(o_4), ab _{ mil^{\prime } }(o_5) \} \rangle .\) The set of abducibles, \( \mathcal{A}_{\mathcal{P}_{ mil }^\mathcal{C}}\), contains six facts and six assumptions about \(o_4\) and \(o_5\):

$$\begin{array}{lll} rich (o_4) \leftarrow \top . &{} \qquad ab _{ mil }(o_4) \leftarrow \top . &{} \qquad ab _{ hard\_wor }(o_4) \leftarrow \top . \\ rich (o_4) \leftarrow \bot . &{} \qquad ab _{ mil }(o_4) \leftarrow \bot . &{} \qquad ab _{ hard\_wor }(o_4) \leftarrow \bot . \\ rich (o_5) \leftarrow \top . &{} \qquad ab _{ mil }(o_5) \leftarrow \top . &{} \qquad ab _{ hard\_wor }(o_5) \leftarrow \top .\\ rich (o_5) \leftarrow \bot . &{} \qquad ab _{ mil }(o_5) \leftarrow \bot . &{} \qquad ab _{ hard\_wor }(o_5) \leftarrow \bot . \end{array} $$

We find six (minimal) explanations for \(\mathcal{O}= \{ mil (o_4), \lnot ab _{ mil^{\prime } }(o_4), \lnot ab _{ mil }(o_4), mil (o_5)\}\), where there are three from which the Conclusion of \(\textsc {S}_{ rich }\) does not follow. Consider one of them, \(\mathcal{E}= \{ ab _{ hard\_wor }(o_4) \leftarrow \top , ab _{ mil }(o_4) \leftarrow \bot , ab _{ mil }(o_5) \leftarrow \top \}\): Given that \(\mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}_{ mil }) = \langle I^\top , I^\bot \rangle \), \(\mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}_{ mil } \cup \mathcal{E})^\mathcal{C}= \langle J^\top , J^\bot \rangle \) is

$$ \begin{array}{l@{}l@{}l} J^\top &{} = I^\top \cup \{ ab _{ hard\_wor }(o_4), { mil (o_4)}, { mil (o_5)}, ab _{ mil }(o_5) \}, \\ J^\bot &{} = I^\bot \cup \{ { ab _{ mil }(o_4)}, { ab _{ mil^{\prime } }(o_4)}, { hard\_wor (o_4)}, { mil^{\prime } (o_4)}, { ab _{ mil^{\prime } }(o_5)} \}. \end{array} $$

According to the definition for skeptical abduction in Sect. 3.5, one explanation for which the Conclusion of \(\textsc {S}_{ rich }\), Some millionaires are not rich people, does not follow, is enough to show that the Conclusion does not follow skeptically from \(\mathcal{P}_{ mil }^\mathcal{C}\), \(\mathcal{I}\mathcal{C}\) and \(\mathcal{O}\). According to [15] this case is neither logically valid nor believable. Almost no one validated \(\textsc {S}_{ rich }\), which complies to the result above, as the Conclusion is not skeptically entailed by \(\mathcal{P}_{ mil }^\mathcal{C}\), \(\mathcal{I}\mathcal{C}\) and \(\mathcal{O}\) either.

6.4 Belief-Bias Effect in Representation and Reasoning Stage

\(\mathcal{P}_{ cig }\) represents the first two premises of \(\textsc {S}_{ cig }\) and consists of

figure i

It is commonly known that Cigarettes are addictive. This belief in the context of Premise 1 leads to

If something is a cigarette, then it is abnormal. (regarding Premise 1 of \(S_{ cig }\))

\(\mathcal{P}_{ cig }\) is extended accordingly. The new program is

$$ \begin{array}{lll} \mathcal{P}^{\mathsf {bias}}_{ cig }= & {} \mathcal{P}_{ cig } \cup \{ ab_{ add^\prime } (X) \leftarrow cig (X) \}. \end{array} $$

Observe that \( ab_{ add^\prime } (X) \leftarrow cig (X)\) overrides \( ab_{ add^\prime } (X) \leftarrow \bot \) under the weak completion of \(\mathcal{P}_{ cig }^{\mathsf {bias}}\). The least model of the weak completion of \(\mathcal{P}_{ cig }^{\mathsf {bias}}\), \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}_{ cig }^{\mathsf {bias}}\), is

$$ \langle \{ cig (o_2), cig (o_3), inex (o_1), inex (o_2) \}, \{ ab _{ add }(o_1), ab _{ inex }(o_2) \} \rangle .$$

Similarly to the previous syllogism, this model does not state anything about the Conclusion, that some addictive things are not cigarettes. Again, the Conclusion of \(\textsc {S}_{ cig }\) is about something, which cannot be \(o_1, o_2\) or \(o_3\). As nothing can be derived about the relation between \( add \) and \( inex \) nor between \( add \) and \( cig \), principle (searchAlt) applies: According to our background knowledge, we know that ‘normal’ addictive things exist, i.e. addictive things for which we do not assume anything abnormal with respect to them being addictive things. Additionally, we cannot be sure that all addictive things are normal, i.e. we know that addictive things exist for which we simply don’t know whether they are normal. We formulate this as an observation about two newly introduced constants, let’s say \(o_4\), representing normal addictive thingsFootnote 2 and \(o_5\) representing addictive things for which it is unknown whether they are normal:

$$ \mathcal{O}= \{ add (o_4), \lnot ab _{ add^\prime }(o_4), \lnot ab _{ add }(o_4), add (o_5) \}.$$

Let us define \(\mathcal{C}= \{ o_1, o_2, o_3, o_4, o_5\}\). \(\mathsf {lm}\,\mathsf {wc}\,\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }\) does not state anything about \(o_4\) nor \(o_5\): All atoms about \(o_4\) and \(o_5\) are unknown in this least model. Given \(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }\), the set of abducibles, \(\mathcal{A}_{\mathcal{P}_{ cig }^{\mathsf {bias},\mathcal{C}}}\) contains six facts and six assumptions about \(o_4\) and \(o_5\):

$$ \begin{array}{lll} cig (o_4) \leftarrow \top . &{} \qquad ab _{ add }(o_4) \leftarrow \top . &{} \qquad ab _{ inex }(o_4) \leftarrow \top . \\ cig (o_4) \leftarrow \bot . &{} \qquad ab _{ add }(o_4) \leftarrow \bot . &{} \qquad ab _{ inex }(o_4) \leftarrow \bot . \\ cig (o_5) \leftarrow \top . &{} \qquad ab _{ add }(o_5) \leftarrow \top . &{} \qquad ab _{ inex }(o_5) \leftarrow \top . \\ cig (o_5) \leftarrow \bot . &{} \qquad ab _{ add }(o_5) \leftarrow \bot . &{} \qquad ab _{ inex }(o_5) \leftarrow \bot . \end{array} $$

The only three (minimal) explanations are

$$ \begin{array}{lllllllll}&\,&\mathcal{E}_1 = \mathcal{E}' \cup \{ cig (o_5)\leftarrow \bot \}, \; \mathcal{E}_2 = \mathcal{E}' \cup \{ ab _{ inex }(o_5) \leftarrow \bot \}, \text{ and } \mathcal{E}_3 = \mathcal{E}' \cup \{ cig (o_5) \leftarrow \top \}, \end{array} $$

where \(\mathcal{E}' = \{ cig (o_4) \leftarrow \bot ,\ ab _{ add }(o_4) \leftarrow \bot ,\ ab _{ add }(o_5) \leftarrow \bot \}\). Given that \(\mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}^\mathsf {bias}_{ cig }) = \langle I^\top , I^\bot \rangle \) as defined above, the least models of the weak completion of \(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }\) together with the corresponding explanations, are as follows:

$$ \begin{array}{lll} \mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig } \cup \mathcal{E}_1) = &{} \langle I^\top \cup \{ { add (o_4)}, { add (o_5)} \}, \\ &{} \;\; I^\bot \cup \{ { cig (o_4)}, { inex (o_4)}, { ab _{ add }(o_4)}, { ab _{ add^\prime }(o_4)}, { add^\prime (o_4)}, \\ &{} \;\; { ab _{ add }(o_5)}, cig (o_5), ab _{ add^\prime }(o_5), inex (o_5), add^\prime (o_5) \} \rangle , \\ \mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig } \cup \mathcal{E}_2) = &{} \langle I^\top \cup \{ { add (o_4)}, { add (o_5)}, ab _{ inex }(o_5) \}, \\ &{} \; \; I^\bot \cup \{{ cig (o_4)}, { inex (o_4)}, { ab _{ add }(o_4)}, { ab _{ add^\prime }(o_4)}, { add^\prime (o_4)}, \\ &{} \;\; { ab _{ add }(o_5)}, inex (o_5), add^\prime (o_5) \} \rangle , \\ \mathsf {lm}\,\mathsf {wc}\,(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig } \cup \mathcal{E}_3) = &{} \langle I^\top \cup \{ { add (o_4)}, { add (o_5)}, cig (o_5), ab _{ add^\prime }(o_5) \}, \\ &{} \;\; I^\bot \cup \{ { cig (o_4)}, { inex (o_4)}, { ab _{ add }(o_4)}, { ab _{ add^\prime }(o_4)}, { add^\prime (o_4)}, \\ &{} \;\; { ab _{ add }(o_5)}, add^\prime (o_5) \} \rangle . \end{array} $$

The Conclusion of \(\textsc {S}_{ add }\), Some addictive things are not cigarettes, follows skeptically from \(\mathcal{P}_{ add }^{\mathsf {bias},\mathcal{C}}\) and \(\mathcal{O}\), as the following derivation follows from all explanations for \(\mathcal{O}\): There exists an object, \(o_4\), such that \(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }, \mathcal{O}\models _{ wcs }^{s} add (o_4) \wedge \lnot cig (o_4)\) and there exists another object, \(o_5\), such that \(\mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }, \mathcal{O}\models _{ wcs }^{s} add (o_5) \quad \text{ and } \quad \mathcal{P}^{\mathsf {bias},\mathcal{C}}_{ cig }, \mathcal{O}\not \models _{ wcs }^{s}~ cig (o_5).\) According to [15], \(\textsc {S}_{ cig }\) is classical logically invalid but psychologically believable and therefore causes a conflict: People are biased and search for a model that confirms their beliefs. This complies with what is entailed skeptically by \(\mathcal{P}_{ cig }^{\mathsf {bias},\mathcal{C}}\), \(\mathcal{I}\mathcal{C}\) and \(\mathcal{O}\).

Note that in this case we need the restriction that explanations are minimal, otherwise \(\mathcal{E}' \cup \{ ab _{ inex }(o_4) \leftarrow \top , \ cig (o_5) \leftarrow \bot \} \supset \mathcal{E}_1\) would be an explanation for \(\mathcal{O}\) as well, and we could not derive that the Conclusion of \(\textsc {S}_{ add }\) follows skeptically anymore.

7 Conclusion

By taking the principles presented in [3] as starting point and extending them with the additional principle Search for Alternative Models, we show how the belief-bias effect can be modeled by discussing the four cases of Evans et al.’s [15] syllogistic reasoning task. The belief-bias effect can be modeled in two stages: The first stage is where the belief bias seems to occur in the representational part of the syllogism, for instance in \(\textsf {S}_{ vit }\). In this case, the belief bias can be modeled by means of abnormality predicates. The belief bias in \(\textsf {S}_{ cig }\) seems to occur in the representational and the reasoning part of the syllogism. The reasoning part can be modeled with skeptical abduction. Additionally, as the last case shows, explanations are required to be minimal.

To the best of our knowledge, the syllogistic reasoning tasks discussed in the literature have never accounted for providing the option ‘I don’t know’ to the participants. As has been discussed in [27], participants who say that no valid conclusion follows, might have problems to actually find a conclusion easily, possibly meaning that they do not know the answer. The authors also point to [28], who suggested that if a conclusion is stated as being not valid this can mean that the reasoning process is exhausted. An experimental study, which allows the participants to distinguish between I don’t know and not valid, might give us more insights about their reasoning processes.