1 Introduction

In Carnielli and Rodrigues (2017) two paraconsistent and paracomplete formal systems were presented, the Basic Logic of Evidence (\({ BLE}\)) and the Logic of Evidence and Truth (\({ LET}_{J}\)). \({ BLE}\) ends up being equivalent to Nelson’s well-known logic N4 but has been conceived to express preservation of evidence instead of truth. Scenarios with conflicting evidence—that is, non-conclusive evidence for the truth and the falsity of \( \alpha \)—as well as scenarios with no evidence at all about \( \alpha \) are possible, so neither explosion nor excluded middle hold in \({ BLE}\)\({ LET}_{J}\) is an extension of \({ BLE}\) equipped with a classicality operator \({\circ }\). When \( \circ \alpha \) holds, classical negation—and so full classical logic—for \( \alpha \) is recovered. According to the intended interpretation, \( {\circ }\alpha \) in \({ LET}_{J}\) means that there is conclusive evidence for the truth or falsity of \( \alpha \), so the truth-value of \( \alpha \) has been established as true or false.

Let us call \( \vdash _C \) and \( \vdash _{BLE} \), respectively, the relation of logical consequence in classical logic and in BLE. Classical consequence is defined in terms of preservation of truth: \(\Gamma \vdash _C \alpha \) just in case there is no model M such that all propositions of \( \Gamma \) are true in M, but \( \alpha \) is not true in M. The intended interpretation of BLE, on the other hand, is not based on preservation of truth, but rather on preservation of evidence: \(\Gamma \vdash _{BLE} \alpha \) means that the availability of evidence for the premises in \( \Gamma \) implies that there is also evidence available for \( \alpha \). Classical logic and BLE, therefore, express different properties of propositions: truth and availability of evidence. The logic \({ LET}_{J}\), in its turn, is able to express preservation of evidence and preservation of truth—it ‘combines’, in one and the same formal system, the relations \( \vdash _C \) and \( \vdash _{BLE} \). The operator \({\circ }\) works like a context switch that divides propositions into those that have a classical and those that have a non-classical behavior, and \({ BLE}\) is the underlying logic of the latter.

Adequate valuation semantics and decision procedures for \({ BLE}\) and \({ LET}_{J}\) have been proposed. These semantics, however, are only able to express the fact that a given proposition \(\alpha \) has or does not have evidence available by attributing, respectively, the semantic value 1 or 0 to \(\alpha \). Evidence, thus, is treated from a purely qualitative point of view. A question that presents itself is whether the amount of evidence available for a given proposition \(\alpha \) could be quantified. Here we give a positive answer to this question.

The aim of this paper is to propose a probabilistic semantics for a modified version of \({ LET}_{J}\) obtained by dropping the implication symbol \(\rightarrow \) and adding a non-classicality operator \({\bullet }\) dual to \({\circ }\). While \({\circ }\alpha \) implies that \(\alpha \) behaves classically, a non-classical behavior of \(\alpha \) implies \({\bullet }\alpha \). The logic so obtained is an extension of the well-known logic of First-Degree Entailment (\({ FDE}\)), and we call it \({ LET}_{F}\), the Logic of Evidence and Truth based on \({ FDE}\). As well as \({ LET}_{J}\), \({ LET}_{F}\) is suitable to an intuitive reading in terms of evidence and truth.

In order to capture this idea of preservation of degrees of evidence a non-classical notion of probability will be employed. The probabilistic semantics proposed here follows the ideas presented in Bueno-Soler and Carnielli (2016, 2017). Let \(P(\alpha )=\epsilon \) mean that \(\epsilon \) is the measure of evidence available for \(\alpha \). We call a probabilistic scenario paracomplete when \(P(\alpha ) + P(\lnot \alpha ) < 1\), and paraconsistent when \(P(\alpha ) + P(\lnot \alpha ) > 1\). These two cases can be explained, respectively, as ‘too little information’ and ‘too much information’ about \(\alpha \).Footnote 1 In both cases, \(P(\circ \alpha ) < 1\), which means that the probability measures of \(\alpha \) and \(\lnot \alpha \) are not behaving classically. So, \(P(\circ \alpha ) < 1\) means that the information available about \(\alpha \) is not reliable, and something must be wrong. If \(P(\circ \alpha ) = 1\), standard probability is recovered for \( \alpha \).

With the purpose of understanding the probabilistic semantics proposed here better, we adopt a notion of information space instead of the standard notion of sample space. The intuitive idea is to collect all the relevant information about a proposition \(\alpha \) (or about a set of propositions \(\Gamma \)) and the corresponding measures of evidence. So, roughly speaking, an information space is constituted by propositions that represent evidence that can be non-conclusive, contradictory or incomplete, more reliable or less reliable, and sometimes conclusive (we return to this point in Sect. 4.3 below). Such a notion of information space requires a generalization of the notion of a partition, and consequently allows us to obtain generalized versions of standard results of probability theory such as total probability theorem and Bayes’ rule.Footnote 2

The remainder of this paper is organized in four sections. Section 2 is dedicated to the logic \({ FDE}\). It is shown that \({ FDE}\) is suited to an interpretation in terms of preservation of evidence. We also present adequate valuation semantics and a decision procedure for \({ FDE}\). In Sect. 3, \({ FDE}\) is extended to \({ LET}_{F}\), and an adequate semantics, a decision procedure, and some relevant results are presented and discussed. In Sect. 4, a probabilistic semantics for \({ LET}_{F}\) is defined, and paraconsistent and paracomplete versions of total probability theorems and Bayes’ rule are also presented and discussed. Finally, in Sect. 5, we discuss some points related to the topics of this paper that could be developed further.

2 \({ FDE}\) as a logic of preservation of evidence

The inference rules of \({ BLE}\) were obtained by asking whether an inference rule preserves evidence. Since evidence can be incomplete (no evidence at all) and contradictory (conflicting evidence), explosion and excluded middle do not hold. In \({ BLE}\), when \( \alpha \) (resp. \(\lnot \alpha \)) holds, the intended meaning is that there is evidence for the truth (resp. falsity) of \( \alpha \). Evidence that \( \alpha \) is true and evidence that \( \alpha \) is false are independent of each other, and are treated as such by the formal system. \({ BLE}\) can express the following four scenarios:

  1. 1.

    Only evidence that \(\alpha \) is true: \( \alpha \) holds, \(\lnot \alpha \) does not hold.

  2. 2.

    Only evidence that \(\alpha \) is false: \(\lnot \alpha \) holds, \( \alpha \) does not hold.

  3. 3.

    No evidence at all: neither \( \alpha \) nor \(\lnot \alpha \) hold.

  4. 4.

    Conflicting evidence: both \( \alpha \) and \(\lnot \alpha \) hold.Footnote 3

Evidence for a proposition \( \alpha \) is explained in Carnielli and Rodrigues (2017, Section 2) as reasons for believing in \( \alpha \), but these reasons may be non-conclusive or even wrong, and do not imply the truth of \( \alpha \), nor the belief in \( \alpha \). Thus, evidence is a notion weaker than truth in the sense that there may be evidence for a proposition \( \alpha \) even if \( \alpha \) is not true. Below, in Sect. 2.2.1, starting from the notion of information proposed by Dunn (2008), we explain evidence in terms of a (perhaps) non-conclusive justification added to a proposition \( \alpha \) or, as Fitting (2016b) puts it, “justifications that might be wrong”.Footnote 4 Notice that the notion of evidence encompasses non-conclusive as well as conclusive evidence, and the latter is evidence that establishes the truth-value of a proposition \( \alpha \).

The logic of First-Degree Entailment (\({ FDE}\)) is a paraconsistent and paracomplete propositional logic in a language with conjunction, disjunction, and negation, with no theorems nor bottom particles (cf. Anderson and Belnap 1963, 1975; Anderson et al. 1992; Belnap 1977a, b; Dunn 1976). \({ FDE}\) is a fragment of \({ BLE}\)/N4, obtained by dropping the implication symbol and the corresponding rules, and it can be interpreted in terms of preservation of evidence, as well as \({ BLE}\)–the four scenarios above clearly correspond to the four truth-values proposed by Belnap (1977a, b) (we return to this point in Sect. 2.2 below).Footnote 5

Definition 1

The Logic of First-Degree Entailment (\({ FDE}\)). Let \(L_{1}\) be a language with a denumerable set of sentential letters \(\left\{ p_{1}, p_{2}, p_{3},\ldots \right\} \), the set of connectives \(\left\{ \lnot , \wedge , \vee \right\} \), and parentheses. The set of formulas of \(L_{1}\) is obtained recursively in the usual way. The logic \({ FDE}\) is defined over the language \(L_{1}\) by the following natural deduction rules:

figure a

A deduction of \( \alpha \) from a set of premises \(\Gamma \), \( \Gamma \vdash _{FDE} \alpha \), is defined as follows: there is a derivation with conclusion \( \alpha \) and all uncancelled hypotheses in \( \Gamma \), and the definition of a derivation is the usual one for natural deduction systems (see e.g. van Dalen 2008, pp. 35–36).

Other deductive systems have already been presented for \({ FDE}\) (see Omori and Wansing 2017, Section 2.2), but the natural deduction system proposed here makes the symmetry between positive and negative rules explicit: \(\wedge I\) and \(\lnot \vee I\) are symmetrical, \(\vee E\) and \(\lnot \wedge E\) are symmetrical, and so on. This mirrors the fact that positive and negative evidence are primitive and non-complementary notions, but have symmetric deductive behavior: the rule \(\wedge I\) expresses the idea that when there is positive evidence available for both \( \alpha \) and \(\beta \), there is positive evidence for \(\alpha \wedge \beta \), while the rule \(\lnot \vee I\) means that when there is negative evidence available for both \( \alpha \) and \(\beta \), there is negative evidence for \(\alpha \vee \beta \).

Theorem 2

Reflexivity, monotonicity, transitivity, and compactness hold for \({ FDE}\).

Proof

These well-known properties of \({ FDE}\) can be easily proved by means of the natural deduction system above. \(\square \)

2.1 Valuation semantics for \({ FDE}\)

We now propose a non-deterministic valuation semantic for \({ FDE}\).

Definition 3

Valuation semantics for \({ FDE}\)

A valuation semantics for FDE is a collection of FDE-valuations defined as follows: a function \(v:L_1 \rightarrow \{0,1\}\) is a \({ FDE}\)-valuation if it satisfies the following clauses:

  1. v1.

    \(v(\alpha \wedge \beta )=1\) iff \(v(\alpha )=1\) and \(v(\beta )=1\),

  2. v2.

    \(v(\alpha \vee \beta )=1\) iff \(v(\alpha )=1\) or \(v(\beta )=1\),

  3. v3.

    \(v(\lnot (\alpha \wedge \beta ))\) = 1 iff \(v(\lnot \alpha ) = 1\) or \(v(\lnot \beta ) = 1\),

  4. v4.

    \(v(\lnot (\alpha \vee \beta )) = 1\) iff \(v(\lnot \alpha ) = 1\) and \(v(\lnot \beta ) = 1\),

  5. v5.

    \(v(\alpha ) = 1\) iff \(v(\lnot \lnot \alpha ) = 1\).

Definition 4

We say that a formula \( \alpha \) is a semantical consequence of \(\Gamma \), \(\Gamma \vDash _{FDE} \alpha \), iff for every valuation v, if \(v(\beta )=1\) for all \(\beta \in \Gamma \), then \(v(\alpha ) = 1\).

This semantics is sound and complete, and provides a decision procedure for \({ FDE}\). From now on, in this section, when there is no risk of ambiguity, we will just write \( \vdash \) and \( \vDash \) in the place of \( \vdash _{FDE}\) and \( \vDash _{FDE}\).

Theorem 5

Soundness. Let \(\Gamma \) be a set of formulas, and \( \alpha \) a formula of \({ FDE}\). So, \(\Gamma \vdash \alpha \text{ implies } \ \Gamma \vDash \alpha \).

Proof

The proof is routine. It shows that assuming there are sound derivations for the premise(s), the derivation obtained by the application of a rule is sound. \(\square \)

Theorem 6

Completeness. Let \(\Gamma \) be a set of formulas, and \( \alpha \) a formula of \({ FDE}\). Then \(\Gamma \vDash \alpha \text{ implies } \ \Gamma \vdash \alpha \).

Proof

Completeness can be proved by a Henkin-style proof. Given \( \Gamma \) and \( \alpha \) such that \(\Gamma \nvdash \alpha \), a set \(\Delta \) maximal w.r.t \( \alpha \) can be obtained in the usual way. So, the proof of the following propositions is straightforward:

  • v1\('.\) \(\alpha \wedge \beta \in \Delta \) iff \(\alpha \in \Delta \) and \(\beta \in \Delta \);

  • v2\('.\) \(\alpha \vee \beta \in \Delta \) iff \(\alpha \in \Delta \) or \(\beta \in \Delta \);

  • v3\('.\) \(\lnot (\alpha \wedge \beta ) \in \Delta \) iff \(\lnot \alpha \in \Delta \) or \(\lnot \beta \in \Delta \);

  • v4\('.\) \(\lnot (\alpha \vee \beta ) \in \Delta \) iff \(\lnot \alpha \in \Delta \) and \(\lnot \beta \in \Delta \);

  • v5\('.\) \(\alpha \in \Delta \) iff \(\lnot \lnot \alpha \in \Delta \).

Let v be the mapping from the language \(L_1\) to \(\{0,1\}\) defined as follows: for every \( \gamma \in L_1\), \(v(\gamma ) =1\) iff \(\gamma \in \Delta \). v is a valuation for \({ FDE}\) such that: for every \(\beta \in \Gamma \), \(v(\beta )=1\), since \( \Gamma \subseteq \Delta \); but \(v(\alpha )=0\), since \( \alpha \notin \Delta \) (\( \Delta \) is maximal w.r.t. \( \alpha \)). Therefore, \(\Gamma \nvDash \alpha \). \(\square \)

The valuation semantics proposed in Definition 3 is non-deterministic in the sense that the semantic value of a formula \(\lnot \alpha \) is not a function of the value of \( \alpha \). The possible values a formula can receive are given by quasi-matrices.Footnote 6 In Example 7 below, we illustrate how a quasi-matrix works.

Example 7

In \({ FDE}\):

  1. 1.

    \(p, \lnot p \vee q \nvDash q \);

  2. 2.

    \( p, \lnot (p \wedge q) \nvDash \lnot q \);

  3. 3.

    \( \lnot p \wedge \lnot q \lnot (p \vee q) \);

  4. 4.

    \( \lnot p \vee \lnot q \lnot (p \wedge q) \).

Proof

Consider the following quasi-matrix:

figure b

The valuations 13 and 14 above show that 1 is invalid, and the valuations 13 and 15 show that 2 is invalid. The remaining cases (De Morgan laws) are left to the reader. \(\square \)

Remark 8

The first four rows of the quasi-matrix above display the semantic values of the propositional variables and the negations of propositional variables that occur in the formulas at stake. The 5th and 6th rows are given by clauses v2 and v3 of Definition 3. Note that the semantic value of \(\lnot p\) is not determined by the value of p: the value of \( \lnot p \) bifurcates into 0 and 1 below \( v(p)=1\) and also below \(v(p)=0 \). So, being n the number of propositional variables of a given formula, the number of valuations is finite and bounded by \( 2^{2n} \). It is intuitively clear that the valuation semantics provides a decision procedure for \({ FDE}\). A detailed algorithm, however, will be presented elsewhere.

2.1.1 Some facts about \({ FDE}\)

Fact 9

Modus ponens and the deduction theorem do not hold in \({ FDE}\) for an implication \(\alpha \rightarrow \beta \) defined as \(\lnot \alpha \vee \beta \).

Proof

That disjunctive syllogism does not hold in \({ FDE}\) is shown by the following valuation: \(v(\alpha )=1\), \(v(\lnot \alpha )=1\), \(v(\beta )=0\). In order to show that the deduction theorem does not hold, suppose \(\Gamma , \alpha \vDash \beta \) implies \(\Gamma \vDash \lnot \alpha \vee \beta \). So, from \( \alpha \vDash \alpha \) we would get \( \vDash \lnot \alpha \vee \alpha \), but the latter is invalid in \({ FDE}\). \(\square \)

Fact 10

Grounding of contradictoriness

A compound formula \( \alpha \) is contradictory in a valuation v, i.e. \(v(\alpha ) = 1\) and \(v(\lnot \alpha )=1\), only if at least one propositional letter p that occurs in \( \alpha \) is contradictory in v.

Proof

Suppose there is a valuation v such that \(v(\alpha ) = v(\lnot \alpha ) = 1\). We prove that there is at least one propositional letter p in \( \alpha \) such that \(v(p) = v(\lnot p) = 1\). If \(\alpha = p\), clearly, \(v(\alpha ) = v(\lnot \alpha ) = v(p) = v(\lnot p) = 1\). The remaining cases are proved by induction on the complexity of \( \alpha \).

Case 1. \(\alpha = \lnot \lnot \beta \). I.H.: if \(v(\beta ) = v(\lnot \beta ) = 1\), there is a p in \(\beta \) such that \(v(p) = v(\lnot p) = 1\). Suppose \(v(\lnot \lnot \beta ) = v(\lnot \lnot \lnot \beta ) = 1\). So, by Definition 3, \(v(\beta ) = v(\lnot \beta ) =1\). The result follows by the inductive hypothesis.

Case 2. \(\alpha = \beta \wedge \gamma \). I.H.: if \(v(\beta ) = v(\lnot \beta ) = 1\), there is a p in \(\beta \) such that \(v(p) = v(\lnot p) = 1\); mutatis mutandis for \( \gamma \). Suppose \(v(\beta \wedge \gamma ) = v(\lnot (\beta \wedge \gamma )) = 1\). So, by Definition 3, \(v(\beta ) = v(\gamma ) =1\), and either \( v(\lnot \beta ) = 1\) or \( v(\lnot \gamma ) = 1\). By the inductive hypothesis, there is a p either in \(\beta \) or in \(\gamma \) such that \(v(p) = v(\lnot p) = 1\). The remaining cases are left to the reader. \(\square \)

Fact 11

Grounding of incompleteness. A compound formula \( \alpha \) is incomplete in a valuation v, i.e. \(v(\alpha )~=~0\) and \(v(\lnot \alpha )~=~0\), only if at least one propositional letter p that occurs in \( \alpha \) is incomplete in v.

Proof

Similar to the proof of Fact 10 above. \(\square \)

It is to be noted that the converse of Facts 10 and 11 do not hold: there may be a contradictory (resp. incomplete) atom p in a formula \( \alpha \) without \( \alpha \) being contradictory (resp. incomplete). Let \( \alpha \) be the formula \(p \vee q\) and consider the valuation v such that \(v(p) = v(\lnot p) = 1\), \(v(q) = 1\) and \(v(\lnot q) = 0\). In this case, p is a contradictory propositional letter, but \(p \vee q\) is not contradictory. On the other hand, in the valuation \(v(p) = v(\lnot p) = 0\), \(v(q) = 1\) and \(v(\lnot q) = 0\), p is a incomplete propositional letter, but \(p \vee q\) is not incomplete. Both valuations make \(v(p \vee q) = 1\) and \(v(\lnot (p \vee q)) = 0\).

2.2 Equivalence with Belnap’s four-valued and Dunn’s relational semantics

The valuation semantics proposed above, as expected, is equivalent both to the two-valued relational semantics proposed by Dunn (1976) and to the four-valued semantics presented by Belnap (1977b).Footnote 7

Definition 12

Dunn’s relational semantics for \({ FDE}\). A Dunn-interpretation for \({ FDE}\) is a relation \(\rho \) between the set of formulas of \({ FDE}\) and the values T and F, \( \rho \subseteq L \times \{T, F \}\), satisfying the following clauses:

  1. 1.

    \(\lnot \alpha \rho T \text{ iff } \alpha \rho F\),

  2. 2.

    \(\lnot \alpha \rho F \text{ iff } \alpha \rho T\),

  3. 3.

    \( (\alpha \wedge \beta ) \rho T \text{ iff } \alpha \rho T \text{ and } \beta \rho T \),

  4. 4.

    \( (\alpha \vee \beta ) \rho T \text{ iff } \alpha \rho T \text{ or } \beta \rho T \),

  5. 5.

    \( (\alpha \wedge \beta ) \rho F \text{ iff } \alpha \rho F \text{ or } \beta \rho F \),

  6. 6.

    \( (\alpha \vee \beta ) \rho F \text{ iff } \alpha \rho F \text{ and } \beta \rho F \).

Definition 13

A formula \(\alpha \) is a Dunn semantic consequence of \(\Gamma \), \(\Gamma \vDash _{D} \alpha \), iff for all Dunn-interpretations \(\rho \), if \(\beta \rho T\) for all \(\beta \in \Gamma \), then \(\alpha \rho T\).

Definition 14

Belnap’s four-valued semantics for \({ FDE}\). A four-valued interpretation for \({ FDE}\) is a function \(v_B\) from the set of formulas of \({ FDE}\) to the semantic values \(\{T, F, B, N\}\) satisfying the following matrices:

figure c

Definition 15

Let \(\mathbf{D } = \{T, B\}\) be the set of designated values of Belnap’s four-valued semantics. A formula \( \alpha \) is a four-valued semantic consequence of \(\Gamma \), \(\Gamma \vDash _{B} \alpha \), iff for all four-valued interpretations \(v_B\), if \(v_B(\beta ) \in \mathbf{D }\) for all \(\beta \in \Gamma \), then \(v_B(\alpha ) \in \mathbf{D }\).

The valuation semantics of Definition 3, Dunn’s relational semantics of Definition 12, and Belnap’s four-valued semantics of Definition 14 intend to represent four scenarios. Belnap (1977b, p. 11) explains the semantic values T, F, N, and B with the notion of a computer ‘being told’, so, these values mean, respectively, ‘just told true’, ‘just told false’, ‘told neither true nor false’, and ‘told both true and false’ (we return to this point in Sect. 2.2.1 below). Dunn (1976, p. 156) explains them in terms of subsets of \(\{T,F\}\), so a proposition can be related to \( \{T\} \), \( \{F\} \), \( \emptyset \), and \( \{T,F\} \). In Sect. 2 above we explained these four scenarios in terms of availability of evidence.

Although both the valuation semantics proposed here and Dunn’s relational semantics are bi-valued, and end up being equivalent, they have an essential difference: a valuation is a function from the set of formulas to \(\{0,1\}\), while a Dunn interpretation is a relation between the set of formulas and \(\{T,F\}\). In the latter, a formula can be related simultaneously to both T and F, when it is, in the Dunn–Belnap reading, both true and false, or not related to T nor F, when it is neither true nor false. But these three semantics, as expected, validate the same inferences, i.e. \(\Gamma \vDash _{FDE}\alpha \) iff \(\Gamma \vDash _D\alpha \) iff \(\Gamma \vDash _B\alpha \).

Definition 16

(Dunn interpretation induced by an \({ FDE}\)-valuation) Given a \({ FDE}\)-valuation v, we define a Dunn-interpretation \(\rho _{v}\), based on v, as follows:

  • \(\alpha \rho _{v} T\) iff \(v(\alpha )=1\)

  • \(\alpha \lnot \rho _{v} T\) iff \(v(\alpha )=0\)

  • \(\alpha \rho _{v} F\) iff \(v(\lnot \alpha )=1\)

  • \(\alpha \lnot \rho _{v} F\) iff \(v(\lnot \alpha )=0\)

Definition 17

(\({ FDE}\)-valuation induced by a Dunn-interpretation) Given a Dunn-interpretation \(\rho \), we define a \({ FDE}\)-valuation \(v_{\rho }\), based on \(\rho \), as follows:

  • \(v_{\rho }(\alpha )=1\) iff \(\alpha \rho T\)

  • \(v_{\rho }(\alpha )=0\) iff \(\alpha \lnot \rho T\)

  • \(v_{\rho }(\lnot \alpha )=1\) iff \(\alpha \rho F\)

  • \(v_{\rho }(\lnot \alpha )=0\) iff \(\alpha \lnot \rho F\)

Lemma 18

Given an \({ FDE}\)-valuation v, then \(\rho _{v}\) is a Dunn-interpretation.

Proof

We have to prove that \(\rho _{v}\) is a Dunn’s relational semantics as in Definition 12.

  1. 1.

    \(\lnot \alpha \rho _{v}T\) iff \(v(\lnot \alpha )=1\) iff \(\alpha \rho _{v}F\)

  2. 2.

    \(\lnot \alpha \rho _{v}F\) iff \(v(\lnot \lnot \alpha )=1\) iff \( v(\alpha )=1 \) iff \(\alpha \rho _{v}T\)

  3. 3.

    \((\alpha \wedge \beta )\rho _{v}T\) iff \(v(\alpha \wedge \beta )=1\) iff \(v(\alpha )=1\) and \(v(\beta )=1\) iff \(\alpha \rho _{v}T\) and \(\beta \rho _{v}T\)

  4. 4.

    \((\alpha \vee \beta )\rho _{v}T\) iff \(v(\alpha \vee \beta )=1\) iff \(v(\alpha )=1\) or \(v(\beta )=1\) iff \(\alpha \rho _{v}T\) or \(\beta \rho _{v}T\)

  5. 5.

    \((\alpha \wedge \beta )\rho _{v}F\) iff \(v(\lnot (\alpha \wedge \beta ))=1\) iff \(v(\lnot \alpha )=1\) or \(v(\lnot \beta )=1\) iff \(\alpha \rho _{v}F\) or \(\beta \rho _{v}F\)

  6. 6.

    \((\alpha \vee \beta )\rho _{v}F\) iff \(v(\lnot (\alpha \vee \beta ))=1\) iff \(v(\lnot \alpha )=1\) and \(v(\lnot \beta )=1\) iff \(\alpha \rho _{v}F\) and \(\beta \rho _{v}F\) \(\square \)

Lemma 19

Given a Dunn-interpretation \(\rho \), then \(v_\rho \) is a \({ FDE}\)-valuation.

Proof

We have to prove that \(v_{\rho }\) is a \({ FDE}\)-valuation as in Definition 3.

  1. 1.

    \(v_{\rho }(\alpha \wedge \beta )=1\) iff \((\alpha \wedge \beta )\rho T\) iff \(\alpha \rho T\) and \(\beta \rho T\) iff \(v_{\rho }(\alpha )=1\) and \(v_{\rho }(\beta )=1\)

  2. 2.

    \(v_{\rho }(\alpha \vee \beta )=1\) iff \((\alpha \vee \beta )\rho T\) iff \(\alpha \rho T\) or \(\beta \rho T\) iff \(v_{\rho }(\alpha )=1\) or \(v_{\rho }(\beta )=1\)

  3. 3.

    \(v_{\rho }(\lnot (\alpha \wedge \beta ))=1\) iff \((\alpha \wedge \beta )\rho F\) iff \(\alpha \rho F\) or \(\beta \rho F\) iff \(v_{\rho }(\lnot \alpha )=1\) or \(v_{\rho }(\lnot \beta )=1\)

  4. 4.

    \(v_{\rho }(\lnot (\alpha \vee \beta ))=1\) iff \((\alpha \vee \beta )\rho F\) iff \(\alpha \rho F\) and \(\beta \rho F\) iff \(v_{\rho }(\lnot \alpha )=1\) and \(v_{\rho }(\lnot \beta )=1\)

  5. 5.

    \(v_{\rho }(\alpha )=1\) iff \((\alpha )\rho T\) iff \(\lnot \alpha \rho F\) iff \(\lnot \lnot \alpha \rho T\) iff \(v_{\rho }(\lnot \lnot \alpha )=1\) \(\square \)

Lemma 20

The valuation semantics (Definition 3) and Dunn-interpretation (Definition 12) are equivalent, that is, given a valuation semantics v there exists a Dunn-interpretation \(\rho _{v}\) such that

  • \(v_{\rho }(\alpha )=1\) iff \(\alpha \rho T\)

  • \(v_{\rho }(\alpha )=0\) iff \(\alpha \lnot \rho T\)

  • \(v_{\rho }(\lnot \alpha )=1\) iff \(\alpha \rho F\)

  • \(v_{\rho }(\lnot \alpha )=0\) iff \(\alpha \lnot \rho F\)

for any proposition \(\alpha \); and vice-versa, given a Dunn-interpretation \(\rho \), there exists a valuation \(v_{\rho }\) such that:

  • \(\alpha \rho _{v} T\) iff \(v(\alpha )=1\)

  • \(\alpha \lnot \rho _{v} T\) iff \(v(\alpha )=0\)

  • \(\alpha \rho _{v} F\) iff \(v(\lnot \alpha )=1\)

  • \(\alpha \lnot \rho _{v} F\) iff \(v(\lnot \alpha )=0\)

for any proposition \(\alpha \).

Proof

Immediate from Lemmas 18 and 19 above. \(\square \)

Lemma 21

The valuation semantics (Definition 3) and Belnap’s four-valued semantics for \({ FDE}\) (Definition 14) are equivalent.

Proof

It follows from Lemma 20 and the well-known fact that Dunn’s and Belnap’s semantics are equivalent. \(\square \)

Theorem 22

The valuation semantics, the Dunn interpretation and the Belnap interpretation define equivalent notions of logical consequence: \(\Gamma \vDash _{FDE} \alpha \) iff \(\Gamma \vDash _{B} \alpha \) iff \(\Gamma \vDash _{D} \alpha \).

Proof

It follows from Lemmas 20 and 21. \(\square \)

2.2.1 On paraconsistency, evidence, and information

\({ FDE}\) is the well-known and widely studied ‘useful four-valued logic’ proposed by Belnap and Dunn as the underlying logic of an artificial information processor, i.e. a computer, capable of dealing with information received from different sources that are not entirely reliable (cf. Belnap 1977a, b; Dunn 1976). The semantic value Both is intended to represent the circumstance in which there is conflicting information about \( \alpha \), i.e. both \( \alpha \) and \( \lnot \alpha \) hold, and None is intended to represent the circumstance in which there is no information at all about \( \alpha \), i.e. neither \( \alpha \) nor \(\lnot \alpha \) holds.

When Belnap explains these four values, he talks about a computer ‘being told’ that a proposition \( \alpha \) is true, or false. The computer should be able to compute the values of complex propositions and draw inferences from the received information, but it “can only accept and report information without divesting itself of it” (Belnap 1977b, p. 9). Of course, contradictory information stored in a database should not be taken as true, as Belnap (1977a, p. 47) remarks that

these sentences have truth-values independently of what the computer has been told; but who can gainsay that the computer cannot use the actual truth-value of the sentences in which it is interested? All it can possibly use as a basis for inference is what it knows or believes, i.e., what it has been told.

The computer, when asked, must provide information based only on what it has been told, otherwise “we would have no way of knowing that its data-base harbored contradictory information” (Belnap 1977b, p. 9).Footnote 8

This notion of ‘a computer being told’ is clearly weaker than truth, since a computer may be told that \( \alpha \) is true even if it is not the case. So, Belnap is not really talking about truth simpliciter. On the other hand, Dunn (1976, p. 157) seems not to be totally comfortable with the interpretation of \({ FDE}\) in terms of the simultaneous truth of \( \alpha \) and \(\lnot \alpha \):

Do not get me wrong—I am not claiming that there are sentences which are in fact both true and false. I am merely pointing out that there are plenty of situations where we suppose, assert, believe, etc., contradictory sentences to be true, and we therefore need a semantics which expresses the truth conditions of contradictions in terms of the truth values that the ingredient sentences would have to take for the contradictions to be true.

Indeed, we should consider Dunn’s relational semantics as a façon de parler, rather than a claim that true contradictions are possible. Obviously, the simultaneous attribution of the semantic value True to a pair of propositions \(\alpha \) and \(\lnot \alpha \) is not to be understood as an acceptance of dialetheism. It is worth noting that at the time Belnap’s and Dunn’s papers were published, although there were already several paraconsistent formal systems available, the conceptual discussion about the nature of contradictions accepted by paraconsistent logics was still in its beginnings. It was a ‘lateral issue’ that had not yet been brought to the center of debate.

That the four values represented by Belnap–Dunn’s semantics correspond to the four scenarios of availability of evidence the logic \({ BLE}\) expresses has been shown in Sect. 2.2 above. The notions of evidence and information, indeed, are akin to each other, and both are well-suited to a non-dialetheist reading of paraconsistency. Let us take a closer look at these two notions.

In Carnielli and Rodrigues (2017, Section 2) the notion of evidence for a proposition \(\alpha \) was explained as ‘reasons for believing and/or accepting \(\alpha \)’. Evidence, when conclusive, gives support to the truth (or falsity) of \(\alpha \), and thus it has to do with the justification of \(\alpha \) (or \(\lnot \alpha \)). The idea behind the recovery operator \({\circ }\), introduced in Sect. 3 below, is that if there is conclusive evidence for the truth, or falsity, of a proposition \( \alpha \), then \( \alpha \) is subjected to classical logic. But evidence can be non-conclusive, and so there may be conflicting evidence for a proposition \(\alpha \). Besides being weaker than truth, evidence does not imply belief: there may be evidence for \(\alpha \), an agent may be aware of such evidence but still does not believe in \(\alpha \). If there is non-conclusive evidence for \(\alpha \), it means that there is some degree of justification for \( \alpha \) that, however, is not conclusive and might be wrong.Footnote 9

Dunn (2008, p. 589) explains a ‘bare-boned’ notion of information as:

what is left from knowledge when you subtract, justification, truth, belief, and any other ingredients such as reliability that relate to justification. Information is, as it were, a mere “idle thought.” Oh, one other thing, I want to subtract the thinker. (...) Anyone who has searched for information on the Web does not have to have this concept drummed home. So much of what we find on the Web has no truth or justification, and one would have to be a fool to believe it (...) [Information] is something like a Fregean “thought,” i.e., the “content” of a belief that is equally shared by a doubt, a concern, a wish, etc.

Information, so understood, is what is expresses by a proposition, indeed similar to a Fregean thought but without its platonic ingredient. It is objective, does not imply belief, does not need to be true. The difference between this bare-boned notion of information and the notion of non-conclusive evidence is that the latter has an epistemic ingredient that is lacking by the former. So, we can characterize non-conclusive evidence as bare-boned information plus a justification that might be wrong. Indeed, situations in which we have something that may be or may be not a justification for some proposition \(\alpha \) are quite common, and there is nothing wrong in saying that evidence, conclusive or non-conclusive, is still information: a proposition \( \alpha \) is information, as well as the claim that \( \alpha \) has been established as true. The notion of information is thus more general than evidence. It is not surprising, therefore, that both \({ BLE}\) and \({ FDE}\) are suitable to a non-dialetheist interpretation in terms of evidence and information.

3 Extending \({ FDE}\) to a logic of evidence and truth

\({ FDE}\) will now be extended to the logic \({ LET}_{F}\), in a similar way to what was done with \({ BLE}\) obtaining \({ LET}_{J}\) in Carnielli and Rodrigues (2017). Both \({ LET}_{J}\) and \({ LET}_{F}\) are Logics of Formal Inconsistency and Undeterminedness (LFIUs) (cf. Carnielli and Rodrigues 2017; Carnielli et al. 2019; Marcos 2005). In LFIs,

  • \( \alpha , \lnot \alpha \nvdash \beta \), while \(\circ \alpha , \alpha , \lnot \alpha \vdash \beta \),

and in LFUs,

  • \( \nvdash \alpha \vee \lnot \alpha \), while \( \circ \alpha \vdash \alpha \vee \lnot \alpha \).Footnote 10

When \( \circ \alpha \) holds, and so excluded middle and explosion are valid, we say that \( \alpha \) is classical. For this reason, in LFIUs, like the logics \({ LET}_{J}\) and \({ LET}_{F}\), we say that \({\circ }\) is a classicality operator.

Like \({ BLE}\), the logic \({ FDE}\), interpreted from the viewpoint of preservation of evidence, is not able to express preservation of truth. Indeed, none of the semantics presented for \({ FDE}\) in Sect. 2.1 can distinguish a context (i) where there there is non-conclusive evidence for \( \alpha \), so \( \alpha \) has not been established as true, but no evidence for \( \lnot \alpha \). from another context (ii) where there is conclusive evidence for \( \alpha \) and so \( \alpha \) has been established as true, and \( \lnot \alpha \) does not hold. In both (i) and (ii), \(\alpha \) and \(\lnot \alpha \) receive respectively the values 1 and 0 by the valuation semantics (Definition 3), or the values T and F by the Belnap’s four valued semantics (Definition 14), and so we cannot distinguish between (i) and (ii). The logic \({ LET}_{F}\), on the other hand, is able to distinguish these contexts.

Definition 23

The Logic of Evidence and Truth based on \({ FDE}\) (\({ LET}_{F}\)). Let \(L_{2}\) be a language with a denumerable set of sentential letters \(\left\{ p_{1}, p_{2}, p_{3},\ldots \right\} \), the set of connectives \(\left\{ {\circ }, {\bullet }, \lnot , \wedge , \vee , \right\} \) and parentheses. The set of formulas of \(L_{2}\) is obtained recursively in the usual way. The logic \({ LET}_{F}\) is defined over the language \(L_{2}\) by adding the following rules to the natural deduction system of \({ FDE}\) (Definition 1):

figure d

A deduction of \( \alpha \) from a set of premises \(\Gamma \) in \({ LET}_{F}\), \( \Gamma \vdash _{LET_F} \alpha \), is defined as follows: there is a derivation with conclusion \( \alpha \) and all uncancelled hypotheses in \( \Gamma \). The definition of a derivation is the usual one for natural deduction systems (see e.g. van Dalen 2008, pp. 35–36).

Theorem 24

The following properties hold for \({ LET}_{F}\):

  1. 1.

    Reflexivity: if \(\alpha \in \Gamma \), then \(\Gamma \vdash _{LET_F} \alpha \);

  2. 2.

    Monotonicity: if \(\Gamma \vdash _{LET_F} \beta \), then \(\Gamma , \alpha \vdash _{LET_F} \beta \), for any \(\alpha \);

  3. 3.

    Transitivity (cut): if \(\Delta \vdash _{LET_F} \alpha \) and \(\Gamma , \alpha \vdash _{LET_F} \beta \), then \(\Delta ,\ \Gamma \vdash _{LET_F} \beta \);

  4. 4.

    Compactness: if \(\Gamma \vdash _{LET_F} \alpha \), then there is \(\Delta \subseteq \Gamma \), \(\Delta \) finite such that \(\Delta \vdash _{LET_F} \alpha \).

Proof

Straightforward, from the definition of a deduction of \(\alpha \) from premises in \(\Gamma \) in \({ LET}_{F}\). \(\square \)

Fact 25

The following rules hold in \({ LET}_{F}\):

Proof

We prove \({{\bullet }}R1\). The proof of \({{\bullet }}R2\) is left to the reader.

\(\square \)

3.1 On the connectives \({\circ }\) and \({\bullet }\)

The rules \(PEM^{\circ }\) and \(EXP^{\circ }\) recover classical logic for propositions in the scope of \({\circ }\) (this claim will be made precise by Fact 31 below). As well as \({ LET}_{J}\), \({ LET}_{F}\) is suitable to an intuitive reading in terms of different contexts concerned with preservation of evidence and preservation of truth. But unlike \({ LET}_{J}\), \({ LET}_{F}\) has a non-classicality operator \({\bullet }\), dual to the classicality operator \({\circ }\). This duality is made clear by the rules above (Fact 25): R1 is the dual of \( EXP^{\circ } \), and R2 is the dual of \( PEM^{\circ }. \)Footnote 11 While \( {\circ }\alpha \) implies a classical behavior for \( \alpha \), a non-classical behavior of \( \alpha \) implies \( {\bullet }\alpha \). Notice that: (i) \( {\circ }\alpha \) does not imply \(\alpha \), rather, it implies that one and at most one between \( \alpha \) and \( \lnot \alpha \) holds; (ii) \({\bullet }\alpha \) does not imply that \( \alpha \) and \(\lnot \alpha \) hold; indeed, according to R2, if both \( \alpha \) and \(\lnot \alpha \) do not hold, \({\bullet }\alpha \) holds.

Strictly speaking, \( {\circ }\alpha \) recovers classical logic for \( \alpha \). The intended interpretation of \( {\circ }\alpha \) is that there is conclusive evidence for \( \alpha \) or \( \lnot \alpha \), and so the truth-value of \( \alpha \) is conclusively established as true or false. On the other hand, if evidence for \( \alpha \) is non-conclusive, or it is contradictory, or there is no evidence at all, then \({\bullet }\alpha \) holds. The rule Cons prohibits the circumstance such that there is and there is not conclusive evidence for \( \alpha \), while Comp expresses the fact that either there is or there is not conclusive evidence for \( \alpha \).

Since \({ LET}_{F}\) distinguishes conclusive from non-conclusive evidence, it is able to express the following six scenariosFootnote 12:

When \({\bullet }\alpha \) holds, four scenarios of non-conclusive evidence can be expressed:

  1. 1.

    Only evidence that \( \alpha \) is true: \( \alpha \) holds, \(\lnot \alpha \) does not hold.

  2. 2.

    Only evidence that \( \alpha \) is false: \(\lnot \alpha \) holds, \( \alpha \) does not hold.

  3. 3.

    No evidence at all: both \( \alpha \) and \(\lnot \alpha \) do not hold.

  4. 4.

    Conflicting evidence: both \( \alpha \) and \(\lnot \alpha \) hold.

    When \(\circ \alpha \) holds, two scenarios can be expressed, tantamount to classical truth and falsity:

  5. 5.

    Conclusive evidence for \( \alpha \): \( \alpha \) is true (\( {\circ }\alpha \wedge \alpha \) holds).

  6. 6.

    Conclusive evidence for \(\lnot \alpha \): \( \lnot \alpha \) is true (\( {\circ }\alpha \wedge \lnot \alpha \) holds).

Of course, a scenario with conclusive evidence for both \( \alpha \) and \(\lnot \alpha \) is not allowed, since it would imply that \( \alpha \) is true and false simultaneously. Indeed, if classical logic holds for \( \alpha \), it cannot be that there is any residual conflicting evidence for \( \alpha \) and \( \lnot \alpha \).

3.2 Valuation semantics for \({ LET}_{F}\)

Definition 26

A valuation semantics for \({ LET}_{F}\) is obtained by adding the following clauses to the valuation semantics of \({ FDE}\) (Definition 3):

  1. v6.

    \( v({\bullet }\alpha ) =~1\) iff \( v({\circ }\alpha ) =~0\),

  2. v7.

    If \(v(\circ \alpha ) = 1\), then \(v(\alpha ) = 1\) if and only if \(v(\lnot \alpha ) = 0\).

Definition 27

We say that a formula \( \alpha \) is a semantical consequence of \(\Gamma \), \(\Gamma \vDash _{LET_{F}} \alpha \) iff for every valuation v, if \(v(\beta )=1\) for all \(\beta \in \Gamma \), then \(v(\alpha ) = 1\).

The valuation semantics given above in Definition 26 is sound, complete, and provides a decision procedure for \({ LET}_{F}\). From now on, when there is no risk of ambiguity, we will just write \( \vdash \) and \( \vDash \) in the place of \( \vdash _{LET_{F}}\) and \( \vDash _{LET_{F}}\).

Theorem 28

Soundness and completeness of \({ LET}_{F}\) w.r.t. the valuation semantics: \( \Gamma \vDash \alpha \) iff \(\Gamma \vdash \alpha \).

Proof

In order to prove completeness, the proof of Theorem 6 has to be extended to include clauses \(6'\) and \(7'\) below:

  • v6\('.\) \(\circ \alpha \in \Delta \) iff \({\bullet }\alpha \notin \Delta \),

  • v7\('.\) \(\circ \alpha \in \Delta \) implies \(\lnot \alpha \in \Delta \) iff \(\alpha \notin \Delta \).

For soundness, it can be shown that rules Cons, Comp, \(EXP^{\circ }\), and \(PEM^{\circ }\) are sound with respect to clauses 6 and 7 of Definition 26 above. Details are left to the reader. \(\square \)

The quasi-matrix below displays the behavior of the connectives \({\circ }\) and \({\bullet }\) in \({ LET}_{F}\).

figure e

The first two lines display the possible values of p and \( \lnot p \). The connectives \({\circ }\) and \({\bullet }\) are primitive and unary, but the semantic values of \({\circ }p\) and \({\bullet }p\) depend (non-deterministically) on the semantic values of p and \(\lnot p\). When \( v(p)=1\) and \(v(\lnot p)=0 \), \( v(p)=0\) and \(v(\lnot p)=1 \), the value of \( {\circ }p \) and \( {\bullet }p \) bifurcates into 0 and 1. This expresses the fact that \( {\circ }p \) is undetermined in \({ LET}_{F}\) when \( v(p) \ne v(\lnot p) \), as explained in page 14 above. In terms of evidence, valuations \( v_1 \) and \( v_6 \) show, respectively, that no evidence at all, as well as conflicting evidence, implies \( v({\bullet }p) = 1 \) and \( v({\circ }p) = 0 \). But if only one holds among p and \(\lnot p\) (valuations \( v_2 \) to \( v_5 \)), then \( v({\bullet }p)\) and \( v({\circ }p)\) are left undetermined. The rationale of this is that in order to say that p is true, or false, only the information that there is evidence for the truth, or for the falsity, of p is not enough. Something else is needed, namely, the information that such evidence is conclusive.Footnote 13

In Example 29 below we illustrate how quasi-matrices work in \({ LET}_{F}\).

Example 29

In \({ LET}_{F}\):

  1. 1.

    \( p \vee \lnot p \nvDash {\circ }p \)

  2. 2.

    \({\bullet }p \nvDash p \wedge \lnot p\)

  3. 3.

    \({\circ }p, p, \lnot p \vee q \vDash q \);

  4. 4.

    \( {\circ }p, p, \lnot (p \wedge q) \vDash \lnot q \);

Proof

Consider the following quasi-matrix (divided into two parts):

figure f

Item 1: since \( v_{24}(p)=v_{24}(\lnot p) =1 \), \(v_{24}(p \vee \lnot p) =1 \), but \(v_{24}({\circ }p)=0 \). Item 2: \(v_{1}({\bullet }p)=1 \), but \(v_{1}(p \wedge \lnot p) =0 \), since \(v_{1}( p)=v_{1}(\lnot p)=0 \). For items 3 and 4, it is easy to check that there is no valuation v such that the premises receive the value 1 but the conclusion receives 0 in v (compare with items 1 and 2 of Fact 7). \(\square \)

Remark 30

The 7th row of the quasi-matrix above is given by clause v7 and the 8th by clause v6 of Definition 26. A quasi-matrix for \({ LET}_{F}\) is finite, and similarly to \({ FDE}\) (see Remark 8), it is intuitively clear that the valuation semantics provides a decision procedure for \({ LET}_{F}\). A detailed algorithm will be presented elsewhere.

3.3 Some facts about \({ LET}_{F}\)

Fact 31 below shows how the operator \({\circ }\) recovers classical logic in \({ LET}_{F}\).

Fact 31

Recovering classical logic in \({ LET}_{F}\). Suppose \(\circ \lnot ^{n_1}\alpha _1, \circ \lnot ^{n_2}\alpha _2, \dots , \circ \lnot ^{n_m}\alpha _m \) hold, for \(n_i \ge 0\ (\)where, \( \lnot ^{n_i}, n_i \ge 0 \), represents \( n_i \) iterations of negations of the formula \( \alpha _i )\). Then, for any formula \(\beta \) formed with \(\alpha _1, \alpha _2, \dots , \alpha _m \) over \(\{\wedge , \vee , \lnot \}\), \(\beta \) behaves classically.

Proof

First, we show that for any value of \(n_i \ge 0\), \(\circ \lnot ^{n_i}\alpha _i \vdash \alpha _i \vee \lnot \alpha _i\) and \(\circ \lnot ^{n_i}\alpha _i, \alpha _i \wedge \lnot \alpha _i \vdash \gamma \), for any \(\gamma \)—i.e. excluded middle and explosion hold for \( \alpha _i \).

Suppose \(\circ \lnot ^{n_i}\alpha _i\) holds. So, \( \lnot ^{n_i}\alpha _i \vee \lnot \lnot ^{n_i}\alpha _i \) and \(\lnot ^{n_i}\alpha _i \wedge \lnot \lnot ^{n_i}\alpha _i \vdash \gamma \) hold. If \(n_i\) is even, \(\lnot ^{n_i}\alpha _i \dashv \vdash \alpha _i\), and if \(n_i\) is odd, \(\lnot ^{n_i}\alpha _i \dashv \vdash \lnot \alpha _i\). So, it is easily proved that \( \lnot ^{n_i}\alpha _i \vee \lnot \lnot ^{n_i}\alpha _i \vdash \alpha _i \vee \lnot \alpha _i\). Since we have that \( \circ \lnot ^{n_i}\alpha _i \vdash \lnot ^{n_i}\alpha _i \vee \lnot \lnot ^{n_i}\alpha _i\), by transitivity, we get \( \circ \lnot ^{n_i}\alpha _i \vdash \alpha _i \vee \lnot \alpha _i\). In order to recover explosion, it can be easily proved that \(\alpha _i \wedge \lnot \alpha _i \vdash \lnot ^{n_i}\alpha _i \wedge \lnot \lnot ^{n_i}\alpha _i\). Since we have that \(\circ \lnot ^{n_i}\alpha _i, \lnot ^{n_i}\alpha _i \wedge \lnot \lnot ^{n_i}\alpha _i \vdash \gamma \), by transitivity, we get \(\circ \lnot ^{n_i}\alpha _i, \alpha _i \wedge \lnot \alpha _i \vdash \gamma \).

Remember that full classical logic can be obtained by adding explosion and excluded middle to the introduction and elimination rules of \( \wedge \) and \(\vee \), \(\alpha _1 \rightarrow \alpha _2 \) being defined as \(\lnot \alpha _1 \vee \alpha _2\). Now, in order to prove the result, it is enough to show that for any formula \(\beta \) formed with \(\alpha _1, \alpha _2, \dots , \alpha _m \) over \(\{\wedge , \vee , \lnot \}\), if \(\circ \lnot ^{n_1}\alpha _1, \circ \lnot ^{n_2}\alpha _2, \dots , \circ \lnot ^{n_m}\alpha _m \) hold, then \(\vdash \beta \vee \lnot \beta \) and \(\beta , \lnot \beta \vdash \gamma \) hold.

Let \(\Gamma = \{\circ \lnot ^{n_1}\alpha _1, \circ \lnot ^{n_2}\alpha _2, \dots , \circ \lnot ^{n_m}\alpha _m \} \).

If \(\beta = \alpha _i\), it has been proved above. The remaining cases are proved by induction on the complexity of \(\beta \).

Case 1. \(\beta = \lnot \delta \). I.H. \(\Gamma , \delta , \lnot \delta \vdash \gamma \) and \(\Gamma \vdash \delta \vee \lnot \delta \). It can be easily proved that \(\Gamma , \lnot \delta , \lnot \lnot \delta \vdash \gamma \) and \(\Gamma \vdash \lnot \delta \vee \lnot \lnot \delta \).

Case 2. \(\beta = \delta _1 \wedge \delta _2\). I.H. \(\Gamma , \delta _1, \lnot \delta _1 \vdash \gamma \) and \(\Gamma \vdash \delta _1 \vee \lnot \delta _1 \), mutatis mutandis for \(\delta _2\). It can be proved that \(\Gamma , \delta _1 \wedge \delta _2, \lnot (\delta _1 \wedge \delta _2)\vdash \gamma \) and \( \Gamma \vdash (\delta _1 \wedge \delta _2) \vee \lnot (\delta _1 \wedge \delta _2)\) The remaining cases are left to the reader. \(\square \)

We have seen in Fact 9 that for an implication \(\alpha \rightarrow \beta \) defined in \({ FDE}\) as \(\lnot \alpha \vee \beta \), modus ponens and the deduction theorem do not hold. Both are recovered for the defined implication in \({ LET}_{F}\) for classical propositions.

Fact 32

1.In \({ LET}_{F}\),for classical propositions, modus ponens holds:\(\circ \alpha , \alpha , \lnot \alpha \vee \beta \vdash \beta \).

Proof

\(\square \)

2. In \({ LET}_{F}\), the following form of the deduction theorem holds: \({\circ }\alpha , \alpha \vdash \beta \) implies \({\circ }\alpha \vdash \lnot \alpha \vee \beta \).

Proof

\(\square \)

Definition 33

Supplementing and complementing negations (Carnielli et al. 2007, pp. 12ff)

  1. 1.

    We say that a unary connective \(*\) in a logic \({\mathcal {L}}\) is a supplementing negation if: (i) for some formula \( \alpha \), \(*\alpha \) is not a bottom particle, and (ii) for any \( \Gamma , \alpha \) and \(\beta \): \(\Gamma , \alpha , *\alpha \vdash _{\mathcal {L}} \beta \).

  2. 2.

    We say that a unary connective \(*\) in a logic \({\mathcal {L}}\) is a complementing negation if: (i) for some formula \( \alpha \), \(*\alpha \) is not a top particle;

    (ii) for any \( \Gamma , \alpha \) and \(\beta \): \( \Gamma , \alpha \vdash _{{\mathcal {L}}} \beta \) and \( \Gamma , *\alpha \vdash _{\mathcal {L}} \beta \) implies \(\Gamma \vdash _{\mathcal {L}} \beta \).

If \(*\) is a complementing negation, for any \( \alpha \), at least one between \( \alpha \) and \(*\alpha \) hold, and if \(*\) is a complementing negation, it cannot be that both \( \alpha \) and \(*\alpha \) hold. Each one expresses one half of classical negation, the former excluded middle, the latter explosion. If a logic \({\mathcal {L}}\) has a (primitive or defined) negation connective that is both supplementing and complementing, then \({\mathcal {L}}\) has a classical negation. A complementing negation and a supplementing negation can be defined in \({ LET}_{F}\).

Definition 34

The following unary connectives can be defined in \({ LET}_{F}\):

  1. 1.

    The connective truth: \({\oplus } \alpha {\mathop {=}\limits ^{{\tiny def}}} {\circ }\alpha \wedge \alpha \);

  2. 2.

    The connective falsity: \({\sim } \alpha {\mathop {=}\limits ^{{\tiny def}}} {\circ }\alpha \wedge \lnot \alpha \);

  3. 3.

    The connective falsity-excluding: \( {\ominus } \alpha {\mathop {=}\limits ^{{\tiny def}}} {\bullet }\alpha \vee \alpha \);

  4. 4.

    The connective truth-excluding: \({\approx } \alpha {\mathop {=}\limits ^{{\tiny def}}} {\bullet }\alpha \vee \lnot \alpha \).

The tables are the following:

figure g

These connectives have been named for the following reasons. According to the proposed interpretation: (1) \( \circ \alpha \wedge \alpha \) means that there is conclusive evidence for \( \alpha \), and so \( \alpha \) is true (\( \oplus \alpha \)); (2) \( \circ \alpha \wedge \lnot \alpha \) means that there is conclusive evidence for the falsity of \( \alpha \), and so \( \alpha \) is false (\( {\sim } \alpha \)); (3) \( {\bullet }\alpha \vee \alpha \) means that there is no conclusive evidence for \( \alpha \), or \( \alpha \) holds, and so it excludes the falsity of \(\alpha \) (\( \ominus \alpha \)); (4) \( {\bullet }\alpha \vee \lnot \alpha \) means that there is no conclusive evidence for \( \alpha \), or \( \lnot \alpha \) holds, and so it excludes the truth of \(\alpha \) (\( {\approx } \alpha \)). It is also clear from the table above and Definition 33 that \( {\sim } \alpha \) is a supplementing negation [if \(v(\alpha )=1 \), \(v({\sim }\alpha )=0 \), they cannot be both 1], while \({\approx } \alpha \) is a complementing negation [if \(v(\alpha )=0 \), \(v({\approx }\alpha )=1 \), they cannot be both 0].Footnote 14 We conjecture that no classical negation can be defined in \({ LET}_{F}\).Footnote 15

These four connectives enjoy some interesting logical relations w.r.t. each other that can be displayed by a square of oppositions:

figure h

\( {\sim } \alpha \) and \({\oplus } \alpha \) are contrary propositions (i.e., they can both be false, but they cannot both be simultaneously true); \({\approx } \alpha \) and \( {\ominus } \alpha \) are subcontrary propositions (i.e., they can both be true, but they cannot both be simultaneously false); \({\oplus } \alpha \) (resp. \({\sim } \alpha \)) is the dual of \({\ominus } \alpha \) (resp. \({\approx } \alpha \)); \({\oplus } \alpha \) (resp. \({\sim } \alpha \)) is the contradictory of \({\approx } \alpha \) (resp. \({\ominus }\)). Notice that in \({ LET}_{F}\), \({\circ }\) is the dual of \({\bullet }\), and \(\lnot \) is the dual of itself [on duality between non-deterministic connectives in Logics of Formal Inconsistency and Undeterminedness, see Carnielli et al. (2019)].

Fact 35

  1. 1.

    \(\circ \alpha \wedge \alpha \wedge \lnot \alpha \), \({\circ }\alpha \wedge {\bullet }\alpha \), \(\oplus \alpha \wedge {\sim } \alpha \), \(\oplus \alpha \wedge {\approx } \alpha \), and \( \ominus \alpha \wedge {\sim } \alpha \) are bottom particles in \({ LET}_{F}\).

  2. 2.

    \({\sim }\) is a supplementing negation in \({ LET}_{F}\).

  3. 3.

    \({\approx }\) is a complementing negation in \({ LET}_{F}\).

Proof

In a few steps from the rules Cons, \(EXP^{\circ }\) and \({\bullet }R2\). \(\square \)

Theorem 36

The following propositions are theorems of \({ LET}_{F}\):

  1. 1.

    \(\circ \alpha \vee {\bullet }\alpha \)

  2. 2.

    \(\alpha \vee \lnot \alpha \vee {\bullet }\alpha \)

  3. 3.

    \(({\bullet }\alpha \wedge \alpha ) \vee ({\bullet }\alpha \wedge \lnot \alpha ) \vee {\bullet }\alpha \vee ({\bullet }\alpha \wedge \alpha \wedge \lnot \alpha ) \vee (\circ \alpha \wedge \alpha ) \vee (\circ \alpha \wedge \lnot \alpha ) \)

  4. 4.

    \( \alpha \vee \lnot \alpha \vee \ominus \alpha \)

  5. 5.

    \( \alpha \vee \lnot \alpha \vee {\approx } \alpha \)

  6. 6.

    \( \oplus \alpha \vee {\sim }\alpha \vee {\bullet }\alpha \)

Proof

Items 1 and 2 follow from the rules Comp and R2. To prove 3, from \( \circ \alpha \vdash \alpha \vee \lnot \alpha \), we obtain \( \circ \alpha \vdash (\circ \alpha \wedge \alpha ) \vee (\circ \alpha \wedge \lnot \alpha ) \), and so \( \circ \alpha \vdash ({\bullet }\alpha \wedge \alpha ) \vee ({\bullet }\alpha \wedge \lnot \alpha ) \vee {\bullet }\alpha \vee ({\bullet }\alpha \wedge \alpha \wedge \lnot \alpha ) \vee (\circ \alpha \wedge \alpha ) \vee (\circ \alpha \wedge \lnot \alpha ) \). On the other hand, \( {\bullet }\alpha \vdash ({\bullet }\alpha \wedge \alpha ) \vee ({\bullet }\alpha \wedge \lnot \alpha ) \vee {\bullet }\alpha \vee ({\bullet }\alpha \wedge \alpha \wedge \lnot \alpha ) \vee (\circ \alpha \wedge \alpha ) \vee (\circ \alpha \wedge \lnot \alpha ) \) holds. Now, use 1 and \( \vee E \). The proofs of 4, 5, and 6 are left to the reader. Notice that item 3 corresponds to the six scenarios presented in Sect. 3.1. \(\square \)

4 Probabilistic semantics for \({ LET}_{F}\)

We now present a probabilistic semantics for \({ LET}_{F}\) and \({ FDE}\).

Definition 37

Given a logic \({\mathcal {L}}\), with a derivability relation \( \vdash \) and a language L, a probability distribution for \({\mathcal {L}}\) is a real-valued function \(P : L \rightarrow {\mathbb {R}} \) satisfying the following conditions:

  1. 1.

    Non-negativity: \(0\le P(\alpha )\le 1\) for all \(\alpha \in L\);

  2. 2.

    Tautologicity: If \(\vdash \alpha \), then \(P(\alpha )=1\);

  3. 3.

    Anti-Tautologicity: If \(\alpha = \bot \), then \(P(\alpha )=0\);

  4. 4.

    Comparison: If \(\alpha \vdash \beta \), then \(P(\alpha )\le P(\beta )\);

  5. 5.

    Finite additivity: \(P(\alpha \vee \beta )=P(\alpha )+P(\beta )-P(\alpha \wedge \beta )\).

The clauses above can be regarded as meta-axioms that define probability functions for an appropriate logic \({\mathcal {L}}\) just by taking \(\vdash \) as the derivability relation of \({\mathcal {L}}\), and so the notion of probability can be regarded as logic-dependent. These clauses define probability functions for both \({ FDE}\) and \({ LET}_{F}\) just by employing respectively \( \vdash _{FDE}\) and \( \vdash _{LET_F}\).Footnote 16 From now on, we will concentrate on \({ LET}_{F}\), but it should be clear that the meta-axioms 1, 4 and 5 above define probability distributions for \({ FDE}\) as well.

Definition 38

\({ LET}_{F}\)-probability distribution. Let \(\Sigma = \{\alpha _{1}, \ldots , \alpha _{n}, \ldots \}\) be a (finite or infinite) collection of propositions in the language \(L_2\) of \({ LET}_{F}\). A \({ LET}_{F}\)-probability distribution over \(\Sigma \) is an assignment of probability values P to the elements of \(\Sigma \) that can be extended to a full probability function \(P : L_2 \rightarrow {\mathbb {R}}\) according to Definition 37.

It is a common view that the classical truth-values true (1) and false (0) can be identified with the endpoints of probabilities in the unit interval [0, 1]. On the other hand, interpretations \(v : L \rightarrow \{0,1\}\) of a formal language L can be regarded as degenerate probability functions \(P : {L}\rightarrow [0,1]\). The class of logics that make possible such an identification can be seen as a special case of probability logic. The standard view, however, is rather the opposite: it claims that probability logic presupposes, and so it depends on, classical logic.Footnote 17 But the connection between logic and probability theory is far from being restricted to classical logic. The fact that probability distributions can be defined based on a non-classical consequence relation, in our view, makes clear that the relation between logic and probability goes beyond the realm of classical logic.

4.1 Conditional probability

The notion of conditional probability of \(\alpha \) given \(\beta \) is defined as usual, for \(P(\beta )~\ne ~0\):

$$\begin{aligned} P(\alpha /\beta )= \begin{array}{l} P(\alpha \wedge \beta )\\ \overline{\quad P(\beta ) \quad } \end{array} \end{aligned}$$

In terms of evidence, a statement \( P(\alpha /\beta ) \) is to be read as a measure of how much the evidence available for \( \beta \) affects the evidence for \( \alpha \).

Some useful theorems of conditional probability of \({ LET}_{F}\)-distributions are the following, with the caveat that \(P(\beta ) \ne 0\) in all cases where \(P(\alpha /\beta )\) is mentioned:

Theorem 39

The following properties hold when the probabilities in the denominators are different from 0.

  1. 1.

    \(P(\alpha _{1}\wedge \cdots \wedge \alpha _{n}) = P(\alpha _{1}/ \alpha _{2}\wedge \cdots \wedge \alpha _{n}) \cdots P(\alpha _{n-1}/\alpha _{n}) P(\alpha _{n})\ (\)Chain Rule).

  2. 2.

    \(P(\alpha /\beta \wedge \gamma ) = \frac{P(\alpha /\gamma ) \cdot P(\beta /\alpha \wedge \gamma )}{ P(\beta /\gamma )}\).

  3. 3.

    \(P(\alpha \wedge \beta /\gamma ) = P(\alpha /\gamma ) \cdot P(\beta /\alpha \wedge \gamma ) = P(\beta /\gamma ) \cdot P(\alpha /\beta \wedge \gamma )\).

  4. 4.

    \(P(\alpha \vee \beta /\gamma ) = P(\alpha /\gamma ) + P(\beta /\gamma ) - P(\alpha \wedge \beta /\gamma ) \).

  5. 5.

    \(P(\alpha \vee \beta /\gamma ) = P(\alpha /\gamma ) + P(\beta /\gamma )\) if \(\alpha \) and \(\beta \) are logically incompatible, i.e., \(\alpha \wedge \beta \) act as a \( \bot \ (\)see Sect. 4.2).

  6. 6.

    \(P(\alpha /\beta ) + P(\lnot \alpha /\beta ) - P(\bullet \alpha /\beta ) \le P(\alpha \vee \lnot \alpha /\beta )\).

  7. 7.

    If \(P(\circ \alpha ){=} 1\), or equivalently \(P(\bullet \alpha ){=} 0\), then \(P(\alpha \vee \lnot \alpha ){=} 1\) and \(P(\alpha \wedge \lnot \alpha ){=} 0 \).

  8. 8.

    \(P(\alpha /\beta ) + P(\lnot \alpha /\beta ) = 1\), if \( P(\circ \alpha ) = 1\).

  9. 9.

    \(P(\beta / {\circ }\beta ) + P(\lnot \beta / {\circ }\beta ) = 1\).

Proof

  • Items 1 to 4 are quite elementary properties coming from the general definition of conditional probability: \(P(\alpha / \beta ) = \frac{P(\alpha \wedge \beta )}{P(\beta )} \), which gives the alternative product rule \( P(\alpha \wedge \beta )= P(\alpha / \beta ) \cdot P(\beta )\). The chain rule (item 1) is derived by successive applications of product rule. Items 2 to 4 are easy consequences of the definition of conditional probability and clause 5 of Definition 37.

  • Item 5: since \(\alpha \wedge \beta \) is a bottom particle in this case, \(P(\alpha \wedge \beta )=0 \), and the result follows from 4.

  • Item 6 is a consequence of Fact 25 (\(\alpha \wedge \lnot \alpha \vdash \bullet \alpha \)), Comparison and elementary inequalities, plus the definition of conditional probability.

  • Item 7: Easy consequence of Definition 22, R1 (Fact 24) and Comparison (Definition 35).

  • Item 8: If \(P(\circ \alpha ) = 1\), then by 9 \(P(\alpha \vee \lnot \alpha )=1\), and by Lemma 40 (below) we have \(P( (\alpha \vee \lnot \alpha )\wedge \beta )= P(\beta )= P( (\alpha \wedge \beta ) \vee (\lnot \alpha \wedge \beta ) )=P(\alpha \wedge \beta ) + P(\lnot \alpha \wedge \beta ) - P(\alpha \wedge \lnot \alpha \wedge \beta )\). Since \(P(\alpha \wedge \lnot \alpha \wedge \beta ) =0\) [\(P({\circ }\alpha )=1 \) implies \(P(\alpha \wedge \lnot \alpha ) = 0 \)], we obtain \(P(\beta )= P(\alpha \wedge \beta ) + P( \lnot \alpha \wedge \beta )\). Dividing both sides by \(P(\beta )\) obtains the result, in view of the definition of conditional probability.

  • Item 9: In \({ LET}_{F}\), \(\vdash \circ \beta \leftrightarrow \circ \beta \wedge (\beta \vee \lnot \beta ) \leftrightarrow (\circ \beta \wedge \beta ) \vee (\circ \beta \wedge \lnot \beta )\) (proof left to the reader). Thus \(P(\circ \beta )= P((\circ \beta \wedge \beta ) \vee (\circ \beta \wedge \lnot \beta ))= P(\circ \beta \wedge \beta ) + P(\circ \beta \wedge \lnot \beta )\) by Finite Additivity, since \({\circ }\beta \wedge \beta \wedge \lnot \beta \) is a bottom particle (Fact 33). Dividing both sides by \(P(\circ \beta )\) yields the result. \(\square \)

4.2 Independence and incompatibility

Intuitively, two propositions are independent if the fact that one holds does not have any effect on whether or not the other holds, and vice-versa. Two propositions \(\alpha \) and \(\beta \), are said to be independent w.r.t. a distribution P if \(P(\alpha \wedge \beta )= P(\alpha ) \cdot P(\beta )\). Two propositions can be independent relative to one probability distribution and dependent relative to another. Alternatively, independence can be defined as follows: \(\alpha \) is independent of \(\beta \) if \(P(\alpha /\beta ) = P(\alpha )\) [or equivalently, \(P(\beta /\alpha ) = P(\beta )\)].Footnote 18 Classically, \(\alpha \) and \(\lnot \alpha \) are never independent (unless one of them has probability zero). In view of item 4 of Theorem 42 below, \(P(\alpha \wedge \lnot \alpha ) \le P({\bullet }\alpha ) \), hence when \(P(\alpha ) \cdot P(\lnot \alpha ) > P({\bullet }\alpha ) \), \(\alpha \) and \(\lnot \alpha \) are not independent. In this way, \(P({\bullet }\alpha ) \) can be regarded as a bound on the ‘degree of independence’ between \(\alpha \) and \(\lnot \alpha \).

Intuitively, two propositions \(\alpha \) and \(\beta \) are logically incompatible if \( \alpha \) cannot hold when \( \beta \) holds, and vice-versa. Two propositions \(\alpha \) and \(\beta \), are said to be logically incompatible if \(\alpha , \beta \vdash \gamma \), for any \(\gamma \), or equivalently, if \(\alpha \wedge \beta \) is a bottom particle. Logically incompatible propositions \(\alpha \) and \(\beta \) with non-zero probabilities are always dependent since \(0= P(\alpha \wedge \beta ) \ne P(\alpha ) \cdot P(\beta )\). Again, for non-zero probabilities, classically \(\alpha \) and \(\lnot \alpha \) are incompatible, and so dependent. In \({ LET}_{F}\), however, they are neither necessarily incompatible nor necessarily dependent, when \( P({\circ }\alpha ) < 1 \). We saw in Fact 35 item 1 that \( \alpha \wedge \lnot \alpha \wedge \circ \alpha \) as well as \({\circ }\alpha \wedge {\bullet }\alpha \) defines a bottom particle in \({ LET}_{F}\). From clause 3 of Definition 37, it follows that for any probability distribution P, \( P(\alpha \wedge \lnot \alpha \wedge \circ \alpha )=0\) and \(P({\circ }\alpha \wedge {\bullet }\alpha )=0\). So, in \({ LET}_{F}\) \(\alpha \) and \({\sim }\alpha \) are always logically incompatible and hence dependent, while \(\alpha \) and \(\lnot \alpha \) can be independent.

An interesting property concerning the behavior of probability measures in \({ LET}_{F}\), related to independence in ‘extreme cases’, occurs when \(P(\alpha ) = 1\). In such cases \(\alpha \) is independent from the probability measure of any other distinct proposition \(\beta \). This kind of property contributes to the dynamics of evidence, in the sense of the interpretation of preservation of conclusive and non-conclusive evidence in \({ LET}_{F}\), in such a way that the increasing of conclusive evidence tends to truth.

Lemma 40

Independence of propositions with maximal probability. If \( P(\alpha ) = 1\) then \( P(\alpha \wedge \beta ) = P(\alpha )\cdot P(\beta )\), for \(\beta \ne \alpha \)

Proof

If \( P(\alpha ) = 1\) then \( P(\alpha \vee \beta ) = 1\) from Comparison, since \(\alpha \vdash \alpha \vee \beta \). By Finite Additivity \(1= P(\alpha \vee \beta ) = P(\alpha ) + P(\beta ) - P(\alpha \wedge \beta ) \). As \(P(\alpha ) =1\), it follows that \(P(\alpha \wedge \beta ) = P(\beta )\). \(\square \)

The restriction \( \alpha \ne \beta \) in the above lemma intends to avoid the problematic cases of ‘self-independence’ of extreme events. As mentioned before, two events \(\alpha \) and \(\beta \) are considered to be independent if \(P(\alpha \wedge \beta )= P(\alpha ) \cdot P(\beta )\), for some probability distribution P. This leads to a puzzling situation concerning events \(\alpha \) such that \( P(\alpha ) =0\) or \( P(\alpha ) =1\). In such cases, \( P(\alpha )= P(\alpha \wedge \alpha )= P(\alpha ) \cdot P(\alpha )\) in both cases. In this way, extreme probabilities can be regarded as independent of themselves, an uncomfortable situation, as recognized in Fitelson and Hájek (2017).

Lemma 40 leads immediately to the independence of consistent and inconsistent propositions in extreme cases:

  1. 1.

    If \( P(\circ \alpha ) = 1\) then \( P(\circ \alpha \wedge \beta ) = P(\beta )\), for \(\beta \ne \circ \alpha \)

  2. 2.

    If \( P(\bullet \alpha ) = 1\) then \( P(\bullet \alpha \wedge \beta ) = P(\beta )\), for \(\beta \ne \bullet \alpha \)

  3. 3.

    If \( P(\beta ) =1\) then \( P(\circ \alpha \wedge \beta ) = P(\circ \alpha )\), for \(\beta \ne \circ \alpha \)

  4. 4.

    If \( P(\beta ) =1\) then \( P(\bullet \alpha \wedge \beta ) = P(\bullet \alpha )\), for \(\beta \ne \bullet \alpha \)

Evidence can be increasing or decreasing in an historical series, leading to a dynamic of evidence. This can be expressed in mathematical terms by elementary series. Let \(lim_{i \rightarrow \infty }P_i(\alpha ) = \lambda \) mean that the sequence of values \( P_1(\alpha ), P_2(\alpha ), \ldots , P_i(\alpha ) \ldots \) is strictly monotonous and converges to \(\lambda \in [0, 1]\).

Lemma 41

The dynamics of evidence

  1. 1.

    If \(lim_{i \rightarrow \infty } P_i({\circ }\alpha )=1\) or \(lim_{i \rightarrow \infty } P_i({\bullet }\alpha )=0\), then \(lim_{i \rightarrow \infty } P_i(\alpha \vee \lnot \alpha ) = 1\) and \(lim_{i \rightarrow \infty } P_i(\alpha \wedge \lnot \alpha ) = 0\).

  2. 2.

    If \(lim_{i \rightarrow \infty } P_i({\circ }\alpha ){=}1\) or \(lim_{i \rightarrow \infty } P_i({\bullet }\alpha ){=}0\), then \( lim_{i \rightarrow \infty } (P_i(\alpha ){+}P_i(\lnot \alpha )) {=} 1\).

Proof

Suppose \(lim_{i \rightarrow \infty } \ P_i({\circ }\alpha ) =1\); by \(PEM^{{\circ }}\) and Comparison, \(P_i({\circ }\alpha ) \le P_i(\alpha \vee \lnot \alpha ) \le 1\). By the Squeeze Theorem of elementary calculus for series (aka the Sandwich Theorem) \(lim_{i \rightarrow \infty } P_i(\alpha \vee \lnot \alpha )=1\). All other limits are proved in similar ways. \(\square \)

The meaning of Lemma 41 is precisely that the values of \(P_i({\circ }\alpha )\) can be interpreted as degrees of classicality, in the sense that greater values of \(P_i({\circ }\alpha )\) indicate that the situation is approaching classicality and, conversely, the values of \( P_i({\bullet }\alpha )\) can be interpreted as degrees of anticlassicality, in the sense that smaller values of \(P_i({\bullet }\alpha )\) indicate that the situation is approaching classicality.

Some useful (though almost all immediate) properties of \({ LET}_{F}\)-distributions are the following:

Theorem 42

  1. 1.

    If \(\alpha \dashv \vdash \beta \), then \( P(\alpha )=P(\beta ) \).

  2. 2.

    \(P(\alpha \vee \beta ) = P(\alpha ) + P(\beta )\), if \(\alpha \) and \(\beta \) are logically incompatible.

  3. 3.

    \(P(\alpha \vee \beta \vee \gamma )= P(\alpha ) + P(\beta ) + P(\gamma ) - P(\alpha \wedge \beta ) - P(\alpha \wedge \gamma ) - P(\beta \wedge \gamma ) + P(\alpha \wedge \beta \wedge \gamma )\).

  4. 4.

    \(P(\alpha \wedge \lnot \alpha )\le P({\bullet }\alpha )\).

  5. 5.

    \(P({\circ }\alpha ) \le P(\alpha \vee \lnot \alpha )\).

  6. 6.

    \(P({\circ }\alpha ) = 1 - P({\bullet }\alpha )\).

  7. 7.

    \(P(\oplus \alpha \wedge \lnot \alpha )= 0\), \(P({\sim }\alpha \wedge \alpha )= 0\).

  8. 8.

    \(P(\circ \alpha \vee {\bullet }\alpha )= 1\), \(P(\alpha \vee \lnot \alpha \vee {\bullet }\alpha )= 1\)

  9. 9.

    \(P({\circ }\alpha \vee (\alpha \wedge \lnot \alpha )) \le P(\alpha \vee \lnot \alpha )\)

  10. 10.

    \(1+ P( (\alpha \vee \lnot \alpha )\wedge {\bullet }\alpha ) = P(\alpha \vee \lnot \alpha ) + P({\bullet }\alpha )\)

  11. 11.

    \(1+ P( (\alpha \wedge \lnot \alpha )\vee {\circ }\alpha ) = P(\alpha \wedge \lnot \alpha ) + P({\circ }\alpha )\)

  12. 12.

    If \(P({\circ }\alpha )=1 \ [or \ equivalently \ P({\bullet }\alpha )=0]\), then \(P(\lnot \alpha ) = 1- P(\alpha )\)

  13. 13.

    If \(P({\circ }\alpha )=1 \ [or \ equivalently \ P({\bullet }\alpha )=0]\), then \(P(\alpha \vee \lnot \alpha ) = 1\) and \(P(\alpha \wedge \lnot \alpha ) = 0\).

Proof

Routine, from the axioms of probability and the derivability relation of \({ LET}_{F}\). We just sketch the proof of items 12 and 13. For 12, suppose \(P({\circ }\alpha ) =1\); by items 4 and 5 above, \(PEM^{{\circ }}\), and Comparison, \( 1=P({\circ }\alpha ) \le P(\alpha \vee \lnot \alpha )\), and \(P(\alpha \wedge \lnot \alpha )\le P({\bullet }\alpha )=0\), hence by Finite Additivity \(P(\alpha \vee \lnot \alpha ) + P(\alpha \wedge \lnot \alpha ) = 1 + 0 = P(\alpha ) + P(\lnot \alpha )\). Hence \(P(\lnot \alpha )= 1- P(\alpha )\). For 13 a similar reasoning as of 12 is obtained. \(\square \)

Items 1 and 2 are usual results in probabilistic logic, and 3 is a particular case of the Inclusion-Exclusion property for finite probability, easily adapted for propositions, that hold for arbitrary finite disjunctions (see Grinstead and Snell 1997). Items 4 and 5 establish constraints on the values of \(P(\circ \alpha )\), \(P(\alpha )\) and \(P(\lnot \alpha )\). Item 7 concerns bottom particles, and 8, theorems of \({ LET}_{F}\) essential for proving total probability theorems (Sect. 4.3 below). Items 12 and 13 show the classical behavior of probabilities when \(P({\circ }\alpha )=1\).

4.3 Total probability theorems for \({ LET}_{F}\)

In the classical approach to probability, total probability theorems compute the probability of an event \( \beta \) in a sample space partitioned into exclusive and exhaustive events. Typically, for a partition in two pieces, a total probability theorem that reflects excluded middle assumes the following form:

$$\begin{aligned} P(\beta ) = P(\beta \wedge \alpha ) + P(\beta \wedge \lnot \alpha ). \end{aligned}$$

Here, however, we are not really talking about sample spaces, about events themselves, but rather about the information related to such events, that we call an information space. In the standard approach to probability theory, we start from a group of events, say, the two outcomes of tossing a coin, and attribute probabilities to these events, whose sum is always equal to 1. Let \( \alpha \) express that the toss of a coin comes up heads. The sample space is thus divided into two parts, \( \alpha \) and \( \lnot \alpha \), corresponding respectively to heads and tails (not heads). If the coin behaves as expected, their probability are the same.

On the other hand, we consider here a language that is able to express information about some event—for example, the result of a referendum—that comes from different sources and may be unreliable. Such information is constituted by evidence for ‘yes’ and for ‘no’ that can be non-conclusive, incomplete, contradictory, more reliable or less reliable, and perhaps even conclusive. Let \( \alpha \) express the result ‘yes’, and \( \lnot \alpha \) the result ‘no’. In this case, the propositions we are concerned with are \( \alpha , \lnot \alpha \), \( \circ \alpha \), \( {\bullet }\alpha \), as well as other propositions of the language of \({ LET}_{F}\) formed from them, for example, \( {\bullet }\alpha \vee \alpha \), \(\alpha \wedge \lnot \alpha \), \(\circ \alpha \wedge \alpha \), etc. A \({ LET}_{F}\)-probability distribution attributes values to these propositions. The information space is thus constituted by such propositions and the measures of probabilities attributed to them by a \({ LET}_{F}\)-probability distribution P. Note that, contrary to the classical case, \( P(\alpha )+P(\lnot \alpha ) \) can be greater or less than 1 precisely because \(\alpha \) and \(\lnot \alpha \) do not establish a partition of the information space.

Now, the question is: since we cannot rely on the classical, mutually exclusive partitions of the sample space, how can total probability theorems be stated? In order to provide such theorems for \({ LET}_{F}\), we have to rely on the connectives \({\circ }\), \({\bullet }\), and on the connectives defined in Fact 34. We also need a bit of terminology.

Definition 43

(Cleavage). Let us call a cleavage a (finite) family of propositions \(\{ \alpha _1 , \alpha _2,\ldots , \alpha _n \}\). A cleavage is said to be exhaustive if \( \alpha _1 \vee \alpha _2 \vee \cdots \vee \alpha _n \) is a tautology, and so it covers all the information space, possibly with intersections. A cleavage is said to be exclusive when \( \alpha _1 \vee \alpha _2 \vee \cdots \vee \alpha _n \) are pairwise logically incompatible. In this case, it does not yield intersection of information (in the sense that \(\alpha _{i} \wedge \alpha _{j}\) for \(i\ne j\) is a bottom particle), and possibly does not cover the whole space. An exhaustive and exclusive cleavage is a partition.

Items 2 to 5 of Theorem 36 cleave the information space exhaustively but not exclusively. Items 1 and 6, on the other hand, cleave the information space in parts that are exhaustive and exclusive, and so they are partitions. Notice that item 4 of Theorem 36 corresponds to the six scenarios of conclusive and non-conclusive evidence that we have seen in Sect. 3.1. These scenarios can be graphically represented as follows:

\({\bullet }\alpha \wedge \alpha \)

\({\bullet }\alpha \wedge \lnot \alpha \)

\({\bullet }\alpha \)

\({\bullet }\alpha \wedge (\alpha \wedge \lnot \alpha )\)

\(\circ \alpha \wedge \alpha \)

\(\circ \alpha \wedge \lnot \alpha \)

1

2

3

4

5

6

Item 1 of Theorem 36 above emphasizes the division between non-conclusive evidence (scenarios 1 to 4) and conclusive evidence (scenarios 5 and 6), while item 6, in addition, splits the conclusive evidence into truth (5) and falsity (6). These propositions can be understood as expressing different ways we can look at the information space. The following total probability theorems can be obtained depending upon certain cleavages, based on Theorem 36.

Theorem 44

Total probability theorems

  1. 1.

    \(P(\beta )= P(\beta \wedge {\circ }\alpha ) + P(\beta \wedge {\bullet }\alpha )\), w.r.t. the cleavage \(\{ {\circ }\alpha ,{\bullet }\alpha \}\).

  2. 2.

    \(P(\beta )= P(\beta \wedge \alpha ) + P(\beta \wedge \lnot \alpha ) + P(\beta \wedge {\bullet }\alpha ) - P(\beta \wedge \alpha \wedge {\bullet }\alpha ) - P(\beta \wedge \lnot \alpha \wedge {\bullet }\alpha ) \), w.r.t. the cleavage \( \{ \alpha ,\lnot \alpha , {\bullet }\alpha \} \).

  3. 3.

    \(P(\beta ) = P(\beta \wedge \circ \alpha \wedge \alpha ) + P(\beta \wedge \circ \alpha \wedge \lnot \alpha ) + P(\beta \wedge {\bullet }\alpha ) - P(\beta \wedge {\bullet }\alpha \wedge \alpha \wedge \lnot \alpha ) \), w.r.t. cleavage \(\{ {\bullet }\alpha \wedge \alpha , {\bullet }\alpha \wedge \lnot \alpha , {\bullet }\alpha , {\bullet }\alpha \wedge \alpha \wedge \lnot \alpha , \circ \alpha \wedge \alpha , \circ \alpha \wedge \lnot \alpha \}\).

  4. 4.

    \(P(\beta )= P(\beta \wedge \alpha )+ P(\beta \wedge \lnot \alpha ) + P(\beta \wedge \ominus \alpha ) - P(\beta \wedge \alpha \wedge \ominus \alpha ) - P(\beta \wedge \lnot \alpha \wedge \ominus \alpha ) \), w.r.t. the cleavage \(\{ \alpha ,\lnot \alpha , \ominus \alpha \}\).

  5. 5.

    \(P(\beta )= P(\beta \wedge \alpha )+ P(\beta \wedge \lnot \alpha ) + P(\beta \wedge {\approx }\alpha )- P(\beta \wedge \alpha \wedge {\approx }\alpha ) -P(\beta \wedge \lnot \alpha \wedge {\approx }\alpha ) \), w.r.t. the cleavage \( \{ \alpha ,\lnot \alpha , {\approx }\alpha \} \).

  6. 6.

    \(P(\beta ) = P(\beta \wedge \oplus \alpha ) + P(\beta \wedge {\sim }\alpha ) + P(\beta \wedge {\bullet }\alpha ) \), w.r.t. the cleavage \(\{ \oplus \alpha ,{\sim }\alpha , {\bullet }\alpha \}\).

Proof

1. \( \beta \dashv \vdash (\beta \wedge {\circ }\alpha ) \vee (\beta \wedge {\bullet }\alpha ) \). So, \(P(\beta ) = P((\beta \wedge {\circ }\alpha ) \vee (\beta \wedge {\bullet }\alpha )) = P(\beta \wedge \circ \alpha ) + P(\beta \wedge {\bullet }\alpha ) - P(\beta \wedge \circ \alpha \wedge {\bullet }\alpha ) = P(\beta \wedge \circ \alpha ) + P(\beta \wedge {\bullet }\alpha ) \). The remaining proofs are left to the reader. In view of Definition 3.1 (connectives \(\oplus \), \({\sim } \), \(\ominus \), \({\approx } \)), some of these cleavages are equivalent. \(\square \)

4.4 Bayes’ rule

As is well-known, Bayes’ rule, or Bayes’ theorem, computes the probability of an event based on previous information related to that event. The standard Bayes’ rule proves that, for \(P(\beta )\ne 0\):

$$\begin{aligned} P(\alpha /\beta )= \begin{array}{l} P(\beta /\alpha )\cdot P(\alpha ) \\ \overline{\quad \quad P(\beta ) \quad \quad } \end{array} \end{aligned}$$

In the equation above, interpreted in terms of measures of evidence rather than standard probabilities, \(P(\alpha )\) denotes the evidence available for \( \alpha \) without taking into consideration any evidence for \( \beta \). The latter is supposed to affect someway the evidence for \(\alpha \), and so \(P(\alpha / \beta )\) is the measure of the evidence for \( \alpha \) after \( \beta \) is taken into account. \(P(\beta /\alpha )\), usually called the ‘likelihood’ in probability theory, is the evidence for \(\beta \) when \(\alpha \) is considered as given, and \(P(\beta ) \), usually called the ‘marginal likelihood’, is the total evidence available for \( \beta \), that takes into account all the possible cases where \(\beta \) may occur. In what follows, we define some relevant versions of Bayes’ rule. Differently from the classical case, these versions are not equivalent. They show how the notion of classicality can modify Bayesian probability updating.

Theorem 45

Bayes’ Conditionalization Rules

  1. 1.
    $$\begin{aligned} P(\alpha /\beta )= & {} \frac{P(\beta /\alpha )\cdot P(\alpha )}{P(\beta /{\circ }\alpha )\cdot P({\circ }\alpha ) + P(\beta /{\bullet }\alpha )\cdot P({\bullet }\alpha )}\\&for\, P(\beta )\ne 0,\,P({\circ }\alpha )\ne 0,\, and\, P({\bullet }\alpha )\ne 0. \end{aligned}$$

Proof

From the definition of conditional probability and Theorem 44, item 1. \(\square \)

  1. 2.
    $$\begin{aligned} P(\alpha /\beta )= \frac{P(\beta /\alpha )\cdot P(\alpha )}{\begin{array}{l}P(\beta /\alpha )\cdot P(\alpha ) + P(\beta /\lnot \alpha )\cdot P(\lnot \alpha ) + P(\beta /{\bullet }\alpha )\cdot P({\bullet }\alpha ) - \\ P(\beta /\alpha \wedge {\bullet }\alpha )\cdot P(\alpha \wedge {\bullet }\alpha ) - P(\beta /\lnot \alpha \wedge {\bullet }\alpha )\cdot P(\lnot \alpha \wedge {\bullet }\alpha )\end{array}}\\ for\, P(\beta )\ne 0,\, P(\alpha \wedge {\bullet }\alpha )\ne 0,\, and\, P(\lnot \alpha \wedge {\bullet }\alpha )\ne 0. \end{aligned}$$

Proof

From the definition of conditional probability and Theorem 44, item 2. \(\square \)

  1. 3.
    $$\begin{aligned} P(\alpha /\beta )= & {} \frac{P(\beta /\alpha )\cdot P(\alpha )}{P(\beta /{\oplus }\alpha )\cdot P(\oplus \alpha ) + P(\beta /{\sim }\alpha )\cdot P({\sim }\alpha ) + P(\beta /{\bullet }\alpha )\cdot P({\bullet }\alpha )}\\&for\, P(\beta )\ne 0,\, P({\bullet }\alpha )\ne 0,\, P(\oplus \alpha )\ne 0, and P(\ominus \alpha )\ne 0. \end{aligned}$$

Proof

From the definition of conditional probability and Theorem 44, item 6. \(\square \)

It should be clear that the process of limit can be easily established for the above formulations of Bayes’ rules. If \(lim \ P_i({\circ }\alpha )=1\) [or equivalently \(lim \ P_i({\bullet }\alpha )=0\)] then item 1 above reduces to \(P({\circ }\alpha /\beta )= 1\). Analogously, if \(lim \ P_i({\bullet }\alpha )=0\) [or equivalently \(lim \ P_i({\circ }\alpha )=1\)], then items 2 and 3 above reduce to the standard form of Bayes’ rule.

5 Final remarks

This paper has been conceived to be a further development of the approach to paraconsistency as preservation of evidence presented in Carnielli and Rodrigues (2017, 2019), where an interpretation of contradictions in terms of non-conclusive evidence was proposed. The underlying assumption is that there are no true contradictions, but rather argumentative contexts in which conflicting evidence, as well as the absence of any evidence, may occur. The valuation semantics is able to express only that there is or there is not evidence for a proposition \( \alpha \), while the probabilistic semantics presented here intends to express the degree of evidence enjoyed by a given proposition. The acceptance of scenarios in which \(P(\alpha ) + P(\lnot \alpha ) > 1\), however, does not mean that there may be something like ‘contradictory sample spaces’, or ‘contradictory probabilistic spaces’. The latter would be the probabilistic counterpart of contradictions in reality, a view on paraconsistency not endorsed by us. In our view, it is the information available about some collection of events that can be contradictory. So, instead of talking about sample spaces, the concept of an information space has been introduced here.

Both \({ LET}_{J}\) and \({ LET}_{F}\) are Logics of Formal Inconsistency and Undeterminedness suitable for an intuitive interpretation in terms of preservation of evidence and truth. The intuition regarding \({\circ }\) and \({\bullet }\) as ‘classically contradictory’ w.r.t. each other had already been presented in Carnielli et al. (2019, Section 4.4). \({ LET}_{F}\), however, as far as we know, is the first formal system where these connectives are both primitive and have the deductive behavior given by rules Cons and Comp, that are in some sense analogous to explosion and excluded middle. The connective \({\bullet }\), and the fact that \( {\circ }\alpha \vee {\bullet }\alpha \) and \( \alpha \vee \lnot \alpha \vee {\bullet }\alpha \) are theorems of \({ LET}_{F}\), are essential for proving total probability theorems and Bayes’ rules (Theorems 44 and 45).

The probabilistic semantics of \({ LET}_{F}\) has been axiomatically stated in Definitions 37 and 38. Accordingly, \( P(\alpha ) + P(\lnot \alpha ) \) can be greater or less than 1, and this is interpreted as scenarios, respectively, of conflicting evidence, and little or no evidence. When \( P(\circ \alpha ) = 1 \), the classical behavior of \( P(\alpha )\) and \( P(\lnot \alpha )\) is restored, and this is interpreted as saying that the evidence available for \( \alpha \) and \( \lnot \alpha \) is subjected to the laws of standard probability theory. But \(P(\circ \alpha )\) may be less than 1, and in this case, according to the axioms, it expresses the degree to which \( P (\alpha ) \) and \(P(\lnot \alpha )\) are expected to behave classically [the value of \(P(\circ \alpha )\) establishes constraints on the values of \( P(\alpha \vee \lnot \alpha )\) and \( P(\alpha \wedge \lnot \alpha ) \), cf. Lemma 41]. Accordingly, \(P(\circ \alpha ) < 1\) can be intuitively interpreted as expressing the reliability of the available evidence for \( \alpha \) and \( \lnot \alpha \): greater reliability corresponds to a greater degree of classicality.

Our treatment here does not intend to express degrees of belief by means of probability measures. The notion of evidence for \( \alpha \), as explained in Sect. 2.2.1, does not imply belief in \( \alpha \). So, the degree of evidence for \( \alpha \) measured by a statement \(P(\alpha )=\epsilon \) is not a measure of the belief of an agent in \( \alpha \). However, nothing a fortiori prevents the formal system proposed here, together with its probabilistic semantics, of being interpreted, or used, as a tool to measure degrees of belief, uncertainty, or some other relation between agents and propositions. Similar remarks apply to the connective \({\circ }\). In \( P({\circ }\alpha ) = \epsilon \), the value of \( \epsilon \) expresses the degree to which it is expected that \( P(\alpha ) \) behaves classically. Indeed, \(\epsilon \) can also be interpreted as the degree of reliability of evidence for \( \alpha \), coherence with previous data or with a historical series of measures of evidence for \( \alpha \), or even with a subjective ingredient, for example, as the degree of trustfulness of the belief in \( \alpha \), or certainty/uncertainty of \( \alpha \).

The rules for \({\circ }\) and \({\bullet }\), due to their dual character, show a symmetry that deserves to be further investigated from the proof-theoretic point of view. There are some extensions of \({ LET}_{J}\) and \({ LET}_{F}\) that also deserve to be studied. The operator \({\bullet }\) and the rules Cons and Comp can be added to \({ LET}_{J}\), obtaining a logic that differs from \({ LET}_{F}\) only in the implication for the non-classical propositions. Two intuitively appealing equivalences are the following:

  • 1. \( \ {{\circ }\alpha } \dashv \vdash {{\circ }{\circ }\alpha }\)

  • 2. \( {{\circ }\alpha } \dashv \vdash {{\circ }\lnot \alpha } \)

It was shown in Carnielli and Rodrigues (2017, Fact 17) that \({ LET}_{J}\) has no theorems of the form \({\circ }\alpha \) (the same result also holds for \({ LET}_{F}\)), and it was argued that \({ LET}_{J}\) (and so \({ LET}_{F}\)) was conceived in such a way that \({\circ }\) has to be introduced from outside the formal system. This is in line with the idea that information about conclusive evidence for a proposition \( \alpha \) comes from outside the formal system. But it is also very reasonable to suppose that once the truth value of a proposition \( \alpha \) has been established, and so \(\circ \alpha \) holds and \(\alpha \) has classical behavior, then \(\circ \alpha \), \({\circ }{\circ }\alpha \), and so on, also have classical behavior. Conversely, it is also reasonable to conclude \(\circ \alpha \) from \({\circ }{\circ }\alpha \), and so on. These ideas are expressed by 1 above. The equivalence 2 above makes explicit inside the system the first part of the result achieved by Fact 31 (to wit: once \( \circ \lnot ^n\alpha \) is proved, and so it follows that \( \lnot ^n\alpha \) is subjected to classical logic, for any formula \( \lnot ^m\alpha \), \( m \ge 0 \), \( \lnot ^m\alpha \) is also subjected to classical logic). Valuation semantics for these rules are straightforward, and adding these rules would produce a decidable formal system.

We believe that the probabilistic semantic relation presented in Sect. 4 will succeed as a tool for dealing with real argumentative contexts, including investigative scenarios and databases concerned with different degrees of evidence attributed to propositions. But this claim needs to be further investigated.