Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Certainty factor and lift are known measures of association rules. The former measure was offered in the expert system Mycin [9], while the latter is widely implemented in both commercial and non-commercial data mining systems [2]. Nevertheless, they do not guarantee accurate evaluation of the strength of dependence between rule’s constituents. In particular, even if there is a strongest possible positive or negative dependence between rule’s constituents X and Y, these measures may reach values quite close to the values indicating independence of X and Y. This might suggest that one deals with a weak dependence, while in fact the dependence is strong. In [4], we proposed a new measure called a dependence factor to overcome this drawback. Unlike in the case of the certainty factor, when defining the dependence factor, we took into account the fact that for a given rule \(X \rightarrow Y\), the minimal conditional probability of the occurrence of Y given X may be greater than 0, while its maximal possible value may less than 1. The dependence factor always takes value 1 if a dependence is strongest possible positive, whereas for a strongest possible negative dependence, it always takes value –1; in the case of independence, it takes value 0.

In [4], we have focused on examining properties of the dependence factor as a measure of dependence between rule’s constituents/events. Our new main contribution in this paper is: (1) the examination of the dependence factor as an interestingness measure with respect to the interestingness postulates formulated by Piatetsky-Shapiro in [7], and (2) the derivation of the relationship among the dependence factor for X and Y, with those for \(\bar{X}\) and Y, X and \(\bar{Y}\), as well as \(\bar{X}\) and \(\bar{Y}\), respectively.

Our paper has the following layout. In Sect. 2, we briefly recall basic notions of association rules, their basic measures (support, confidence) as well as lift and certainty factor. In Sect. 3, we recall maximal and minimal values of examined measures in the case when probabilities of rule’s constituents are fixed, as shown in [4]. In Sect. 4, we recall the definition and properties of the dependence factor after [4]. Our new contribution is presented in Sects. 5 and 6. In Sect. 5, we examine the usefulness of the dependence factor as an interestingness measure, while in Sect. 6, we identify the relationship between the dependence factors for events and their complements. Section 7 concludes our work.

2 Basic Notions and Properties

In this section, we recall the notion of association rules after [1].

Definition 1

Let \(I=\{i_{1}, i_{2}, \ldots , i_{m} \}\) be a set of distinct literals, called items (e.g. products, features, symptoms). Any subset X of the set I is called an itemset. A transaction database is denoted by \(\fancyscript{D}\) and is defined as a set of itemsets. Each itemset T in \(\fancyscript{D}\) is a transaction. An association rule is an expression associating two itemsets:

Itemsets and association rules are typically characterized by support and confidence, which are simple statistical parameters.

Definition 2

Support of an itemset X is denoted by sup(X) and is defined as the number of transactions in \(\fancyscript{D}\) that contain X; that is

$$ {\textit{sup}}(X) = {\vert }\{ T \in {\fancyscript{D}} {\vert } X \subseteq T \}{\vert }. $$

Support of a rule \(X \rightarrow Y\) is denoted by sup(\(X \rightarrow Y)\) and is defined as the support of \(X \cup Y\); that is,

$$ {\textit{sup}}(X \rightarrow Y) = {\textit{sup}}(X \cup Y). $$

Clearly, the probability of the event that itemset X occurs in a transaction equals \({ sup}(X) / {\vert } {\fancyscript{D}} {\vert }\), while the probability of the event that both X and Y occur in a transaction equals \({ sup}(X \cup Y) / {\vert } \mathrm{D} {\vert }\). In the remainder, the former probability will be denoted by P(X), while the latter by \(P({ XY})\).

Definition 3

The confidence of an association rule \(X \rightarrow Y\) is denoted by conf(\(X \rightarrow Y)\) and is defined as the conditional probability that Y occurs in a transaction provided X occurs in the transaction; that is:

$$ {\textit{conf}}(X \rightarrow Y)=\text { }\frac{sup(X\rightarrow Y{)}}{sup{(}X{)}}={ }\frac{P(XY{)}}{P{(}X{)}}. $$

A large amount of research was devoted to strong association rules understood as those association rules the supports and confidences of which exceed user-defined support threshold and confidence threshold, respectively. However, it has been argued in the literature that these two measures are not sufficient to express different interestingness, usefulness or unexpectedness aspects of association rules [3, 58, 1012]. In fact, a number of such measures of association rules was proposed (see e.g. [3, 58, 1012]. Among them very popular measures are lift [2] and certainty factor [9].

Definition 4

The lift of an association rule \(X \rightarrow Y\) is denoted by lift(\(X \rightarrow Y)\) and is defined as the ratio of the conditional probability of the occurrence of Y in a transaction given X occurs there to the probability of the occurrence of Y; that is:

$$ {\textit{lift}}(X\rightarrow Y)={ }\frac{conf(X\rightarrow Y{)}}{P{(}Y{)}}. $$

Lift may be also defined in an equivalent way in terms of probabilities only:

Property 1

$$ {\textit{lift}}(X\rightarrow Y)=\frac{P{(}XY{)}}{P{(}X{)}\times P{(}Y{)}}. $$

Definition 5

The certainty factor of an association rule \(X \rightarrow Y\) is denoted by cf(\(X \rightarrow Y)\) and is defined as the degree to which the probability of the occurrence of Y in a transaction can change when X occurs there as follows:

$$ \displaystyle cf(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{conf(X\rightarrow Y)-P(Y)}{1-P(Y)} &{} \text {if}\,\, conf(X\rightarrow Y)>P(Y)\text {,} \\ 0 &{} \text {if}\,\, conf(X\rightarrow Y)=P(Y)\text {,} \\ -\frac{P(Y)-conf(X\rightarrow Y)}{P(Y)-0} &{} \text {if}\,\, conf(X\rightarrow Y)<P(Y). \end{array} }} \right. $$

The definition of the certainty factor is based on the assumption that the probability of the occurrence of Y in a transaction given X occurs there (conf(\(X\rightarrow Y))\) can be increased from P(Y) up to 1 and decreased from P(Y) down to 0. In Fig. 1, we visualize the meaning of the absolute value of the certainty factor as the ratio of the lengths of respective intervals.

Fig. 1
figure 1

Calculating the absolute value of the certainty factor as the ratio of the lengths of respective intervals when conf(\(X\rightarrow Y) > P(Y)\) (on the left-hand side) and when conf(\(X\rightarrow Y) < P(Y)\) (on the right-hand side)

As shown in Property 2, the certainty factor can be expressed equivalently in terms of unconditional probabilities (by multiplying the numerator and denominator of the formula in Definition 5 by P(X)) or lift (by dividing the numerator and denominator of the original cf formula by P(Y)).

Property 2

 

  1. (a)

    \(cf(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{P(XY)-P(X)\times P(Y)}{P(X)-P(X)\times P(Y)} &{} \mathrm{if}\,\,P(XY)>P(X)\times P(Y)\text {,} \\ 0 &{} \mathrm{if}\,\,P(XY)=P(X)\times P(Y)\text {,} \\ -\frac{P(X)\times P(Y)-P(XY)}{P(X)\times P(Y)-0}&{} \mathrm{if}\,\,P(XY)<P(X)\times P(Y). \\ \end{array} }} \right. \)

  2. (b)

    \(cf(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{lift(X\rightarrow Y)-1}{\frac{1}{P(Y)}-1} &{} \qquad \,\,\,\,\,\, \mathrm{if}\,\,lift(X\rightarrow Y)>1\text {,} \\ 0 &{} \qquad \,\,\,\,\,\, \mathrm{if}\,\,lift(X\rightarrow Y)=1\text {,} \\ -\frac{1-lift(X\rightarrow Y)}{1-0} &{} \qquad \,\,\,\,\,\, \mathrm{if}\,\,lift(X\rightarrow Y)<1. \\ \end{array} }} \right. \)

Both lift and certainty factor are related to the notion of (in)dependence of events, where two events are treated as independent if the product of the probabilities of their occurrences equals the probability that the two events co-occur. Otherwise, they are regarded as dependent. Note that this notion of dependence does not indicate which event is a reason of the other. However, it allows formulating whether the dependence between events is positive or negative in the case when the events are dependent on each other.

Definition 6

X and Y are:

  • independent if \(P({ XY}) = P(X) \times P(Y)\),

  • dependent positively if \(P({ XY}) > P(X) \times P(Y)\),

  • dependent negatively if \(P({ XY}) < P(X) \times P(Y)\).

In Table 1, we provide equivalent conditions in terms of P, conf, lift and cf for independence, positive dependence and negative dependence, respectively, between two itemsets.

Table 1 Conditions for independence, positive dependence and negative dependence

In general, one may distinguish between symmetric (two direction) measures of association rules and asymmetric (one direction) ones.

Definition 7

A measure m is called symmetric (two direction) if \(m(X\rightarrow Y)=m(Y\rightarrow X)\) for any X and Y. Otherwise, it is called an asymmetric (one direction) measure.

Property 3

 

  1. (a)

    \({\textit{conf}}(X \rightarrow Y) = {\textit{conf}}(Y\rightarrow X)\) is not guaranteed to hold.

  2. (b)

    \({\textit{lift}}(X\rightarrow Y) = {\textit{lift}}(Y \rightarrow X)\).

  3. (c)

    \({\textit{cf}}(X \rightarrow Y) = {\textit{cf}}(Y \rightarrow X)\) is not guaranteed to hold if conf(\(X\rightarrow Y) > P(Y)\).

  4. (d)

    \({\textit{cf}}(X\rightarrow Y) = {\textit{cf}}(Y\rightarrow X)\) if conf(\(X\rightarrow Y) \le P(Y)\).

As follows from Property 3, conf is an asymmetric measure and lift is a symmetric measure. On the other hand, we observe that strangely cf has a mixed nature—asymmetric for positive dependences and symmetric for negative dependences and independences. This observation provoked us to revisit the definition of cf and to propose its modification in [4]. When defining the dependence factor there, we took into account the fact that in some circumstances it may be infeasible to increase the probability of the occurrence of Y in a transaction under the presence of X (conf(\(X\rightarrow Y))\) from P(Y) up to 1 as well as it may be infeasible to decrease it from P(Y) down to 0.

3 Maximal and Minimal Values of Rule Measures

In this section, we first recall global maximal and minimal values of rule measures (Table 2). Next, following [4], we recall maximal and minimal values of rule measures for given values of P(X) and P(Y).

Table 2 Global maximal and minimal values of rule measures

In the remainder of the paper, we denote maximal probability and minimal probability of the co-occurrence of X and Y given P(X) and P(Y) are fixed by \({ max\_P}({ XY}{\vert }_{P(X), P(Y)})\) and \({ min\_P}({ XY}{\vert }_{P(X), P(Y)})\), respectively. Analogously, maximal confidence and minimal confidence (maximal lift, minimal lift, maximal certainty factor, minimal certainty factor) of \(X\rightarrow Y\) given P(X) and P(Y) are fixed are denoted by \({ max\_conf}(X\rightarrow Y{\vert }_{P(X), P(Y)})\) and \({ min\_conf}(X\rightarrow Y{\vert }_{P(X), P(Y)}) \quad ({ max\_lift}(X\rightarrow Y{\vert }_{P(X), P(Y)})\), \({ min\_lift}(X\rightarrow Y{\vert }_{P(X), P(Y)})\), \({ max\_cf}(X\rightarrow Y{\vert }_{P(X), P(Y)})\), \({ min\_cf}(X\rightarrow Y{\vert }_{P(X), P(Y)})\), respectively.

Property 4

 

  1. (a)

    \({ max\_conf(X\rightarrow Y{\vert }_{P(X), P(Y)})}=\frac{{ max\_P(XY|_{P(X),P(Y)} )}}{{ P(X)}}\)

  2. (b)

    \({ min\_conf(X\rightarrow Y{\vert }_{P(X), P(Y)})}=\frac{{ min\_P(XY|_{P(X),P(Y)} )}}{ {P(X)}}\)

  3. (c)

    \({ max\_lift(X\rightarrow Y{\vert }_{P(X), P(Y)})}=\frac{{ max\_conf\mathrm{(}XY|_{P(X),P(Y)} )}}{{ P(Y)}}=\frac{{ max\_P(XY|_{P(X),P(Y)} )}}{{ P(X)\times P(Y)}}\)

  4. (d)

    \({ min\_lift(X\rightarrow Y{\vert }_{P(X), P(Y)})}=\frac{{ min\_conf\mathrm{(}XY|_{P(X),P(Y)} )}}{ {P(Y)}}=\frac{{ min\_P(XY|_{P(X),P(Y)} )}}{{ P(X)\times P(Y)}}\)

  5. (e)

    \({ max\_cf(X\rightarrow Y{\vert }_{P(X), P(Y)})}=\frac{{ max\_conf(X\rightarrow Y|_{P(X),P(Y)} )-P(Y)}}{{ 1-P(Y)}}\)

    \(= \frac{max\_P(XY|_{P(X),P(Y)} )-P(X)\times P(Y)}{P(X)-P(X)\times P(Y)}=\frac{max\_lift(XY|_{P(X),P(Y)} )-1}{\frac{1}{P(Y)}-1}\)

  6. (f)

    min_cf(\(X\rightarrow Y{\vert }_{P(X), P(Y)})=-\frac{P(Y)-min\_conf(X\rightarrow Y|_{P(X),P(Y)} )}{P(Y)-0} \)

    \(= -\frac{P(X)\times P(Y)-min\_P(XY|_{P(X),P(Y)} )}{P(X)\times P(Y)-0}=-\frac{1-min\_lift(XY|_{P(X),P(Y)} )}{1-0}\)

In Proposition 1, we show how to calculate \({ min\_P}({ XY}{\vert }_{P(X), P(Y)})\) and \({ max\_P}({ XY}{\vert }_{P(X), P(Y)})\). We note that neither \({ max\_P}({ XY}{\vert }_{P(X), P(Y)})\) necessarily equals 1 nor \({ min\_P}({ XY}{\vert }_{P(X), P(Y)})\) necessarily equals 0. Figure 2 illustrates this.

Proposition 1

 

  1. (a)

    \({ max\_P}({ XY}{\vert }_{P(X), P(Y)}) = \min \{ P(X), P(Y)\}\)

  2. (b)

    \({ min\_P}({ XY}{\vert }_{P(X), P(Y)})=\left\{ {{\begin{array}{ll} 0 &{}\mathrm{if}\,\, P{(}X{)}+P{(}Y{)}\le 1 \\ P{(}X{)}+P{(}Y{)}- 1 &{} \mathrm{if}\,\, P{(}X{)}+P{(}Y{)}> 1 \\ \end{array} }} \right. \)

    \(= \max \{0, P{(}X{)}+P{(}Y{)}-1 \}\)

The next proposition follows from Property 4 and Proposition 1.

Fig. 2
figure 2

a \({ max\_P}({ XY}{\vert }_{P(X), P(Y)}) = \min \{ P(X), P(Y)\} = \min \left\{ \frac{3}{6}, \frac{2}{6} \right\} = \frac{2}{6}\). b \({ min\_P}({ XY}{\vert }_{P(X), P(Y)}) = 0 \,\, \text {if}\,\, P(X) + P(Y) \le 1\). c \({ min\_P}({ XY}{\vert }_{P(X), P(Y)})=P(X) + P(Y) - 1 =\frac{5}{6}+\frac{4}{6}-1 = \frac{3}{6}\,\) if \(P(X) + P(Y) > 1\)

Proposition 2

 

  1. (a)

    \({ max\_conf}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\frac{\min \{P(X),P(Y)\}}{P(X)}=\left\{ {{\begin{array}{ll} 1 &{} \mathrm{if}\,\, P(X)\le P(Y)\text {,} \\ \frac{P(Y)}{P(X)} &{} \mathrm{if}\,\, P(Y)<P(X)\text {.} \\ \end{array} }} \right. \)

  2. (b)

    \({ min\_conf}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\frac{\max \{0,P(X)+P(Y)-1\}}{P(X)}\)

    \(=\left\{ {{\begin{array}{ll} 0 &{} \mathrm{if}\,\, P(X)+P(Y)\le \text {1,} \\ \frac{P(X)+P(Y)-\text {1}}{P(X)} &{} \mathrm{if}\,\, P(X)+P(Y)>\text {1.} \\ \end{array} }} \right. \)

  3. (c)

    \({ max\_lift}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\frac{\min \{P(X),P(Y)\}}{P(X)\times P(Y)}=\frac{1}{max\{P(X),P(Y)\}}\).

  4. (d)

    \({ min\_lift}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\frac{\max \{0,P(X)+P(Y)-1\}}{P(X)\times P(Y)} \)

    \( =\left\{ {{\begin{array}{ll} 0 &{} \mathrm{if}\,\, P(X)+P(Y)\le \text {1,} \\ \frac{P(X)+P(Y)-\text {1}}{P(X)\times P(Y)} &{} \mathrm{if}\,\, P(X)+P(Y)>\text {1.} \\ \end{array} }} \right. \)

  5. (e)

    \({ max\_cf}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\frac{\min \{P(X),P(Y)\}-P(X)\times P(Y)}{P(X)-P(X)\times P(Y)}\)

    \(=\frac{\frac{1}{\max \{P(X),P(Y)\}}-1}{\frac{1}{P(Y)}-1}\text { }=\left\{ {{\begin{array}{ll} 1 &{} \mathrm{if}\,\, P(X)\le P(Y), \\ \frac{\frac{1}{P(X)}-1}{\frac{1}{P(Y)}-1} &{} \mathrm{if}\,\, P(X)>P(Y). \\ \end{array} }} \right. \)

  6. (f)

    \({ min\_cf}(X\rightarrow Y{\vert }_{P(X), P(Y)})=\text { }-\frac{P(X)\times P(Y)-\max \{0,P(X)+P(Y)-1\}}{P(X)\times P(Y)-0} \)

    \(=\text { }\frac{\max \{0,P(X)+P(Y)-1\}}{P(X)\times P(Y)}-1=\left\{ {{\begin{array}{ll} -1 &{} \mathrm{if}\,\, P(X)+P(Y)\le 1 \\ \frac{P(X)+P(Y)-1}{P(X)\times P(Y)}-1&{} \mathrm{if}\,\, P(X)+P(Y)>1. \\ \end{array} }} \right. \)

In Table 3, we summarize real achievable maximal and minimal values of P(XY), conf(\(X\rightarrow Y),{ lift}(X\rightarrow Y)\) and cf(\(X \rightarrow Y)\) for given values of P(X) and P(Y).

Table 3 Real achievable maximal and minimal values of P(XY), conf \((X \rightarrow Y)\), lift \((X \rightarrow Y)\) and cf \((X \rightarrow Y)\) for given values of P(X) and P(Y)

4 Dependence Factor

In this section, we recall the definition of the dependence factor of a rule \(X\rightarrow Y\), which we offered in [4] as a modification of the certainty factor. Unlike the certainty factor, it is based on real maximal and minimal values of conf(\(X \rightarrow Y)\) for given values of P(X) and P(Y). Then we present the properties of this measure.

Definition 8

The dependence factor of \(X \rightarrow Y\) is denoted by df(\(X \rightarrow Y)\) and is defined as the ratio of the actual change of the probability of the occurrence of Y in a transaction given X occurs there to its maximal feasible change as follows:

$$ df(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{conf(X\rightarrow Y)-P(Y)}{max\_conf(X\rightarrow Y|\text { }_{P{(}X\text {), }P{(}Y{)}} )-P(Y)} &{} \mathrm{if}\,\, conf(X\rightarrow Y)>P(Y)\text {,} \\ 0 &{} \mathrm{if}\,\, conf(X\rightarrow Y)=P(Y)\text {,} \\ -\frac{P(Y)-conf(X\rightarrow Y)}{P(Y)-min\_conf(X\rightarrow Y|_{P{(}X\text {), }P{(}Y{)}} )} &{} \mathrm{if}\,\, conf(X\rightarrow Y)<P(Y). \\ \end{array} }} \right. $$

The dependence factor not only determines by how much the probability of the occurrence of Y in a transaction changes under the presence of X with respect to by how much it could have changed, but also it determines by how much the probability of the occurrence of X and Y in a transaction differs from the probability of their common occurrence under independence assumption with respect to by how much it could have been different (see Proposition 3a). In addition, the dependence factor determines by how much the value of the lift of a rule \(X \rightarrow Y\) differs from the value 1 (that is, from the value indicating independence of rule’s constituents in terms of the lift measure) with respect to by how much it could have been be different (see Proposition 3b).

Proposition 3

 

  1. (a)

    \(df(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{P(XY)-P(X)\times P(Y)}{max\_P(XY|_{P(X),P(Y)} )-P(X)\times P(Y)} &{} \mathrm{if}\,\, P(XY)>P(X)\times P(Y)\text {,} \\ 0 &{} \mathrm{if}\,\, P(XY)=P(X)\times P(Y)\text {,} \\ -\frac{P(X)\times P(Y)-P(XY)}{P(X)\times P(Y)-min\_P(XY|_{P(X),P(Y)} )} &{} \mathrm{if}\,\, P(XY)<P(X)\times P(Y). \\ \end{array} }} \right. \)

  2. (b)

    \(df(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{lift(X\rightarrow Y)-1}{max\_lift(X\rightarrow Y|_{P(X),P(Y)} )-1} &{} \qquad \,\,\,\, \mathrm{if}\,\, lift(X\rightarrow Y)>1\text {,} \\ 0 &{} \qquad \,\,\,\, \mathrm{if}\,\, lift(X\rightarrow Y)=1\text {,} \\ -\frac{1-lift(X\rightarrow Y)}{1-min\_lift(X\rightarrow Y|_{P(X),P(Y)} )} &{} \qquad \,\,\,\, \mathrm{if}\,\, lift(X\rightarrow Y)<1. \\ \end{array} }} \right. \)

Theorem 1

 

  1. (a)

    If \(P({ XY}) > P(X) \times P(Y)\), then df(\(X\rightarrow Y) \in \) (0, 1].

  2. (b)

    If \(P({ XY}) = P(X) \times P(Y)\), then df \((X \rightarrow Y) = 0\).

  3. (c)

    If \(P({ XY}) < P(X) \times P(Y)\), then df(\(X\rightarrow Y) \in \) [\(-\)1, 0).

Proof

Follows from Proposition 3a. \(\square \)

As follows from Proposition 3a, the dependence factor is a symmetric measure.

Theorem 2

df \((X \rightarrow Y) = { df}(Y\rightarrow X)\).

Based on Proposition 1 and 3a, we will express the dependence factor df \((X \rightarrow Y)\) in terms of P(XY), P(X) and P(Y), which will be useful for examining properties of this measure.

Theorem 3

$$ df(X\rightarrow Y)=\left\{ {{\begin{array}{ll} \frac{P(XY)-P(X)\times P(Y)}{\min \{P(X),P(Y)\}-P(X)\times P(Y)} &{} \mathrm{if}\,\, P(XY)>P(X)\times P(Y)\text {,} \\ 0 &{} \mathrm{if}\,\, P(XY)=P(X)\times P(Y)\text {,} \\ -\frac{P(X)\times P(Y)-P(XY)}{P(X)\times P(Y)-\max \{0,P(X)+P(Y)-1\}} &{} \mathrm{if}\,\, P(XY)<P(X)\times P(Y). \\ \end{array} }} \right. $$

One may easily note that df \((X \rightarrow Y)\) reaches 1 when P(XY) is maximal for given values of P(X) and P(Y); that is, when P(XY) = min \(\{ P(X), P(Y)\}\) or, in other words, when the dependence between X and Y is strongest possible positive for given values of P(X) and P(Y). Analogously, df \((X \rightarrow Y)\) reaches –1 when P(XY) is minimal for given values of P(X) and P(Y); that is, when P(XY) = max \(\{0, P(X)+P(Y) -1\}\) or, in other words, when the dependence between X and Y is strongest possible negative for these probability values (Table 4).

Table 4 Maximal and minimal values of df \((X \rightarrow Y)\) for any given values of P(X) and P(Y)

Based on Theorem 3 and Property 2a, one may derive relations between the dependence factor and the certainty factor as follows:

Theorem 4

 

  1. (a)

    \({ df}(X\rightarrow Y) \ge { cf} (X \rightarrow Y)\)                                                   if  \(P({ XY}) > P(X) \times P(Y)\),

  2. (b)

    \({ df}(X \rightarrow Y) = { cf}(X \rightarrow Y) = 0\)                                if   \(P ({ XY}) = P(X) \times P(Y)\),

  3. (c)

    \({ df}(X\rightarrow Y) \le { cf} (X \rightarrow Y)\)                                       if   \(P({ XY}) < P(X) \times P(Y)\),

  4. (d)

    \({ df}(X \rightarrow Y) = \max \{{ cf} (X \rightarrow Y), { cf}(Y\rightarrow X)\}\)        if   P(XY) \(> P(X) \times P(Y)\),

  5. (e)

    \({ df}(X \rightarrow Y) = { cf} (X \rightarrow Y)\)                                     if   \(P({ XY}) < P(X) \times P(Y)\) and \(P(X)+P(Y) < 1\),

  6. (f)

    \({ df}(X\rightarrow Y) < { cf} (X \rightarrow Y)\)                                     if   \(P ({ XY}) < P(X) \times P(Y)\) and \(P(X)+P(Y) > 1\).

Tables 56 illustrate the findings expressed as Theorem 4. In particular, Table 5 shows values of lift \((X \rightarrow Y)\), cf \((X \rightarrow Y)\) and df \((X \rightarrow Y)\) for P(X) = 0.6 and P(Y) = 0.3; that is, in the case when \(P(X)+P(Y) \le \) 1. For these values of P(X) and P(Y), the maximal possible value for P(XY) equals min \(\{P(X)\), \(P(Y)\}\) = 0.3. The fact of reaching the maximal possible value by P(XY) for the given values of P(X) and P(Y) is reflected by the value of df \((X \rightarrow Y)\) = 1, which means that the dependence between X and Y is strongest possible positive. On the other hand, cf \((X \rightarrow Y)\) = 0.29 does not reflect this fact. In general, the real dependence of Y on X may be underestimated when expressed in terms of cf \((X \rightarrow Y)\). Also the value 1.67 of lift \((X \rightarrow Y)\) itself does not reflect the strong positive dependence between X and Y in the considered case in the view that the lift may reach very large values (close to infinity) in general.

Table 5 Comparison of values of lift \((X \rightarrow Y)\), cf \((X \rightarrow Y)\) and df \((X \rightarrow Y)\) when \(P(X)+P(Y) \le 1\)
Table 6 Comparison of values of lift \((X \rightarrow Y)\), cf \((X \rightarrow Y)\) and df \((X \rightarrow Y)\) when \(P(X)+P(Y) >\) 1

Table 6 shows values of lift \((X \rightarrow Y)\), cf \((X \rightarrow Y)\) and df \((X \rightarrow Y)\) for P(X) = 0.8 and P(Y) = 0.6; that is, in the case when \(P(X)+P(Y) >\) 1. For these values of P(X) and P(Y), the minimal possible value of P(XY) equals \(P(X)+P(Y) - 1\, \)= 0.4. Then the dependence between X and Y is strongest possible negative. This is reflected by the value of df \((X \rightarrow Y)\) = \(-\)1. On the other hand, cf \((X \rightarrow Y)\) = –0.17 does not reflect this fact by itself. Also the value 0.83 of lift \((X \rightarrow Y)\) itself does not reflect the strong negative dependence between X and Y as it is positioned closer to the value 1 characteristic for independence rather than to the value 0.

5 Dependence Factor as an Interestingness Measure

In [7], Piatetsky-Shapiro postulated that a good interestingness measure of an association rules \(X\rightarrow Y\) should fulfill the following conditions:

  1. 1.

    be equal to 0 if X and Y are independent; that is, if P(XY) = \(P(X) \times P(Y)\),

  2. 2.

    be increasing with respect to P(XY) given P(X) and P(Y) are fixed,

  3. 3.

    be decreasing with respect to P(X) given P(XY) and P(Y) are fixed or be decreasing with respect to P(Y) given P(XY) and P(X) are fixed.

According to [7], the following rule interest measure \({ ri}(X \rightarrow Y) = {\vert } {\fancyscript{D}} {\vert } \times [P ({ XY}) - P(X) \times P(Y)]\) fulfills the above postulates. Nevertheless, we notice that this measure does not always satisfy the third postulate. Beneath we present the case in which the ri measure violates this postulate:

Let P(Y) = 0. Then, P(XY) = 0. In this case, ri \((X \rightarrow Y)\) = 0 for each value of P(X) in the interval [0, 1]. Thus, ri \((X \rightarrow Y)\) is not guaranteed to be decreasing with respect to P(X) given P(XY) and P(Y) are fixed. Analogically, we would derive that ri \((X \rightarrow Y)\) = 0 for each value of P(Y) in the interval [0, 1] if P(X) = 0. So, ri \((X \rightarrow Y)\) is not guaranteed to be decreasing with respect to P(X) given P(XY) and P(Y) are fixed. As a result, ri \((X \rightarrow Y)\) does not fulfill the third postulate if P(X) or P(Y) equals 0.

In fact, the novelty \((X \rightarrow Y)\) measure, which was defined in [5] as [P(XY)–P(X\(\times \) P(Y)], violates the third postulate in the same way as ri \((X \rightarrow Y)\).

Now, we will focus on examining if the dependence factor fulfills the postulates of rule interestingness. We start with formulating the properties of probabilities of events which will be useful in our examination.

Proposition 4

 

  1. (a)

    If \(P(X) = 0\) or \(P(Y) = 0\) or \(P(X) = 1\) or \(P(Y) = 1\), then \(P({ XY}) = P(X) \times P(Y)\).

  2. (b)

    If \(P({ XY}) \ne P(X) \times P(Y)\), then \(P(X), P(Y) \in (0, 1)\).

Proof

  Ad (a) Trivial.

Ad (b) Follows from Proposition 4a. \(\square \)

Theorem 5

Let \(X\rightarrow Y\) be an association rule.

  1. (a)

    df \((X \rightarrow Y)\) = 0 iff P(XY) = \(P(X) \times P(Y)\).

  2. (b)

    df is increasing with respect to P(XY) given P(X) and P(Y) are fixed.

  3. (c)

    df is non-increasing with respect to P(X) given P(XY) and P(Y) are fixed. In addition, df is decreasing with respect to P(X) given P(XY) and P(Y) are fixed, \(P(Y) \notin \{ 0, P ({ XY}), 1\}\) and P(XY) \(\ne \) 0.

  4. (d)

    df is non-increasing with respect to P(Y) given P(XY) and P(X) are fixed. In addition, df is decreasing with respect to P(Y) given P(XY) and P(X) are fixed,\( P(X) \notin \{0, P({ XY}), 1\} \) and \(P ({ XY}) \ne 0\).

Proof

Ad (a, b) Follow trivially from Theorems 1 and 3.

Ad (c) Let us first determine the derivative \({{\textit{df}}^{\,\,\prime }}\) \((X \rightarrow Y)\) of df \((X \rightarrow Y)\) as a function of variable P(X) based on Theorem 3 in all possible cases when P(XY) \(\ne P(X)\) \(\times \) P(Y). We will use the fact that in such cases P(X), \(P(Y) \in \) (0, 1) (by Proposition 4b).

Case P(XY) \(> P(X) \times P(Y)\) and min\(\{ P(X)\), \(P(Y)\}\) = P(X).

Then P(XY) \(> P(X) \times P(Y) >\) 0 and

\({{\textit{df}}^{\,\,\prime }}\)(\(X\rightarrow Y)= \frac{P{(XY)} \times (1-P(Y))}{(P(X)-P(X) \times P(Y))^{2}} < 0\).

Case P(XY) \(> P(X) \times P(Y)\) and min\(\{ P(X)\), \(P(Y)\} = P(Y)\).

Then

\({{\textit{df}}^{\,\,\prime }}\) \((X \rightarrow Y) = \frac{P{(Y)} \times (P(XY) - P(Y))}{(P(Y)-P(X) \times P(Y))^{2}}.\)

Hence:

  • If P(XY) = P(Y), then \({ df}^{\prime }(X \rightarrow Y)\) = 0.

  • If P(XY) \(\ne P(Y)\), then P(XY) \(< P(Y)\), so \({{\textit{df}}^{\,\,\prime }}\)(\(X \rightarrow Y) <\) 0.

Case P(XY) \(< P(X) \times P(Y)\) and \( \max \{0,P(X)+P(Y)-1 \}\) = 0.

Then

\({{ df}^{\prime }}(X \rightarrow Y) = \frac{P{XY} \times (P(Y)}{(P(X)\times P(Y))^{2}}.\)

Hence:

  • If P(XY) = 0, then \({{\textit{df}}^{\,\,\prime }}\)(\(X \rightarrow Y)\) = 0.

  • If P(XY) \(\ne \) 0, then \({{\textit{df}}^{\,\,\prime }}\)(\(X \rightarrow Y) <\) 0.

Case P(XY) \(< P(X) \times P(Y)\) and \(\max \{0,P(X)+P(Y)-1\}=P(X)+P(Y)-1\).

Then

\({{\textit{df}}^{\,\,\prime }}\) \((X \rightarrow Y) = \frac{(1-P(Y))\times (P(XY)-P(Y))}{(P(X)\times P(Y)-(P(X)+P(Y)-1))^2} = \frac{(1-P(Y))\times (P(XY)-P(Y))}{\left( (1-P(X)) \times (1-P(Y)) \right) ^2} .\)

Hence:

  • If P(XY) = P(Y), then \({{\textit{df}}^{\,\,\prime }}\)(\(X \rightarrow Y)\) = 0.

  • If P(XY) \(\ne P(Y)\), then P(XY) \(< P(Y)\), so \({{\textit{df}}^{\,\,\prime }}\)(\(X \rightarrow Y) <\) 0.

Now, let us consider the case when P(XY) = \(P(X) \times P(Y)\) and P(Y) \(\in \) (0, 1). Then P(X) may take only one value, namely \(\frac{P(XY)}{P(Y)}\).

Finally, we note that for P(Y) = 0 as well as for P(Y) = 1, P(XY) = \(P(X) \times P(Y)\) (by Proposition 4a), and so, df \((X \rightarrow Y)\) = 0 for each value of P(X) in the interval [0, 1].

Thus, df is a non-increasing function with respect to P(X) given P(XY) and P(Y) are fixed. However, if \(P(Y) \notin \{0, P ({ XY}), 1\}\) and \(P ({ XY}) \ne 0\), then df is a decreasing function with respect to P(X) given P(XY) and P(Y) are fixed.

Ad (d) Analogous to the proof of Theorem 5c. \(\square \)

Corollary 1

df \((X \rightarrow Y)\) fulfills the first and second Piatetsky-Shapiro postulates. In addition, it fulfills the third Piatetsky-Shapiro postulate if \(P(Y) \notin \{0, P({ XY}), 1\}\) and P(XY) \(\ne \) 0 or if \(P(X) \notin \{0, P({ XY}), 1\}\) and \(P({ XY}) \ne 0\).

Proof

By Theorem 5. \(\square \)

6 Dependence Factors for Events and Their Complements

In this section, we examine the relationship between the dependence factors for events and their complements. We start with determining extreme values of joint probabilities of events and their complements. Next, we prove that the character of the (in)dependence between X and Y determines uniquely the character of the (in)dependence between \(\bar{X}\) and YX and \(\bar{Y}\), as well as \(\bar{X}\) and \(\bar{Y}\), respectively. Eventually, we derive the relationship among the dependence factor for X and Y, with those for \(\bar{X}\) and YX and \(\bar{Y}\), as well as \(\bar{X}\) and \(\bar{Y}\), respectively.

Proposition 5

 

  1. (a)

    \({ max\_P}({ XY}{\vert }_{P(X),P(Y)}) = 1\) iff \(P(X) = P(Y) = 1\).

  2. (b)

    \({ min\_P}({ XY}{\vert }_{P(X),P(Y)}) = 0\) iff \(P(X) + P(Y) \le 1\).

  3. (c)

    \(P(X)+P(Y) \le 1\) iff \((1 - P(X)) + (1 - P(Y)) \ge 1\) iff \(P(\bar{X}) + P(\bar{Y}) \ge 1\).

Proof

Ad (a) Follows from Proposition 1a.

Ad (b) Follows from Proposition 1b.

Ad (c) Trivial. \(\square \)

Proposition 6

 

  1. (a)

    \({ max\_P}(\bar{X}\bar{Y}\vert _{P(\bar{X}), P(\bar{Y})}) = \min \{P (\bar{X}), P(\bar{Y})\} = \min \{1 - P(X), 1 - P(Y) \} = 1 - \max \{P(X), P(Y)\}\)

  2. (b)

    \({ min\_P}(\bar{X}\bar{Y}\vert _{P(\bar{X}), P(\bar{Y})}) = \max \{0, P(\bar{X}) + P (\bar{Y}) - 1\} = \)

    \(\max \{0, (1 - P(X)) + (1 - P(Y)) - 1\} = \max \{0, 1 - P(X)-P(Y)\}\)

  3. (c)

    \({ max\_P}(X\bar{Y}\vert _{P({X}), P(\bar{Y})}) = \min \{ P ({X}), P (\bar{Y})\} = min\{ P(X), 1 - P(Y) \}\)

  4. (d)

    \({ min\_P}(X\bar{Y}\vert _{P({X}), P(\bar{Y})}) = \max \{0, P(X) + P(\bar{Y}) - 1\} =\)

    \( \max \{0, P(X) + (1 - P(Y)) - 1\} = \max \{0, P(X) - P(Y)\}\)

  5. (e)

    \({ max\_P}(\bar{X}{Y}\vert _{P(\bar{X}), P({Y})}) = \min \{P(\bar{X}), P({Y})\}= \min \{1 - P(X), P(Y)\}\)

  6. (f)

    \({ min\_P}(\bar{X}{Y}\vert _{P(\bar{X}), P({Y})}) = \max \{0, P(\bar{X}) + P(Y) - 1\} =\)

    \(\max \{0, (1 - P(X)) + P(Y) - 1\} = \max \{0, P(Y) - P(X)\}\)

Proof

Ad (a, c, e) Follows from Proposition 1a, saying that \({ max\_P}({ VZ}{\vert }_{P(V),P(Z)}) = \min \{P(V), P(Z)\}\).

Ad (b, d, f) Follows Proposition 1b, saying that \({ min\_P}({ VZ}{\vert }_{P(V),P(Z)}) = \max \{0, P{(}V{)}+P{(}Z{)}-\text {1}\}\). \(\square \)

Lemma 1

 

  1. (a)

    \(P({ XY}) > P(X) \times P(Y)\) iff \(P(\bar{X}\bar{Y}) > P(\bar{X}) \times P(\bar{Y})\) iff \(P({X}\bar{Y}) < P({X}) \times P(\bar{Y})\) iff \(P(\bar{X}{Y}) < P(\bar{X}) \times P(Y)\).

  2. (b)

    \(P({ XY}) = P(X) \times P(Y)\) iff \(P(\bar{X}\bar{Y}) = P(\bar{X}) \times P(\bar{Y})\) iff \(P({X}\bar{Y}) = P(X) \times P(\bar{Y})\) iff \(P(\bar{X}{Y}) = P(\bar{X}) \times P(Y)\).

  3. (c)

    \(P({ XY}) < P(X) \times P(Y)\) iff \(P(\bar{X}\bar{Y}) < P(\bar{X}) \times P(\bar{Y})\) iff \(P(X\bar{Y}) > P(X) \times P(\bar{Y})\) iff \(P(\bar{X}Y) > P(\bar{X}) \times P(Y)\).

Proof

We will proof the proposition using the following equations:

  • \(P(\bar{X}) = 1 - P(X)\),    \(P(\bar{Y}) = 1 - P(Y)\),

  • \(P(\bar{X}Y)=P(Y)-P({ XY})\), \(P(X\bar{Y}) = P(X)-P({ XY})\),

  • \(P(\bar{X}\bar{Y}) = P(\bar{X}) - P(\bar{X}Y) = 1 - P(X)-P(Y)+P({ XY})\).

Ad (a)

  • \(P(\bar{X}\bar{Y}) > P(\bar{X}) \times P(\bar{Y})\) iff \(1 - P(X)-P(Y) + P({ XY}) > (1 - P(X)) \times (1 - P(Y))\) iff \(P({ XY}) > P(X) \times P(Y)\).

  • \(P(X\bar{Y}) < P(X) \times P(\bar{Y})\) iff \(P(X)- P({ XY}) < P(X) \times (1 - P(Y))\) iff \(P({ XY}) > P(X) \times P(Y)\).

  • \(P(\bar{X}Y) < P(\bar{X}) \times P(Y)\) iff \(P(Y)-P({ XY}) < (1 - P(X)) \times P(Y)\) iff \(P({ XY}) > P(X) \times P(Y)\).

Ad (b, c) Analogous to the proof of Lemma 1a. \(\square \)

Proposition 7

 

  1. (a)

    X and Y are dependent positively iff \(\bar{X}\) and \(\bar{Y}\) are dependent positively iff X and \(\bar{Y}\) are dependent negatively iff \(\bar{X}\) and Y are dependent negatively.

  2. (b)

    X and Y are independent iff \(\bar{X}\) and \(\bar{Y}\) are independent iff X and \(\bar{Y}\) are independent iff \(\bar{X}\) and Y are independent.

  3. (c)

    X and Y are dependent negatively iff \(\bar{X}\) and \(\bar{Y}\) are dependent negatively iff X and \(\bar{Y}\) are dependent positively iff \(\bar{X}\) and Y are dependent positively.

Proof

Follows from Lemma 1. \(\square \)

Lemma 2

(Proof in Appendix)

  1. (a)

    \({ df}(X \rightarrow Y) = { df}(\bar{X} \rightarrow \bar{Y})\)

  2. (b)

    \({ df}(X \rightarrow \bar{Y}) = { df}(\bar{X} \rightarrow Y)\)

  3. (c)

    \({ df}(X \rightarrow \bar{Y}) = -{ df}(X \rightarrow Y\mathbf{)}\)

Theorem 6 follows immediately from Lemma 2.

Theorem 6

$$ {\textit{df}}(X \rightarrow Y) = { df}(\bar{X} \rightarrow \bar{Y}) = -{ df}(X \rightarrow \bar{Y}) = -{ df}(\bar{X} \rightarrow Y). $$

Corollary 2

 

  1. (a)

    \({ df}(X \rightarrow Y)\) reaches maximum iff \({ df}(\bar{X} \rightarrow \bar{Y})\) reaches maximum iff \({ df}(X \rightarrow \bar{Y})\) reaches minimum iff \({ df}(\bar{X} \rightarrow Y)\) reaches minimum.

  2. (b)

    \({ df}(X \rightarrow Y)\) reaches minimum iff \({ df}(\bar{X} \rightarrow \bar{Y})\) reaches minimum iff \({ df}(X \rightarrow \bar{Y})\) reaches maximum iff \({ df}(\bar{X} \rightarrow Y)\) reaches maximum.

7 Conclusions

In [4], we have offered the dependence factor as a new measure for evaluating the strength of dependence between rules’ constituents. Unlike in the case of the certainty factor, when defining the dependence factor, we took into account the fact that for a given rule \(X \rightarrow Y\), the minimal conditional probability of the occurrence of Y given X may be greater than 0, while its maximal possible value may less than 1. df \((X \rightarrow Y)\) always reaches 1 when the dependence between X and Y is strongest possible positive, –1 when the dependence between X and Y is strongest possible negative, and 0 if X and Y are independent. Unlike the dependence factor, the certainty factor itself as well as lift are misleading in expressing the strength of the dependence. In particular, if there is strongest possible positive dependence between X and Y, cf \((X \rightarrow Y)\) is not guaranteed to reach its global maximum value 1 (in fact, its value can be quite close to 0 that suggests independence). On the other hand, if there is strongest possible negative dependence between X and Y, cf \((X \rightarrow Y)\) is not guaranteed to reach its global minimum value –1 (in fact, its value can be quite close to 0). Similarly, lift may reach values close to the value 1 (that means independence in terms of this measure) even in the cases when the dependence between X and Y is strongest possible positive or strongest possible negative. Thus, we find the dependence factor more accurate measure of a rule constituents’ dependence than the certainty factor and lift.

In this paper, we have: (1) examined the dependence factor as an interestingness measure with respect to the interestingness postulates formulated by Piatetsky-Shapiro in [7], and (2) derived the relationship among the dependence factor for X and Y with those for \(\bar{X}\) and YX and \(\bar{Y}\), as well as \(\bar{X}\) and \(\bar{Y}\), respectively. We have proved that the dependence factor df \((X \rightarrow Y)\) fulfills all Piatetsky-Shapiro interestingness postulates if \(P(Y) \notin \{0, P({ XY}), 1\}\) and P(XY) \(\ne \) 0 or if \(P(X) \notin \{0, P({ XY}), 1\}\) and P(XY) \(\ne \) 0. Otherwise, it fulfills the first two postulates entirely and the third postulate partially as df \((X \rightarrow Y)\) is a non-increasing function rather than decreasing with respect to the marginal probability of an event given the joint probability and the marginal probability of the other event are fixed. On the other hand, it can be observed that several interestingness measures of association rules proposed and/or discussed in the literature does not fulfill all interestingness postulates from [7], including the rule interest ri [7] and novelty [5], which violate the third postulate for zero marginal probabilities.

In this paper, we have found that the character of the (in)dependence between X and Y determines uniquely the character (positive/negative) of the (in)dependence between \(\bar{X}\) and YX and \(\bar{Y}\), as well as \(\bar{X}\) and \(\bar{Y}\), respectively. We have also found that the absolute value of the dependence factors is the same for events and their complements. We find this result justified as the marginal and joint probabilities of events and all their complements depend uniquely on the triple of the probabilities \(\langle P(X), P(Y), P({ XY}) \rangle \).