1 Introduction

The market basket of a shopper, which consists of those product categories which she purchases during a shopping trip at a store (Russell et al. 1997, 1999), is the observable result of multicategory choices. Multicategory choices are often interdependent, i.e., the choice of one category has an effect on the choice of another category. Two categories are considered as purchase complements (purchase substitutes) if their cross effects are positive (negative), i.e., if these categories are purchased jointly more (less) frequently than expected under stochastic independence (Betancourt and Gautschi 1990; Hruschka et al. 1999). Because of these interdependences market basket analysis requires models which measure such cross effects between product categories, whereas using a collection of one-category models would not to be appropriate.

In this paper, we only look at purchase incidences and do not consider other response variables like purchase quantities,Footnote 1 or several of these response variables.Footnote 2 Therefore, we represent the market basket as binary vector whose dimension equals the number of considered categories. Measurement of cross effects between product categories constitutes the main objective of our work.

The models most prominent in the marketing literature which are capable of measuring cross-category effects for purchase incidence data are the multivariate logit (MVL) and the multivariate probit (MVP) models. These models may also include explanatory variables such as marketing instruments or consumer demographics. To our knowledge, Manchanda et al. (1999) published the first application of the MVP model (four categories), followed by Chib et al. (2002) and Duvvuri et al. (2007) who analyzed 12 and 6 categories, respectively. In a recent paper, Duvvuri and Gruca (2010) analyzed four subcategories of the instant coffee category. They specify the overall price sensitivity which is indicated by higher-order factor scores as function of household-level covariates. Russell and Petersen (2000) applied the MVL model to four categories. Boztuğ and Hildebrandt (2008) studied cross-category effects by means of the MVL model for six and five categories.

The overwhelming majority of the previous studies on both multicategory purchase incidence and other multicategory response variables consider rather small numbers of categories. For the MVL model, this limitiation can be explained by the computational intractability of maximum likelihood estimation, as the number of possible market baskets grows exponentially with the number of categories (see Sect. 3 for more details). Extant publications dealing with higher numbers of categories refrain from simultaneously estimating all parameters of a comprehensive model and follow a two-step approach (Hruschka et al. 1999; Boztuğ and Reutterer 2008). In contrast to simultaneous approaches, the results of the first step are not changed in view of estimation results of the second step. Because results of the first step restrict the models estimated in the second step, two-step approaches usually do not attain the statistical performance of simultaneous approaches. This weakness of two-step approaches is well known from, e.g., conjoint analysis models (Wedel and Kamakura 1998).

 Hruschka et al. (1999) estimated one binary logit model for each of 73 categories investigated and determine which interactions with the remaining 72 categories are significant in the first step. In the second step, these authors estimated a model for all 73 categories with interaction terms selected in the first step setting the remaining interaction terms to zero. Boztuğ and Reutterer (2008) started from basket data on 65 categories. In the first step, they used a vector quantization method to determine clusters of market baskets. In the second step they estimated one MVL model for each cluster considering the four or five categories with highest purchase probabilities conditional on the respective cluster.

The model we introduce is known as restricted Boltzmann machine (RBM). The RBM is related to the MVL model from which it differs by the inclusion of binary hidden variables. So far, the RBM has been mostly applied to solve pattern recognition problems, e.g., recognition of handwritten digits or classification of documents (Hinton and Salakhutdinov 2006). Compared to previous work using MVL models the RBM allows us to analyze a higher number of product categories by simultaneous estimation for all categories investigated.

Due to data limitations, the models estimated by us do not include explanatory variables. This procedure can also be justified by the results of two previous studies. The best MVL model with price effects in and Russell and Petersen (2000) performs only slightly better than the MVL model without explanatory variables in terms of log likelihood for the validation data (the former model has six parameters more and a log likelihood values which is higher by 2.92). Similarly, price effects are not statistically significant in the MVL model of Boztuğ and Hildebrandt (2008).

We begin by juxtaposing the specifications of the well-known MVL model and the RBM. Then, we explain maximum likelihood (ML) estimation of RBMs and the independence model and maximum pseudo-likelihood (MPL) estimation of the the MVL model. We also present the criteria we use to evaluate different RBMs, the independence model and the MVL model. We derive an approach to measure asymmetric marginal cross-category effects based on estimation results of a RBM. In the empirical part of our paper, we present descriptive statistics of the data and select one of the estimated models. Based on this model, we interpret the hidden variables and discuss strong and significant cross effects between pairs of categories. In the final chapter, we give an overview of our results, discuss managerial implications, and mention several related future research opportunities.

2 Model

In this section, we present both the MVL and the RBM models to clarify their relationship. Indices of market baskets are \(i=1,\ldots ,I\), indices of categories \(j=1,\ldots ,J\). Basket \(i\) is characterized by a binary column vector of purchase incidences \({\varvec{y}}_{i} = (y_{i1},\ldots ,y_{iJ})^{\prime }\). If market basket \(i\) contains category \(j\), then \(y_{ij}=1\), otherwise \(y_{ij}=0\).

Russell and Petersen (2000) applied the MVL model (which is also known as autologistic model) to market basket analysis drawing upon earlier work of Cox (1972) and Besag (1974). According to the MVL model, the probability of market basket \(i\) is:

$$\begin{aligned} p({\varvec{y}}_i) = \frac{\exp ({\varvec{a}}^{\prime } {\varvec{y}}_i + {\varvec{y}}_{i}^{\prime } {\varvec{V}} {\varvec{y}}_i ) }{Z_\mathrm{{MVL}}} \end{aligned}$$
(1)

with

$$\begin{aligned} Z_\mathrm{{MVL}} = \sum _{{\varvec{y}} \in \{0,1\}^J} \exp ( {\varvec{a}}^{\prime } {\varvec{y}} + {\varvec{y}}^{\prime } {\varvec{V}} {\varvec{y}} ) \end{aligned}$$

Column vector \({\varvec{a}}\) contains constants for the categories. The symmetric \( J \times J\) matrix \({\varvec{V}}\) contains zero diagonal elements (\(V_{jj} = 0\)) and symmetric pairwise cross-category coefficients (\(V_{jl} = V_{lj}\)). Paths in the upper part of Fig. 1 symbolize constants and pairwise cross-category coefficients of the MVL model. A positive (negative) cross-category coefficient \(V_{jl}\) indicates that categories \(j\) and \(l\) are purchased together more (less) frequently compared to conditional independence. On the other hand, a cross category coefficient equal to zero implies conditional independence.

Fig. 1
figure 1

Path diagrams of MVL and RBM models (top MVL, bottom RBM)

The normalization constant \(Z_\mathrm{{MVL}}\) is obtained by summing over the \(2^J\) possible market baskets. Dividing by this normalization constant ensures that expression (1) provides a valid probability. That is why the MVL model can also be interpreted as multinomial logit model which considers choice of one of \(2^J\) alternatives (Russell and Petersen 2000).

Conditional probabilities of purchase incidences \(y_{ij}\) for \(j=1,\ldots ,J\) given purchase incidences of the other \(J-1\) categories \({\varvec{y}}_{i,-j}\) are derived from expression (1). They have binary logit functional forms:

$$\begin{aligned} p(y_{ij} | {\varvec{y}}_{i,-j}) = 1\Bigg /\left(1+\exp \left(-\left(a_j + \sum _ {l \ne j} V_{jl} y_{il} \right) \right) \right)\quad \text{ for} \text{ all}\ i,j \end{aligned}$$
(2)

The RBM introduced by Smolensky (1986) encompasses a column vector \({\varvec{h}}_i = (h_{i1},\ldots ,h_{iK})^{\prime }\) of K binary hidden variables. Each hidden variable is connected to each observed binary variable (here: to each purchase incidence). The RBM is called restricted, because it does not have connections among hidden variables as well as among observed variables (see lower part of Fig. 1).

According to the RBM, the joint probability of market basket \({\varvec{y}}_i\) and hidden variables \({\varvec{h}}_i \) is:

$$\begin{aligned} p({\varvec{y}}_i, {\varvec{h}}_i ) = \frac{\exp ({\varvec{b}}^{\prime } {\varvec{y}}_i+{\varvec{h}}_{i}^{\prime }{\varvec{W}}{\varvec{y}}_i)}{Z_\mathrm{{RBM}}} \end{aligned}$$
(3)

with

$$\begin{aligned} Z_\mathrm{{RBM}} = \sum _{{\varvec{y}} \in \{0,1\}^J} \sum _{{\varvec{h}} \in \{0,1\}^K} \exp ( {\varvec{b}}^{\prime } {\varvec{y}} + {\varvec{h}}^{\prime } {\varvec{W}} {\varvec{y}} ) \end{aligned}$$

Column vector \({\varvec{b}}\) contains constants for the categories. \(K\times J\) Matrix \({\varvec{W}}\) contains coefficients which link each hidden variable to each category. Contrary to the MVL model cross effects between categories do not depend on pairwise coefficients, but on coefficients linking purchase incidence and hidden variables (Sect. 4 deals with measurement of cross effects for the RBM). The normalization constant \(Z_\mathrm{{RBM}}\) is obtained by summing over the \(2^J\) possible market baskets \({\varvec{y}} \in \{0,1\}^J\) and the \(2^K\) possible configurations of hidden variables \({{\varvec{h}} \in \{0,1\}^K}\). Dividing by the normalization constant \(Z_\mathrm{{RBM}}\) ensures that expression (3) gives a valid probability.

From expression (3), we obtain the probability \(p({\varvec{y}}_i)\) of market basket \({\varvec{y}}_i\) by summing over the \(2^K\) possible configurations of hidden variables \({\varvec{h}} \in \{0,1\}^K\):

$$\begin{aligned} p({\varvec{y}}_i) = \frac{p^{*}({\varvec{y}}_i)}{Z_\mathrm{{RBM}}} \quad \text{ with} \quad p^{*}({\varvec{y}}_i) = \sum _{{\varvec{h}} \in \{0,1\}^K} \exp ( {\varvec{b}}^{\prime } {\varvec{y}}_i + {\varvec{h}}^{\prime } {\varvec{W}} {\varvec{y}}_i) \end{aligned}$$
(4)

In the following, we call \(p^{*}({\varvec{y}}_i)\) unnormalized probability.

Conditional probabilities of purchase incidences \(y_{ij}\) given hidden variables \({\varvec{h}}_i\) for \(j=1,\ldots ,J\) and hidden variables \(h_{ik}\) given purchase incidences \({\varvec{y}}_i\) for \(k=1,\ldots ,K\) are derived from expression (3). Quite similar to the MVL we get binary logit functional forms:

$$\begin{aligned}&p(y_{ij} | {\varvec{h}}_i) = 1\Bigg /\left(1+\exp \left(-\left(b_j + \sum _ k W_{kj} h_{ik} \right) \right)\right) \quad \text{ for} \text{ all} \; i,j \end{aligned}$$
(5)
$$\begin{aligned}&p(h_{ik}| {\varvec{y}}_i) = 1\Bigg /\left(1+\exp \left(-\sum _ j W_{kj} y_{ij} \right)\right) \quad \text{ for} \text{ all} \; i,k \end{aligned}$$
(6)

3 Estimation and model comparison

Computation of the normalization constant of the MVL model makes ML estimation for higher numbers of categories impractical, because it requires in each iteration to sum over all \(2^J\) possible market baskets, e.g., over more than \(10^{18}\) market baskets in the case of 60 categories (see expression (1)).

Looking at expression (3) one could get the impression that one faces the same problem with the RBM, but by factorizing over categories, we can rewrite its normalization constant with \({\varvec{W}}_{\cdot ,j }\) denoting the \(j\)th column of \({\varvec{W}}\) as follows (Larochelle et al. 2010):

$$\begin{aligned} Z_\mathrm{{RBM}} = \sum _{{\varvec{h}} \in \{0,1\}^K } \prod _{j=1}^{J} \left(1+ \exp ({\varvec{W}}_{\cdot ,j }^{^{\prime }} {\varvec{h}} + b_j) \right) \end{aligned}$$
(7)

Expression (7) makes clear that estimation is significantly less-involved compared to the MVL model , if the number of hidden variables is much lower than the number of categories. For example, one has to sum and multiply over only \(2^5 \times 60=1{,}920\) terms in the case of five hidden variables and 60 categories.

Therefore a moderate number of hidden variables allows us to estimate RBMs by ML. For ML estimation unnormalized probabilities defined in (4) are rewritten with \({\varvec{W}}_{k ,\cdot }\) denoting the kth row of \({\varvec{W}}\) (Larochelle et al. 2010):

$$\begin{aligned} p^{*}({\varvec{y}}_i) = \exp ( {\varvec{b}}^{\prime } {\varvec{y}}_i) \prod _{k=1}^{K} \left( 1+ \exp ({\varvec{W}}_{k ,\cdot } {\varvec{y}}_i ) \right) \end{aligned}$$
(8)

To our knowledge, our paper is the first to develop and apply ML estimation of RBMs. Of course, estimation of RBMs with a high number of hidden variables is only practical by approximate methods. The most widespread of these approximate methods is contrastive-divergence estimation (Hinton 2002) which in each iteration includes blocked Gibbs updates of expressions (5) and (6).

We estimate coefficients of all considered RBMs by maximizing the log likelihood (LL) across baskets (more details on ML estimation of RBMs can be found in Appendix 1):

$$\begin{aligned} \text{ LL} = \sum _i \log p({\varvec{y}}_i) = \sum _i \log p^{*}({\varvec{y}}_i) - I \log Z_\mathrm{{RBM}} \end{aligned}$$
(9)

According to the independence model, the relative frequency of any basket configuration equals the product of relative marginal frequencies of purchases or non-purchases of the categories. The independence model can be obtained as the special case of the RBM by setting \( {\varvec{W}}=0\). ML estimates of the coefficients of the independence model are:

$$\begin{aligned} b_j = \log \left(\sum _i y_{ij}\right) - \log \left(I-\sum _i y_{ij}\right) \quad \text{ for}\ j=1,\ldots ,J \end{aligned}$$
(10)

The normalization constant \(Z_0\) of the independence model equals:

$$\begin{aligned} Z_0 = 2^J \prod _j (1+\exp (b_j)) \end{aligned}$$
(11)

As ML estimation of the MVL model is not possible for a higher number of categories, we estimate its coefficients by the maximum pseudo-likelihood method (MPL) introduced by Besag (1974). The pseudo-likelihood of the MVL model equals the product of conditional probability functions across all J categories which are derived from expression (2). Its log, the log pseudo-likelihood (LPL), can be written as:

$$\begin{aligned}&\text{ LPL} = \sum _i \text{ LPL}_i \nonumber \\&\quad \text{ with} \quad \text{ LPL}_i \!=\! \sum _j [ Y_{ij} \log ( p(y_{ij} | {\varvec{y}}_{i,-j})) \!+\! (1\!-\!Y_{ij})(1\!-\!\log ( p(y_{ij} | {\varvec{y}}_{i,-j})))]\qquad \qquad \end{aligned}$$
(12)

\(\text{ LPL}_i\) denotes the log pseudo-likelihood of basket \(i\). Comets (1992) showed that MPL estimates are consistent and asymptotically normal . Appendix 2 provides further information on MPL estimation of the MVL model.

The models which we compare consist of RBMs which differ with regard to the number of hidden variables, the independence model and the MVL model. Comparisons of RBMs and the independence model are based on log likelihood values. We measure performance of these models by log likelihood values both for estimation data and for validation data. Log likelihood values for validation data are computed using coefficients determined for the estimation data. In addition, we also compute values of the Bayesian information criterion (BIC) for the models investigated (LL is the loglikelihood value for the estimation data, \(n\) the number of parameters of a model):

$$\begin{aligned} \text{ BIC} = - 2 \text{ LL} + n \, \log I \end{aligned}$$
(13)

To compare the RBM and the MVL model, we cannot use likelihood for the first, and pseudo-likelihood values for the latter. That is why we follow a predictive model selection approach and use the absolute deviation between observed and model generated purchase incidences as evaluation criterion (Carlin and Louis 2000; Laud and Ibrahim 1995). We generate 500 artificial datasets each with 10,000 market baskets by Gibbs sampling over the appropriate conditional probability functions (i.e., for the MVL model over expression (2) using MPL coefficients and for the RBM over expressions (5) and (6) using ML coefficients). We define the absolute deviation (AD) with \(\hat{y}_{ijs}\) denoting the \(s\)th sampled purchase incidence value of category \(j\) in basket \(i\) as follows:

$$\begin{aligned} \text{ AD} = \frac{1}{500} \sum _{s=1}^{500} \sum _i \sum _j |y_{ij} - \hat{y}_{ijs}| \end{aligned}$$
(14)

Gibbs sampling produces samples from the joint distribution (expressions (1) and (3) for the MVL and the RBM, respectively) by sampling from the conditional distribution of each variable conditional on current values of the other variables. In their seminal paper, Geman and Geman (1984) proved that samples generated this way are a Markov chain with a unique stationary distribution which equals the joint distribution (also see Casella and George 1992 for another explanation of Gibbs sampling). Blockwise sampling is possible for RBMs, because according to their respective conditional distributions, purchases only depend on hidden variables (expression (5)) and hidden variables only depend on purchases (expression (6)).

We assess the statistical significance of coefficients of the RBM by parametric bootstrap (see, e.g., Cameron and Trivedi 2005). For each of 500 artificial data sets generated on the basis of the RBM by blockwise Gibbs sampling we compute ML estimates. We calculate \(t\)-statistics of coefficients by dividing means by standard deviations across these 500 estimates. The parametric bootstrap allows to asses significance of quantities which are functions of coefficients quite easily using these 500 estimates of coefficient vectors.

4 Marginal cross effects

To derive marginal cross effects, we apply mean field theory (Hertz et al. 1991; Salakhutdinov and Hinton 2010) and replace the fluctuating values of purchases of categories and hidden variables by shares, i.e., average values. Then the share \(\langle y_j\rangle \) of category \(j\) and the share \( \langle h_k \rangle \) of hidden variable \(k\) can be calculated as:

$$\begin{aligned} \langle y_j \rangle&= 1\bigg /\left(1+\exp \left(-\left(b_j + \sum _ k W_{kj} \, \langle h_{k} \rangle \right) \right)\right)\end{aligned}$$
(15)
$$\begin{aligned} \langle h_k \rangle&= 1\bigg /\left(1+\exp \left(-\sum _ j W_{kj} \, \langle y_j \rangle \right)\right) \end{aligned}$$
(16)

We define the marginal cross effect of category \(l\) on category \(j\) by the first derivative of the purchase share of category \(j\) given in expression 15 with respect to the purchase share of another category \(l\):

$$\begin{aligned} \frac{\partial \langle y_j \rangle }{\partial \langle y_l \rangle }&= \langle y_j \rangle \, (1 - \langle y_j \rangle ) \sum _k W_{kj} \frac{\partial \langle h_{k} \rangle }{\partial \langle y_l \rangle } \nonumber \\&= \langle y_j \rangle \, (1 - \langle y_j \rangle ) \sum _k W_{kj} W_{kl} \langle h_k \rangle \, (1-\langle h_k \rangle ) \end{aligned}$$
(17)

A positive (negative) cross effect indicates that a higher share of category \(l\) is associated with a higher (lower) share of category \(j\). Marginal cross effects are positive (negative) if \(W_{kj} W_{kl} > 0\) (\(W_{kj} W_{kl} < 0\)) for each \(k\). In other words, the cross effects between two categories are positive (negative) if their coefficients have the same (different) signs for each hidden variable. Note that marginal cross effects are asymmetric (i.e., \(\partial \langle y_j \rangle / \partial \langle y_l \rangle \ne \partial \langle y_l \rangle / \partial \langle y_j \rangle \)) except for special constellations (\( \langle y_j \rangle = \langle y_l \rangle \) or \( \langle y_j \rangle = 1- \langle y_l \rangle \)). Marginal cross effects for two independent categories \(l\) and \(j\) are zero which follows from the fact that their coefficients \(W_{kj}, W_{kl}\) equal zero for each \(k\).

Asymmetric marginal cross effects are not peculiar to the RBM. Although coefficients of the MVL are symmetric (i.e., \(V_{jl}=V_{lj}\)), we obtain asymmetric marginal cross effects for this model as well, because \( \langle y_j \rangle \, (1 - \langle y_j \rangle ) V_{jl}\) usually differs from \( \langle y_l \rangle \, (1 - \langle y_l \rangle ) V_{jl}\).

For the RBM, estimates of \( \langle y_j \rangle \) and \(\langle h_k \rangle \) cannot be determined analytically. We therefore set \( \langle y_j \rangle \) and \(\langle h_k \rangle \) in expression (17) equal to averages of categories and averages of hidden variables in 10,000 pseudo-baskets. We generate these pseudo-baskets by blockwise Gibbs sampling of expressions (5) and (6) using coefficients estimated by ML. We repeat this procedure for each of the 500 coefficient vectors obtained by parametric bootstrap which was described in Sect. 3. Dividing means by standard deviations over 500 estimated marginal cross effects gives \(t\) values of marginal cross effects.

We interpret each hidden variable by its larger (in absolute terms) partial marginal cross effects. We define a partial marginal cross effect between categories \(j\) and \(l\) as contribution of hidden variable \(k\) to the overall marginal cross effect between these two categories:

$$\begin{aligned}&\langle y_j \rangle (1 - \langle y_j \rangle ) \, W_{kj} W_{kl} \, \langle h_k \rangle \, (1-\langle h_k \rangle ) \quad \text{ for} \quad k=1,\ldots ,K \end{aligned}$$
(18)

5 Empirical study

5.1 Data

Twenty-thousand market baskets collected at a medium-sized supermarket during four consecutive saturdays are randomly divided into two data sets of equal size. One set (estimation data) is required for estimation, the second set (validation data) is used to determine the predictive accuracy of models. Out of a total of 209 categories, we analyze the 60 categories with the highest univariate marginal purchase frequencies (see Table 1). This number of categories is five times as big as the maximum number of categories in the previous studies in which MVP or MVL models are simultaneously estimated (see Sect. 1). Table 2 contains the relative bivariate marginal frequencies for the 60 most frequent pairs of product categories.

Table 1 Relative univariate marginal frequencies of analyzed categories

 

Table 2 Relative bivariate marginal frequencies

 

5.2 Estimation results

All the RBMs considered perform better than the independence model in terms of log likelihood values and values of the Bayesin information criterion. This result clearly shows that cross effects between categories should not be ignored.

Log likelihood values of RBMs increase with the number of hidden variables. Values of the Bayesian information criterion improve with the number of hidden variables also. RBMs turn out to be robust, as log likelihood values almost stay the same for the validation data (Table 3).

Table 3 Comparison of alternative models

Whereas the RBM with four hidden variables implies 14 additional significant cross effects compared to the RM with three hidden variables (see Table 4), the significant cross effects of the RBM with five hidden variables consist of the same category pairs like the RBM with four hidden variables. Moreover, bootstrapped \(t\) values of differences of cross effects between these models are not significant. As we are mainly interested in cross effects, we present only estimation results of the less-complex RBM with four hidden variables in the following, though log likelihood values and values of the Bayesian information criterion of the RBMs with five hidden variables are better.

By computing absolute deviations as explained in Sect. 3, we obtain rounded values of \(117,796\) and \(97,818\) for the MVL model and the RBM model with four hidden variables, respectively. We think that the lower number of coefficients is one reason for the better performance of the RBM. For 60 categories one has to estimate \(1{,}830 = 60 \,+\, 60 \,\times \, 59/2\) coefficients for the MVL model, but only \(300=60 \,+\, 4 \,\times \, 60 \) coefficients for the RBM model with four hidden variables.

Marginal cross effects implied by the RBM with four hidden variables whose absolute \(t\) values are >1.65 are given in Table 4 and also illustrated by the graph of Fig. 2.

Table 4 Significant marginal cross effects for the RBM with four hidden variables
Fig. 2
figure 2

Graph of marginal cross effects (solid: positive, dashed: negative)

For positive (negative) \(W_{kj}\), the conditional probability of hidden variable \(k\) being equal to one (zero) increases if category \(j\) is purchased. The conditional probability of the first hidden variable being equal to one (zero) increases, for example, if beer (bread) is purchased. To give another example, the conditional probability of the fourth hidden variable being equal to one (zero) increases if cigarettes (vegetables) are purchased. Table 5 provides coefficients \(W_{kj}\) linking hidden variables and categories whose absolute values are greater than \(1.10\). A value greater than \(1.10\) (less than \(-1.10\)) means that the odds of hidden variable \(k\) defined as \(p(h_{ik}| {\varvec{y}}_i)/(1-p(h_{ik}| {\varvec{y}}_i))\) are multiplied by a factor greater than 3 (less than \(1/3\)) if category \(j\) is purchased (these factors equal \(\exp {(W_{kj})}\)).

Table 5 Selected \(W_{kj}\) coefficients

We interpret the four hidden variables by means of partial marginal cross effects as defined in expression 18 (see Table 6). All four hidden variables contribute to the positive interdependence of fruits and vegetables. Hidden variables 1 and 2 indicate that fruits and vegetables exert negative effects on sliced cheese.

Table 6 Partial marginal cross effects for each hidden variable

We now deal with partial cross effects which are specific to a hidden variable. Hidden variable 1 contributes on one hand to negative effects of periodicals, bread, milk, and yogurt on beer, on the other hand to negative effects of beer on fruits and vegetables. Hidden variable 3 reflects positive effects of both fruits and vegetables on soft drinks and tea. For hidden variable 4, both fruits and vegetables excert positive effects on fat, soups, and spices, negative effects on chocolate and cigarettes. Moreover, hidden variable 4 contributes to negative effects of both periodicals and cigarettes on vegetables and fruits.

Note that none of the categories having the 26 lowest univariate marginal frequencies is part of any significant cross effect (in Table 1 these are the categories beginning with the category gifts). Fruits and vegetables form the center of the graph of Fig. 2 as they are part of each of the significant cross effects. Fruits and vegetables are interdependent and exert a positive effect on milk, yogurt, and cheese. Milk, yogurt and cheese, on the other hand, have a positive effect on vegetables and fruits.

The majority of the cross effects of fruits and vegetables run one-way. These cross effects are positive with respect to water, butter, pasta, spices, soups, fat, eggs, seasonal, pickles, tea, delicatessen 1 and UHT milk.

By also looking at Table 1, it is possible to examine whether more frequently purchased categories tend to exert significant cross effects or tend to be subject to significant cross effects. Equating categories with high purchase frequencies with categories being part of high marginal cross effects turns out to be misleading (e.g., for six of the nine most frequently purchased categories, namely bread, periodicals, chocolate, soft drinks, cigarettes). This result can be explained by a relative high share of purchases whose baskets do not contain another category for these categories. The lowest share of such single category baskets amounts to 0.073 for these six categories. The corresponding highest share for the remaining three of the nine most frequently purchased categories (namely fruits, yogurt, milk) equals 0.003.

Moreover, comparing Table 4 and Fig. 2 to Table 2 makes it clear that the results of our model differ from those of a simple pairwise cross-tabulation of purchase incidences in several respects:

  • High pairwise frequencies of bread and periodicals with other categories are due to the high relative marginal purchase frequencies of bread and periodicals and not related to cross effects with other categories, i.e., the corresponding marginal cross effects are not significant (bread with fruits, vegetable, yogurt, milk, periodicals, sliced cheese, butter and cheese; periodicals with bread, fruits and vegetables).

  • Of course, it is not possible to discover negative marginal cross effects by means of pairwise purchase frequencies (for our data this holds for the cross effects of fruits and vegetables on both beer and sliced cheese).

  • High pairwise purchase frequencies between yogurt and milk, cheese and butter are traced back to the high cross effect of fruits and vegetables on these categories.

6 Conclusion

Extending the well-known MVL model by adding binary hidden variables enables us to analyze a large number of categories. For a moderate number of hidden variables, we can estimate the resulting model by ML and do not have to resort to two-step estimation methods in contrast to the previous studies which dealt with higher number of categories. To compare the RBM to the MVL, we estimate the latter by maximum pseudo-likelihood. For a validation data set, the RBM performs better than the MVL model estimated this way.

For our data, about 75 % of the baskets are reproduced by the independence model [(75 % is the relative frequency of hidden variables \(h_k = 0\) for all \(k\) as estimated by blockwise Gibbs sampling over expressions (5) and (6)]. Still, 25 % of the baskets obviously cannot be adequately modeled if cross effects among categories are ignored. Log likelihood values of the RBM and the independence model both for estimation and validation data as well als BIC values of these models also give evidence to this fact.

On the basis of our model, we infer asymmetric marginal cross effects among variables. The number of significant cross effects we obtain appears to be manageable, i.e., 38 cross effects as opposed to the 1,770 possible pairwise effects for 60 categories. We found only four negative cross effects between categories. Our results show that assuming that categories with high marginal cross effects are simply categories with high marginal purchase frequencies or categories with high pairwise purchase frequencies may be misleading.

A study of Russel and Kamakura (1997) shows that due to positively correlated category preferences promotions in one category lead to higher sales not only in that category but also in other positively related categories. Russell and Petersen (2000) emphasize strong positive cross-category effects, because promotional activity in one category simultaneously causes higher sales in other related categories.

Therefore, knowledge on categories which are part of significant cross effects is of benefit to a manager who wants to increase the number of purchases in other categories. To this end, she can make the independent category more attractive to customers by, e.g., sales promotion, advertising or assortment decisions. We give a few examples for the supermarket whose data we analyzed.

Marketing measures in categories which have no effect on purchases of other categories are not appropriate to increase purchases in other categories. For our data, many most-frequent categories share this property, e.g., bread, periodicals, chocolate, and cigarettes.

If purchases across the assortment should be increased, marketing measures for fruits or vegetables are obvious because of the centrality of these two categories in terms of cross effects. Marketing measures for fruits and vegetables are also advisable if purchases should be increased in categories which are positively influenced (milk, yogurt, cheese, water, butter, pasta, spices, soups, fat, eggs, seasonal, tea, pickles, delicatessen 1, UHT milk). Marketing measures for milk, yogurt and cheese, on the other hand, are suitable to increase purchases of fruits and vegetables.

Assuming, e.g., the retailer wants to increase cheese purchases, the manager should undertake marketing measures for fruits or vegetables. Similar measures in categories like, e.g., bread, rolls, red wine, or beer, are not appropriate to increase cheese purchases. Though these products are often consumed together with cheese, they have no effect on cheese purchases.

Marketing measures in other categories do not make sense if the retailer wants to increase purchases of beer or sliced cheese. Management should even avoid such measure for the vegetables or fruits categories, because they have negative effects on beer or sliced cheese purchases.

Applications of the RBM in other application areas, e.g., non-food retailing or financial services, may be fruitful. Adding explanatory variables seems to be an obvious next research task. The development of similar models for market basket analysis based on other response variables (e.g., purchase rates or purchase quantities), though somewhat challenging, constitute interesting research problems.