1 Introduction

The aggregation of information is a fundamental process in many fields of science. Many times, this process is carried out through functions that, from a given data set, return a single value. Obviously, the choice of the function is of crucial importance in this process (see, for instance, Zhang and Xu 2014). For instance, in the field of statistics, it is well known that the mean is quite sensitive to extreme values and, consequently, the sample mean is not a robust estimator of the population mean. So, the mean value may not be very representative in data where outliers may occur. For overcoming this shortcoming, other estimators have been proposed in the literature (see, for instance, Huber and Ronchetti 2009). Two of the most popular are the trimmed means and the Winsorized means. In the trimmed means, the lowest and the highest values are removed before calculating the mean whereas in the Winsorized means they are replaced with the less extreme adjacent values.

Trimmed and Winsorized means can also be seen as symmetric aggregation functions. In fact, they are specific cases of OWA operators (Yager 1988): Trimmed means are OWA operators associated with weighting vectors of the form \((0, \ldots ,0,1/m, \ldots ,1/m,0, \ldots ,0)\) (where we are supposing that \(n-m\) values have been removed) while Winsorized means are OWA operators associated with weighting vectors of the form \((0, \ldots ,0,(r+1)/n,1/n, \ldots ,1/n,(s+1)/n,0, \ldots ,0)\) (where we are supposing that the r lowest values and the s highest values have been replaced). The use of symmetric functions means that all information sources (for instance, criteria or experts’ opinions) are treated equally. However, in many practical applications (for instance, in many multiattribute decision making (MADM) problems) it is necessary to consider that the information sources have different importance. But, as what we have previously reported for the mean value, the use of the weighted mean, which is known in the MADM literature as the simple additive weighting (SAW) method, may not be adequate in presence of outliers. For this reason, several proposals have been suggested in the literature with the purpose of introducing different weights for the information sources in an OWA-type aggregation [see, among others, Torra (1997), Llamazares (2013, 2015a), and Beliakov and Dujmović (2016)].

It is interesting to note that the three families of functions introduced by the above-mentioned authors are specific cases of the Choquet integral (Choquet 1953). Since its appearance, the Choquet integral has received increased attention from the scientific community, due mainly to its simplicity, versatility and good properties (for instance, it is well known that the Choquet integral is continuous, monotonic, idempotent, compensative, and ratio scale invariant; see Grabisch et al 2009). For these reasons, the Choquet integral has been a widely used tool in economics to deal with problems related to decision making under risk and uncertainty, finance, insurance, social welfare and quality of life (see, for instance, Heilpern 2002). Likewise, the integral of Choquet has also received much attention in the MADM literature because it allows to take into account the interaction that usually exists between the information sources, which is very useful in this field (among the vast literature that exists on this topic, see, for instance, Grabisch 1995; Grabisch and Labreuche 2010, 2016).

The aim of this paper is to present a natural generalization of the Winsorized means that allows us to consider weighting vectors for the information sources. The new functions, which we will call Winsorized weighted means, are also a specific case of the Choquet integral whose capacities are known. The analysis of these functions is carried out through the study of several indices that provide us with a more precise knowledge about their behavior in the aggregation processes. In this sense, we give closed-form expressions for the following indices: the orness degree (Marichal 2004), k-conjunctiveness and k-disjunctiveness indices (Marichal 2007), veto and favor indices (Marichal 2004, 2007), Shapley values (Shapley 1953), and interaction indices (Owen 1972; Murofushi and Sugeno 1993); and also for their Möbius transform (Rota 1964). Of particular interest are the closed-form expressions obtained for the Shapley values because they reflects the global importance of each information source (as it is well known, in the case of a Choquet integral with respect to a capacity \(\mu \), the importance of the ith information source is no given by \(\mu (\{i\})\) but by importance indices, and, among them, Shapley values are usually the most used). From them, we can determine the weights that allow us to get Shapley values previously fixed, and, in this way, we can aggregate data so that each information source has the desired weighting and outliers have no influence in the aggregated value (the procedure used will be illustrated with an example in Sect. 5).

The remainder of the paper is organized as follows. In Sect. 2 we recall some basic concepts on Choquet integrals. Section 3 is devoted to introduce the Winsorized weighted means. In Sect. 4 we show the main results of the paper: the Möbius transform and several indices of the Winsorized weighted means are given in closed-form expressions. In Sect. 5 we show by an example how Winsorized weighted means can be applied. Finally, some concluding remarks are provided in Sect. 6. All proofs are given in the Appendix.

2 Choquet Integral

The following notation will be used throughout the paper: \(N=\{1, \ldots ,n\}\); given \(T\subseteq N\), |T| will denote the cardinality of T; and vectors will be denoted in bold. Given a vector \(\varvec{x} \in \mathbb {R}^n\), \((\cdot )\) and \([\cdot ]\) will denote permutations such that \(x_{(1)}\le \cdots \le x_{(n)}\) and \(x_{[1]}\ge \cdots \ge x_{[n]}\).

Choquet integral was introduced by Choquet (1953) by using the concept of capacity (Choquet 1953), which was also introduced in an independent way by Sugeno (1974) under the name of fuzzy measure.

Definition 1

A capacity (or fuzzy measure) \(\mu \) on N is a set function, \(\mu :2^{N} \longrightarrow [0,\infty )\) satisfying \(\mu (\varnothing )=0\) and \(\mu (A)\le \mu (B)\) whenever \(A\subseteq B\). A capacity \(\mu \) is said to be normalized if \(\mu (N)=1\).

The Choquet integral is a functional that generalizes the Lebesgue integral (see, for instance, Choquet 1953, Murofushi and Sugeno 1991, and Denneberg 1994). Nevertheless, in the discrete case, it can be seen as an aggregation function over \(\mathbb {R}^n\) (Grabisch et al 2009, p. 181). Notice that we define the Choquet integral for all vectors of \(\mathbb {R}^n\) instead of nonnegative vectors given that we are actually considering the asymmetric Choquet integral with respect to \(\mu \) (on this, see again Grabisch et al 2009, p. 182).

Definition 2

Given a capacity \(\mu \) on N, the Choquet integral with respect to \(\mu \) is the function \(\mathcal {C}_{\mu }:\mathbb {R}^n \longrightarrow \mathbb {R}\) defined by

$$\begin{aligned} \mathcal {C}_{\mu } (\varvec{x}) = \sum _{i=1}^n \mu (A_{(i)})\big (x_{(i)}-x_{(i-1)}\big ), \end{aligned}$$

where \(A_{(i)}=\{(i), \ldots ,(n)\}\), and we adopt the convention that \(x_{(0)}=0\).

Weighted means and OWA operators (Yager 1988) are two of best well-known specific cases of Choquet integrals. Both are defined through weighting vectors; that is, non-negative vectors whose components sum to one.Footnote 1

Definition 3

  1. 1.

    Given a weighting vector \(\varvec{p}\), the weighted mean associated with \(\varvec{p}\) is the function \(M_{\varvec{p}}:\mathbb {R}^n \longrightarrow \mathbb {R}\) defined by

    $$\begin{aligned} M_{\varvec{p}}(\varvec{x}) = \sum _{i=1}^n p_i x_i. \end{aligned}$$
  2. 2.

    Given a weighting vector \(\varvec{w}\), the OWA operator associated with \(\varvec{w}\) is the function \(O_{\varvec{w}}:\mathbb {R}^n \longrightarrow \mathbb {R}\) defined by

    $$\begin{aligned} O_{\varvec{w}}(\varvec{x}) = \sum _{i=1}^n w_i x_{[i]}, \end{aligned}$$

    where as we have said, \([\cdot ]\) is a permutation on N such that \(x_{[1]}\ge \cdots \ge x_{[n]}\).

3 Winsorized Weighted Means

Winsorized means received this name in honor of Winsor, who suggested replacing the magnitude of extreme observations by the magnitude of the next largest (or smallest) observation (see Dixon 1960; Wainer 1976). Although it is common to define the Winsorized means through the proportion of the replaced values, we will use, for convenience, the number of replaced values (see, for instance, Barnett and Lewis 1994).

Definition 4

Given an integer r, with \(0\le r < n/2\), the r-fold Winsorized mean is defined by

$$\begin{aligned} M^{r}(\varvec{x}) = \frac{1}{n}\left( r x_{(r+1)} + \sum _{i=r+1}^{n-r} x_{(i)} + r x_{(n-r)}\right) \end{aligned}$$

Notice that the arithmetic mean (when \(r=0\)) and the median (when r is the largest integer less than n / 2) are specific cases of the Winsorized means. The above definition can be easily extended to consider asymmetrical Winsorized means. For this, we will use the following notation:

$$\begin{aligned} \mathcal {R}=\{(r,s)\in \{0,1, \ldots ,n-1\}^2 \;\vert \; r + s < n\}. \end{aligned}$$

Definition 5

Let \((r,s)\in \mathcal {R}\). The (rs)-fold Winsorized mean is defined by

$$\begin{aligned} M^{(r,s)}(\varvec{x}) = \frac{1}{n}\left( r x_{(r+1)} + \sum _{i=r+1}^{n-s} x_{(i)} + s x_{(n-s)}\right) . \end{aligned}$$

Note that the kth order statistic \(({{\,\mathrm{OS}\,}}_k(\varvec{x}) = x_{(k)}\)) is obtained with the pair \((k-1,n-k)\in \mathcal {R}\); and the extreme cases of these pairs allow us to get the minimum and the maximum. Obviously, asymmetrical Winsorized means are specific cases of OWA operators.

Winsorized means can also be easily extended to include a weighting vector \(\varvec{p}\): Given \(\varvec{x}\), it is sufficient to replace the r lowest values and the s highest values by \(x_{(r+1)}\) and \(x_{(n-s)}\), respectively, and to consider the weighted mean associated with \(\varvec{p}\).

Definition 6

Let \(\varvec{p}\) be a weighting vector and \((r,s)\in \mathcal {R}\). The (rs)-fold Winsorized weighted mean is defined by

$$\begin{aligned} M_{\varvec{p}}^{(r,s)}(\varvec{x}) = \left( \sum _{i=1}^{r} p_{(i)} \right) x_{(r+1)} + \sum \limits _{i=r+1}^{n-s} p_{(i)}x_{(i)} + \left( \sum _{i=n-s+1}^{n} p_{(i)} \right) x_{(n-s)}. \end{aligned}$$

Notice that \(M_{\varvec{p}}^{(0,0)} = M_{\varvec{p}}\) for any weighting vector \(\varvec{p}\), and \(M_{\varvec{p}}^{(k-1,n-k)} = {{\,\mathrm{OS}\,}}_k\) for any \(k\in N\) and any weighting vector \(\varvec{p}\). Moreover, it is worth emphasizing that Winsorized weighted means have been obtained as specific cases of SUOWA operators (Llamazares 2018b); which are in turn particular cases of the Choquet integral [on SUOWA operators see Llamazares (2015a, b, 2016a, b, 2018a, b, 2019a, b)]. It is easy to see (Llamazares 2018b) that the capacity associated with the (rs)-fold Winsorized weighted mean is

$$\begin{aligned} \mu _{\varvec{p}}^{(r,s)}(T) = {\left\{ \begin{array}{ll} 0, &{} \text {if } |T| \le s, \\ \sum \nolimits _{i\in T} p_{i}, &{} \text {if } s< |T| < n-r, \\ 1, &{} \text {if } |T| \ge n-r, \end{array}\right. } \end{aligned}$$
(1)

where \(T\subseteq N\). In the specific case of order statistics, where \(r+s=n-1\), the capacity is

$$\begin{aligned} \mu _{\varvec{p}}^{(r,s)}(T) = {\left\{ \begin{array}{ll} 0, &{} \text {if } |T| \le s, \\ 1, &{} \text {if } |T| \ge n-r = s + 1. \end{array}\right. } \end{aligned}$$
(2)

4 The Results

There exist in the literature several indices such as the orness and andness degrees, the tolerance indices, the importance and interaction indices, etc., that allow us to know the behavior of the functions used in the aggregation processes. The aim of this section is to analyze the Winsorized weighted means by showing closed-form expressions of the following indices: the orness degree, k-conjunctiveness and k-disjunctiveness indices, veto and favor indices, the Shapley values, and interaction indices. Moreover, we also provide the expression of the Möbius transforms of the capacities associated with the Winsorized weighted means.

4.1 The Orness Degree

The notion of orness was introduced by Yager (1988) in the field of OWA operators to measure the degree to which the aggregation is disjunctive (i.e., it is like an or operation). Subsequently, and by using the notion of average value, Marichal (1998, 2004) generalized it to the case of Choquet integrals and gave an expression in terms of the capacity.

Remark 1

Let \(\mu \) be a normalized capacity on N. Then

$$\begin{aligned} {{\,\mathrm{orness}\,}}(\mathcal {C}_{\mu })=\frac{1}{n-1}\sum _{t=1}^{n-1}\frac{1}{\left( {\begin{array}{c}n\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N \\ |T|=t \end{array}}\mu (T). \end{aligned}$$

The orness degree of the functions \(M_{\varvec{p}}^{(r,s)}\) is shown in the next proposition.

Proposition 1

Let \(\varvec{p}\) be a weighting vector and \((r,s)\in \mathcal {R}\). Then,

$$\begin{aligned} {{\,\mathrm{orness}\,}}\big (M_{\varvec{p}}^{(r,s)}\big ) = \frac{1}{2} + \frac{r(r+1) - s(s+1)}{2n(n-1)}. \end{aligned}$$

It is worth noting that the orness degree of \(M_{\varvec{p}}^{(r,s)}\) does not depend on the weighting vector \({\varvec{p}}\); that is, \({{\,\mathrm{orness}\,}}\big (M_{\varvec{p}}^{(r,s)}\big ) = {{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big )\) for any weighting vector \({\varvec{p}}\). Other immediate consequences of the previous proposition are gathered in the following corollary.

Corollary 1

Let \((r,s)\in \mathcal {R}\). Then:

  1. 1.

    If \(r>s\) then \({{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big ) > 0.5\); if \(r=s\) then \({{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big )=0.5\); and if \(r<s\) then \({{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big )<0.5\).

  2. 2.

    \({{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big ) = 1 - {{\,\mathrm{orness}\,}}\big (M^{(s,r)}\big )\).

  3. 3.

    If \(r+s \le n-2\), then

    $$\begin{aligned} {{\,\mathrm{orness}\,}}\big (M^{(r+1,s)}\big )&= {{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big ) + \frac{r+1}{n(n-1)}, \\ {{\,\mathrm{orness}\,}}\big (M^{(r,s+1)}\big )&= {{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big ) - \frac{s+1}{n(n-1)}. \end{aligned}$$
  4. 4.

    If \(r+s \le n-3\), then

    $$\begin{aligned} {{\,\mathrm{orness}\,}}\big (M^{(r+1,s+1)}\big ) = {{\,\mathrm{orness}\,}}\big (M^{(r,s)}\big ) + \frac{r-s}{n(n-1)}. \end{aligned}$$

The properties given in the previous corollary can be easily observed in Table 1, where we show the orness degree of the functions \(M^{(r,s)}\) when \(n=5\).

Table 1 Orness degree of the functions \(M^{(r,s)}\) when \(n=5\)

Notice that Table 1 has a triangular structure (there are no values below the main diagonal of the table) so that the values increase as we move down or to the right. Moreover, in the main diagonal appear the orness degree of the order statistics (from the minimum to the maximum), and in the upper right corner appears the orness degree of the weighted mean.

It is also interesting to note that the properties given in Corollary 1 are reflected in the structure of Table 1:

  1. 1.

    The values below the secondary diagonal are greater than 0.5; the values in the secondary diagonal are 0.5; and the values above the secondary diagonal are less than 0.5.

  2. 2.

    The symmetric values with respect to the secondary diagonal sum to one.

  3. 3.

    The value of a cell in the row labeled r increases the amount \(\frac{r+1}{n(n-1)}\) when we move a position down; the value of a cell in the column labeled s decreases the amount \(\frac{s+1}{n(n-1)}\) when we move a position to the left (alternatively, the value of a cell in the column labeled s increases the amount \(\frac{s}{n(n-1)}\) when we move a position to the right).

  4. 4.

    The value of a cell labeled (rs) varies the amount \(\frac{r-s}{n(n-1)}\) when we move down to the left (alternatively, the value of a cell labeled (rs) increases the amount \(\frac{r+s+1}{n(n-1)}\) when we move down to the right).

4.2 k-Conjunctiveness and k-Disjunctiveness Indices

The notions of k-conjunctive and k-disjunctive Choquet integrals (which were originally called at most k-intolerant and at most k-tolerant Choquet integrals) were introduced by Marichal (2007) to determine the conjunctive/disjunctive character of aggregation (see also Komorníková and Mesiar 2011).

Definition 7

Let \(k\in N\) and let \(\mu \) be a normalized capacity on N.

  1. 1.

    \(\mathcal {C}_{\mu }\) is k-conjunctive if \(\mathcal {C}_{\mu } \le {{\,\mathrm{OS}\,}}_k\); i.e., \(\mathcal {C}_{\mu }(\varvec{x}) \le x_{(k)}\) for any \(\varvec{x} \in \mathbb {R}^n\).

  2. 2.

    \(\mathcal {C}_{\mu }\) is k-disjunctive if \(\mathcal {C}_{\mu } \ge {{\,\mathrm{OS}\,}}_{n-k+1}\); i.e., \(\mathcal {C}_{\mu }(\varvec{x}) \ge x_{(n-k+1)}\) for any \(\varvec{x} \in \mathbb {R}^n\).

Since k-conjunctive and k-disjunctive Choquet integrals are infrequent in practice, Marichal (2007) suggested two indices for measuring the degree to which a Choquet integral is k-conjunctive or k-disjunctive.

Definition 8

Let \(k\in N{\setminus } \{n\}\) and let \(\mu \) be a normalized capacity on N. The k-conjunctiveness and k-disjunctiveness indices for \(\mathcal {C}_{\mu }\) are defined by

$$\begin{aligned} {{\,\mathrm{conj}\,}}_{k}(\mathcal {C}_{\mu })&= 1 - \frac{1}{n-k}\sum _{t=1}^{n-k}\frac{1}{\left( {\begin{array}{c}n\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N \\ |T|=t \end{array}}\mu (T), \\ {{\,\mathrm{disj}\,}}_{k}(\mathcal {C}_{\mu })&= \frac{1}{n-k}\sum _{t=k}^{n}\frac{1}{\left( {\begin{array}{c}n\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N \\ |T|=t \end{array}}\mu (T) - \frac{1}{n-k} = \frac{1}{n-k}\sum _{t=k}^{n-1}\frac{1}{\left( {\begin{array}{c}n\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N \\ |T|=t \end{array}}\mu (T). \end{aligned}$$

In the following propositions we show the k-conjunctiveness and k-disjunctiveness indices for the functions \(M_{\varvec{p}}^{(r,s)}\).

Proposition 2

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\), and \(k\in N{\setminus } \{n\}\). Then,

$$\begin{aligned} {{\,\mathrm{conj}\,}}_{k}\!\big (M_{\varvec{p}}^{(r,s)}\big ) = {\left\{ \begin{array}{ll} 1 &{} \text {if } k\ge n-s,\\ \frac{n(n-1) + s(s+1) - k(k-1)}{2n(n-k)} &{} \text {if } r< k < n-s,\\ \frac{n(n-1) + s(s+1) - r(r+1)}{2n(n-k)} &{} \text {if } k \le r. \end{array}\right. } \end{aligned}$$

Proposition 3

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\), and \(k\in N{\setminus } \{n\}\). Then,

$$\begin{aligned} {{\,\mathrm{disj}\,}}_{k}\!\big (M_{\varvec{p}}^{(r,s)}\big ) = {\left\{ \begin{array}{ll} 1 &{} \text {if } k\ge n-r,\\ \frac{n(n-1) + r(r+1) - k(k-1)}{2n(n-k)} &{} \text {if } s< k < n-r,\\ \frac{n(n-1) + r(r+1) - s(s+1)}{2n(n-k)} &{} \text {if } k \le s. \end{array}\right. } \end{aligned}$$

As in the case of the orness degree, k-conjunctiveness and k-disjunctiveness indices do not depend on the weighting vector \({\varvec{p}}\); that is, \({{\,\mathrm{conj}\,}}_{k}\!\big (M_{\varvec{p}}^{(r,s)}\big ) = {{\,\mathrm{conj}\,}}_{k}\!\big (M^{(r,s)}\big )\) and \({{\,\mathrm{disj}\,}}_{k}\!\big (M_{\varvec{p}}^{(r,s)}\big ) = {{\,\mathrm{disj}\,}}_{k}\!\big (M^{(r,s)}\big )\) for any weighting vector \({\varvec{p}}\). Notice also that \({{\,\mathrm{conj}\,}}_{k}\!\big (M^{(r,s)}\big ) = {{\,\mathrm{disj}\,}}_{k}\!\big (M^{(s,r)}\big )\).

4.3 Veto and Favor Indices

The notions of veto and favor (originally called dictator) were suggested by Dubois and Koning (1991) in the framework of the social choice functions.

Definition 9

Let \(j\in N\) and let \(\mu \) be a normalized capacity on N.

  1. 1.

    \(j\,\) is a veto for \(\mathcal {C}_{\mu }\) if \(\mathcal {C}_{\mu }(\varvec{x}) \le x_j\) for any \(\varvec{x} \in \mathbb {R}^n\).

  2. 2.

    \(j\,\) is a favor for \(\mathcal {C}_{\mu }\) if \(\mathcal {C}_{\mu }(\varvec{x}) \ge x_j\) for any \(\varvec{x} \in \mathbb {R}^n\).

Since veto and favor criteria are rather unusual, Marichal (2004, 2007) proposed two indices to measure the degree with which the behavior of a criterion is like a veto or a favor.

Definition 10

Let \(j\in N\) and let \(\mu \) be a normalized capacity on N. The veto and favor indices of criterion \(j\,\) with respect to \(\mu \) are defined by

$$\begin{aligned} {{\,\mathrm{veto}\,}}(\mathcal {C}_{\mu },j)&= 1 - \frac{1}{n-1}\sum _{t=1}^{n-1}\frac{1}{\left( {\begin{array}{c}n-1\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N{\setminus } \{j\} \\ |T|=t \end{array}} \mu (T),\\ {{\,\mathrm{favor}\,}}(\mathcal {C}_{\mu },j)&= \frac{1}{n-1}\sum _{t=0}^{n-1} \frac{1}{\left( {\begin{array}{c}n-1\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N{\setminus } \{j\} \\ |T|=t \end{array}} \mu (T\cup \{j\}) - \frac{1}{n-1}. \end{aligned}$$

In the following propositions we show the veto and favor indices of criteria with respect to the capacities \(\mu _{\varvec{p}}^{(r,s)}\).

Proposition 4

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\) and \(j\in N\). Then,

$$\begin{aligned} {{\,\mathrm{veto}\,}}\!\big (M_{\varvec{p}}^{(r,s)},j\big ) = 1 - \frac{r}{n-1} - \frac{(1-p_{j})\big ((n-r)(n-r-1)-s(s+1)\big )}{2(n-1)^2}. \end{aligned}$$

Proposition 5

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\) and \(j\in N\). Then,

$$\begin{aligned} {{\,\mathrm{favor}\,}}\!\big (M_{\varvec{p}}^{(r,s)},j\big )&= (1-p_{j}) \frac{(n-1)(n-2) + r(r+1) - s(s-1)}{2(n-1)^2}\\&\quad + p_{j}\left( 1 - \frac{s}{n-1}\right) . \end{aligned}$$

4.4 The Shapley Values

The Shapley values were introduced in the cooperative game theory as a solution to the problem of distributing the value \(\mu (N)\) among the players. They can be interpreted as a type of average of the contribution of player \(j\,\) alone in all coalitions (see Shapley 1953; Marichal 2007).

Definition 11

Let \(j\in N\) and let \(\mu \) be a normalized capacity on N. The Shapley value of criterion \(j\,\) with respect to \(\mu \) is defined by

$$\begin{aligned} \phi _{j}(\mu ) = \frac{1}{n}\sum _{t=0}^{n-1}\frac{1}{\left( {\begin{array}{c}n-1\\ t\end{array}}\right) } \sum _{\begin{array}{c} T\subseteq N{\setminus } \{j\} \\ |T|=t \end{array}}\Big (\mu (T\cup \{j\})-\mu (T)\Big ). \end{aligned}$$

The Shapley values of the capacities \(\mu _{\varvec{p}}^{(r,s)}\) are shown in the following proposition.

Proposition 6

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\) and \(j\in N\). Then,

$$\begin{aligned} \phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) = \frac{r+s}{n-1}\,\frac{1}{n} + \left( 1 - \frac{r+s}{n-1}\right) p_{j}. \end{aligned}$$
(3)

It is worth mentioning that the Shapley value \(\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big )\) is a convex combination between 1 / n and \(p_{j}\), which are the Shapley values of the capacities of any OWA operator (for instance, \(\mu _{\varvec{p}}^{(r,n-1-r)}\) or \(\mu _{\varvec{p}}^{(n-1-s,s)}\)) and \(\mu _{\varvec{p}}^{(0,0)}\), respectively. Moreover, given \(j\in N\), the values \(\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big )\) are the same for all capacities \(\mu _{\varvec{p}}^{(r,s)}\) having the same value of \(r+s\). Other immediate consequences of the previous proposition are gathered in the following corollary.

Corollary 2

Let \(\varvec{p}\) be a weighting vector and \(j\in N\).

  1. 1.

    If \(r+s = n-1\) then \(\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) = 1/n\).

  2. 2.

    If \(r+s< n-1\) then

    $$\begin{aligned} p_{j}> \frac{1}{n}&\;\Rightarrow \; \phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) > \frac{1}{n}, \\ p_{j} = \frac{1}{n}&\;\Rightarrow \; \phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) = \frac{1}{n}, \\ p_{j}< \frac{1}{n}&\;\Rightarrow \; \phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) < \frac{1}{n}. \end{aligned}$$

Since the Shapley value reflects the global importance of each criterion, it seems very interesting to be able to determine the weights that allow us to obtain Shapley values previously fixed. Notice that when \(r+s<n-1\) we can express the weight \(p_{j}\) in terms of \(\phi _{j}(\mu _{\varvec{p}}^{(r,s)})\):

$$\begin{aligned} p_{j}&= \frac{n-1}{n-1-(r+s)}\!\left( \!\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) - \frac{r+s}{n-1}\frac{1}{n} \right) \\&= \frac{n-1}{n-1-(r+s)}\!\left( \!\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) - \frac{1}{n} + \frac{1}{n} \!\left( 1 - \frac{r+s}{n-1} \right) \right) \\&= \frac{1}{n} + \frac{n-1}{n-1-(r+s)}\!\left( \!\phi _{j}\big (\mu _{\varvec{p}}^{(r,s)}\big ) - \frac{1}{n}\right) \!. \end{aligned}$$

From the previous expressions it is easy to check that \(p_{j} \ge 0\) if and only if \(\phi _{j}(\mu _{\varvec{p}}^{(r,s)}) \ge \frac{r+s}{n(n-1)}\), and \(\sum _{j=1}^{n} p_{j} = 1\). Therefore we have the following corollary.

Corollary 3

Let \((\phi _1, \ldots ,\phi _n)\) be a weighting vector. Given \((r,s) \in \mathcal {R}\) such that \(r+s<n-1\), the following conditions are equivalent:

  1. 1.

    \(\displaystyle \min _{j\in N} \phi _{j} \ge \frac{r+s}{n(n-1)}\).

  2. 2.

    The vector \(\varvec{p}\) defined by

    $$\begin{aligned} p_{j} = \frac{1}{n} + \frac{n-1}{n-1-(r+s)}\left( \phi _{j} - \frac{1}{n}\right) ,\quad j=1, \ldots ,n, \end{aligned}$$

    is a weighting vector such that \(\phi _{j}(\mu _{\varvec{p}}^{(r,s)}) = \phi _{j}\) for any \(j\in N\).

4.5 Interaction Indices

Interaction indices allow us to measure the interaction degree between two elements of N. This concept was initially proposed by Owen (1972) under the name of covalue, and, in an independent way, by Murofushi and Soneda (1993). Afterwards, it has been generalized by Grabisch (1997) to subsets of N with any number of elements.

Definition 12

Let \(\mu \) be a normalized capacity on N. The interaction index of elements \(j,k\in N\) is defined by

$$\begin{aligned} I_{jk}\big (\mu \big )&= \frac{1}{n-1} \sum _{t=0}^{n-2} \frac{1}{\left( {\begin{array}{c}n-2\\ t\end{array}}\right) }\\&\quad \cdot \sum _{\begin{array}{c} T\subseteq N{\setminus } \{j,k\} \\ |T|=t \end{array}} \Big (\mu (T\cup \{j,k\}) - \mu (T\cup \{j\}) - \mu (T\cup \{k\})+\mu (T)\Big ). \end{aligned}$$

We next give the interaction indices of the capacities \(\mu _{\varvec{p}}^{(r,s)}\).

Proposition 7

Let \(\varvec{p}\) be a weighting vector, \((r,s)\in \mathcal {R}\) and \(j,k\in N\). Then,

$$\begin{aligned} I_{jk}\big (\mu _{\varvec{p}}^{(r,s)}\big ) = {\left\{ \begin{array}{ll} \frac{1}{n-1} &{} \text {if } r = 0 \text { and } s=n-1,\\ \frac{1}{n-2}\left( p_{j}+p_{k} -\frac{1}{n-1}\right) &{} \text {if } r = 0 \text { and } 0< s< n-1\\ -\frac{1}{n-2}\left( p_{j}+p_{k} -\frac{1}{n-1}\right) &{} \text {if } s = 0 \text { and } 0< r < n-1\\ -\frac{1}{n-1} &{} \text {if } s = 0 \text { and } r=n-1,\\ 0 &{} \text {otherwise}. \end{array}\right. } \end{aligned}$$

4.6 The Möbius Transform

The Möbius transform (Rota 1964) is a relevant concept in several fields such as combinatorics and cooperative game theory. In this last field it is known as Harsanyi dividends (Harsanyi 1959) and it can be interpreted as the contribution of each coalition by itself, without considering its parts.

Definition 13

Let \(\mu \) be a normalized capacity on N. The Möbius transform of \(\mu \) is the set function \(m^{\mu }\) on N defined by

$$\begin{aligned} m^{\mu }(A) = \sum _{B\subseteq A}(-1)^{|A{\setminus } B|}\mu (B) \qquad (A\subseteq N). \end{aligned}$$

It is worth mentioning that given \(m^{\mu }\), it is possible to recover \(\mu \) by the expression

$$\begin{aligned} \mu (A)=\sum _{B\subseteq A} m^{\mu }(B) \qquad (A\subseteq N), \end{aligned}$$

which it is known as the Zeta transform. Moreover, the Shapley values are related to the Möbius transform through the following expression:

$$\begin{aligned} \phi _{j}(\mu ) = \sum _{j\in A \subseteq N} \frac{m^{\mu }(A)}{|A|}. \end{aligned}$$

In the following proposition we give the Möbius transforms of the capacities \(\mu _{\varvec{p}}^{(r,s)}\).

Proposition 8

Let \(\varvec{p}\) be a weighting vector and \((r,s)\in \mathcal {R}\). Then,

$$\begin{aligned} m^{\mu _{\varvec{p}}^{(r,s)}}(A) = {\left\{ \begin{array}{ll} 0 &{} \text {if } |A| \le s,\\ \displaystyle (-1)^{|A|-s-1} \left( {\begin{array}{c}|A|-2\\ s-1\end{array}}\right) \left( \sum _{i\in A} p_{i} \right) &{} \text {if } s< |A| < n-r,\\ \begin{aligned} &{}\!\left( \sum _{i\in A} p_{i} \right) \left( (-1)^{|A|-s-1} \left( {\begin{array}{c}|A|-2\\ s-1\end{array}}\right) \right. \\ &{}\!\left. - (-1)^{|A|-n+r} \left( {\begin{array}{c}|A|-2\\ n-r-2\end{array}}\right) \right) \\ &{}\!+ (-1)^{|A|-n+r} \left( {\begin{array}{c}|A|-1\\ n-r-1\end{array}}\right) \end{aligned}&\text {if } |A| \ge n-r. \end{array}\right. } \end{aligned}$$

5 Example

In this section we show the use of Winsorized weighted means from a practical point of view. For this, we will use an example taken from Llamazares (2019b). Suppose that the Department of Mathematics in a Faculty of Economics offers a research assistantship for the students accepted into the M. Sc. in Economics. Applicants are evaluated with respect to seven subjects: Mathematics I (MatI), Mathematics II (MatII), Mathematics III (MatIII), Statistics I (StaI), Statistics II (StaII), Econometrics I (EcoI), and Econometrics II (EcoII), and the members of the committee would like to take into account the following aspects:

  1. 1.

    Each one of the first three subjects is considered twice as important as each one of the remaining four.

  2. 2.

    Outliers should be discarded.Footnote 2

It is worth to notice that the above requirements fit perfectly into a Winsorized weighted mean-type aggregation where, in principle, the weighting vector \(\varvec{p}\) is (0.2, 0.2, 0.2, 0.1, 0.1, 0.1, 0.1). Table 2 collects the marks obtained by three students (marks are given on a scale from 0 to 10). Note that, in Statistics I, student A gets its highest mark whereas students B and C get their lowest grades. Furthermore, these marks are also very different from those obtained in the other subjects and, as we will see later, they could be considered outliers.

Table 2 Marks of three students in the different subjects

Table 3 gathers the ranking of the three students when the Winsorized weighted means are used. As one can observe, student A wins when the weighted mean, the minimum or the maximum are used. Moreover, he/she also wins with the pairs (0, 5), (0, 4), and (0, 3). In the case of student B, he/she is the winner with the pairs (4, 2), (4, 1), and (4, 0); and he/she ties with C in the first position with the order statistics \({{\,\mathrm{OS}\,}}_{2}\), \({{\,\mathrm{OS}\,}}_{3}\), \({{\,\mathrm{OS}\,}}_{4}\) and \({{\,\mathrm{OS}\,}}_{6}\). In the remaining cases, the winner is student C.

Table 3 Ranking of the students when the functions \(M_{\varvec{p}}^{(r,s)}\) are used

In Table 4 we collect the Shapley values of the Winsorized weighted means obtained by using expression (3). Notice that when \(r+s=0\) (that is, in the case of the weighted mean) we get the desired values. However, as the value of \(r+s\) increases, the Shapley values are approaching each other until they finally coincide, what happens when \(r+s=6\) (in this case, the Winsorized weighted means are order statistics and their Shapley values are 1 / 7).

Table 4 Shapley values of the Winsorized weighted means

Given that the Shapley values reflect the global importance of each criterion, the weighting vector \(\varvec{p}\) should be chosen so that the Shapley values are 0.2 (for the subjects MatI, MatII and MatIII) and 0.1 (for the remaining subjects). This can be done by using Corollary 3. Notice that, since \(\displaystyle \min _{j\in N} \phi _{j} =0.1\), the weights, which are shown in Table 5, can only be obtained when \(r+s\le 4\).

Table 5 Weights for which the Shapley values of the subjects are 0.2 and 0.1

Table 6 shows the ranking of the students obtained through \(M_{\varvec{p}}^{(r,s)}\) when the new weighting vectors are considered. It is interesting to note that now, in all cases, except for the weighted mean, the winner is student C. In fact, in all cases, except for the pairs (0, 0) and (0, 1), the ranking is C \(\succ \) B \(\succ \) A.

Table 6 Ranking of the students when the new weights are used

Notice that, from a practical point of view, it is necessary to choose a Winsorized weighted mean to rate the students. For that, we can use the following strategies:

  1. 1.

    To use a pair (rs) previously fixed (some typical percentages are 10%, 15%, 20% or 25%; although smaller percentages are also used; see Hoitash and Hoitash 2009). In this case the pair (rs) is chosen independently of the outliers present in the data. For instance, suppose we choose a percentage of 20% at both ends. Then, in our example, \(0.2\cdot 7 = 1.4\), and when we round down 1.4 to the nearest integer we have \(r=s=1\). As we can see in Table 6, the winner is student C.

  2. 2.

    To choose a pair (rs) so that the outliers of all students are removed. For that, first of all we have to choose a method to detect outliers.Footnote 3 Usual procedures to detect outliers in the case of one-dimensional data are the boxplot rule (Tukey 1977) and the MAD–median rule (see, for instance, Iglewicz and Hoaglin 1993, Wilcox 2012, and Leys et al 2013). In our example, with both methods we get the following outliers:

    1. (a)

      For the student A: 9.8.

    2. (b)

      For the student B: 5.2.

    3. (c)

      For the student C: 5.2. Therefore, to remove the outliers of the three students we should take the pair (1, 1),Footnote 4 and the winner is student C (see again Table 6).

6 Conclusion

In this paper we have generalized Winsorized means to include weighting vectors. For this, we have introduced, in a natural way, a new family of functions called Winsorized weighted means. These functions are a specific case of Choquet integrals (in fact, they are a specific case of SUOWA operators) and their capacities are relatively simple. This fact allows that several indices such as the orness degree, the k-conjunctiveness and k-disjunctiveness indices, the veto and favor indices, etc., can be given in closed-form expressions. Of particular interest are the closed-form expressions obtained for the Shapley values because they reflects the global importance of each information source. From them, we can determine the weights that allow us to get Shapley values previously fixed, and, in this way, we can aggregate data so that each information source has the desired weighting and outliers have no influence in the aggregation. It is important to note that other families of functions built in the framework of the Choquet integral do not exhibit this behavior (see Beliakov 2018).