1 Introduction

Economic globalization and fierce market competition are driving companies to seek overseas suppliers (Awasthi et al. 2018; Kim et al. 2018). Global sourcing benefits companies, but it also increases the risks of supplier selection as the range of alternatives expands. Supplier selection represents a strategic managerial issue that is closely related to corporate performance and sustainability (Viswanadham and Samvedi 2013). Such a complex decision is difficult for a single person or a small organization to reach. In practice, a special committee (including relevant department heads and experts) or a special board meeting will be established to conduct in-depth assessments. This configuration is a typical large-group decision-making scenario where multiple decision makers (DMs) attempt to select a common solution from a set of alternatives. Large-group decision making (LGDM) has received much attention in the field of decision analysis. Generally, a GDM problem can be called an LGDM problem when the number of DMs is more than 20 (Ding et al. 2019; Gou et al. 2018). And yet in the studies of Palomares et al. (2014) and Wu et al. (2018), the illustrative examples included 50 experts. Situations involving large-scale DMs are considered to present new challenges for the decision process, such as scalability, constant preference supervision, different behaviors towards consensus, time cost (Labella et al. 2018; Rodríguez et al. 2018; Xu et al. 2019).

The DMs themselves play a central role in solving decision problems, but the form in which the evaluation information is expressed is also important for the decision output. Linguistic assessment information can express natural language well, and it has been one of the most commonly used forms of information expression in modeling realistic decision-making scenarios (Herrera et al. 1996, 2009; Liu et al. 2016; Merigó and Gil-Lafuente 2013; Zuheros et al. 2018). The characteristics of several existing linguistic models are described in Table 1. The traditional linguistic assessment information generally allows DMs to express their preferences with single linguistic terms. However, sometimes, it is difficult to depict complex qualitative information only by one linguistic term. For instance, a DM may evaluate an item as “very good”, “good” or “somewhat good” but be unsure of how good the item is. To address such issues, Rodríguez et al. (2012) proposed the concept of hesitant fuzzy linguistic term sets (HFLTSs). HFLTSs collect all possible linguistic terms provided by the DMs, and all of these terms are assigned the same importance. In reality, a DM may have different preferences for multiple linguistic terms when judging an item. Therefore, Pang et al. (2016) extended the HFLTSs to a more general concept, named as probabilistic linguistic term sets (PLTSs). The PLTSs enable DMs to assign different weights to possible linguistic terms. To date, most studies on the GDM events with probabilistic linguistic information have focused on scenarios involving small-scale DMs, typically no more than 20. (e.g., Bai et al. 2018; Zhang et al. 2017).

Table 1 A review of the characteristics of several existing linguistic models

A probabilistic linguistic large-group decision making (PL-LGDM) problem can be defined as a situation in which a large number of DMs seek to choose a common solution from a set of alternatives in the context of probabilistic linguistic information. We characterize its features as follows: (1) the number of DMs is large (usually more than 20); (2) most or all DMs provide their evaluation information by means of PLTSs; and (3) obtaining a high-consensus output requires the implementation of a consensus reaching process (CRP). Generally, solving a PL-LGDM problem requires the following four processes: the probabilistic linguistic representation of the evaluation information, clustering process, consensus reaching process and selection process.

To address the scalability issue in LGDM, we assume that among a large number of DMs, there will be subgroups of them with similar opinions (Rodríguez et al. 2018). A clustering process can be used to divide the large group into several small-scale subgroups based on opinion similarity. Many clustering methods have been proposed, such as the K-means algorithm (Wu and Xu 2018), fuzzy c-means algorithm (Bezdek et al. 1984), and hierarchical clustering algorithm (Dong et al. 2006; Johnson 1967). This study adopts the hierarchical clustering method to classify large-scale DMs.

The CRP is considered particularly important in LGDM because opinions among a large number of DMs can easily be controversial. In contrast to the ideal consensus that requires unanimity, the notion of soft consensus allows differences of opinion within a reasonable range (Kacprzyk and Fedrizzi 1988). Various consensus models have been proposed to address GDM problems under different situations. However, research on consensus in the context of probabilistic linguistic information has just begun. Zhang et al. (2017) designed a CRP with probabilistic linguistic preference relations. Wu and Liao (2019) proposed a consensus-based probabilistic linguistic gained and lost dominance score method. We find that the case studies in the above literature focus on GDM problems composed of three experts and do not involve large-scale DMs.

Faced with the challenges of classical consensus models for LGDM and the scarcity of research on consensus in PL-LGDM, this study develops a hierarchical punishment-driven consensus model and applies it to global supplier selection. The following efforts are made to address the consensus in PL-LGDM, which can overcome the scalability challenge and soften the human supervision of the CRP.

  1. 1

    Problem framework construction: We define probabilistic linguistic large-group decision making and characterize its constituent elements. We present novel operational laws to preprocess probabilistic linguistic information.

  2. 2

    Clustering process: A hierarchical clustering method is utilized to divide the large group into several small-scale subgroups. The cluster is regarded as the basic unit in the decision-making process.

  3. 3

    Consensus reaching process: A hierarchical punishment-driven consensus model is employed to manage the differences among the opinions of clusters. According to the current group consensus index, the model can adopt different strategies to adjust the clusters’ opinions.

The remainder of this paper is organized as follows: Sect. 2 reviews the basic concepts of PLTSs and CRPs and formulates a GDM event with PLTSs. In Sect. 3, a PL-LGDM problem is established and the hierarchical clustering method is used to classify large-scale DMs. In Sect. 4, a hierarchical punishment-driven consensus model is developed. Section 5 applies the proposed model to a case study on global supplier selection. Section 6 presents the comparative analysis and managerial implications. This paper ends with conclusions in Sect. 7.

2 Preliminaries

In this section, we first review some concepts related to PLTSs and present the construction of a GDM problem with probabilistic linguistic information. The CRP for LGDM problems is then described.

2.1 Group Decision Making with Probabilistic Linguistic Information

As an extension of HFLTSs, the concept of PLTSs was originally proposed by Pang et al. (2016), which enables DMs to express possible linguistic terms with different weights. Clearly, Pang et al. (2016) defined PLTSs on the basis of an additive linguistic evaluation scale, while Zhang et al. (2016) used PLTSs based on a subscript-symmetric linguistic evaluation scale.

Definition 1

(Pang et al. 2016) Let \(S = \left\{ {\left. {s_{\alpha } } \right|\alpha = - \tau , \ldots - 1,0,1, \ldots ,\tau } \right\}\) (τ is a positive integer) be a linguistic term set; then a PLTS is defined as

$$L\left( p \right) = \left\{ {\left. {L^{\left( u \right)} \left( {p^{\left( u \right)} } \right)} \right|L^{\left( u \right)} \in S, \, p^{\left( u \right)} \ge 0, \, u = 1,2, \ldots ,\# L\left( p \right),\sum\limits_{u = 1}^{\# L\left( p \right)} {p^{\left( u \right)} } \le 1} \right\},$$
(1)

where \(L^{\left( u \right)} \left( {p^{\left( u \right)} } \right)\) is a probabilistic linguistic element that includes the linguistic term \(L^{\left( u \right)}\) and the associated probability \(p^{\left( u \right)}\), and \(\# L\left( p \right)\) is the number of different linguistic terms in \(L\left( p \right)\).

A probabilistic linguistic element consists of two parts: a linguistic term and its corresponding probability. Based on this unique construct, a PLTS can be obtained in two ways: the statistical aggregation of single linguistic terms (see Pang et al. 2016) and the aggregation of multiple PLTSs (see Zhang et al. 2017). Formally, a multicriteria GDM problem with probabilistic linguistic information includes (1) a set of alternatives \(X = \left\{ {x_{1} ,x_{2} , \ldots ,x_{m} } \right\} \, \left( {m \ge 2} \right)\), which are the possible solutions to the problem; (2) a set of criteria \(A = \left\{ {a_{1} ,a_{2} , \ldots ,a_{n} } \right\} \, \left( {n \ge 2} \right)\) used to evaluate the alternatives in X; and (3) a set of DMs \(E = \left\{ {e_{1} ,e_{2} , \ldots ,e_{q} } \right\} \, \left( {q \ge 2} \right)\) who provide judgments on the alternatives. Let \(V_{l} = \left( {v_{l,ij} } \right)_{m \times n}\) be the individual opinion given by DM \(e_{l} \left( {l = 1,2, \ldots ,q} \right)\), where \(v_{l,ij}\) represents the evaluation of DM \(e_{l}\) on alternative \(x_{i} \in X\) with respect to criterion \(a_{j} \in A\). \(v_{l,ij}\) can be either a single linguistic term or a PLTS. The former expression represents a relatively precise evaluation value, while the latter reflects the DM’s hesitation. To deal with the above different types of linguistic information, we attach a probability of 1 to the single linguistic term to convert it to a PLTS. For example, a given linguistic term \(s_{3}\) can be changed to \(\left\{ {s_{3} \left( 1 \right)} \right\}\). In particular, if all individual opinions are expressed by means of single linguistic terms, a collective opinion can be obtained as a PLTS by aggregating linguistic terms (Song and Li 2019). In this study, we focus on a GDM in which most or all DMs provide evaluation information in the form of PLTSs. This is consistent with the fact that, due to the complexity of LGDM problems, it is difficult for DMs to express opinions by only using single linguistic terms.

2.2 Consensus Reaching Process in LGDM Problems

The increasing complexity of current decision-making scenarios requires the participation of large-scale DMs. Two main differences between classical GDM and LGDM are that (1) the latter case usually involves a larger number of DMs, and (2) implementing a CRP to achieve a high-consensus outcome is considered both more necessary and more difficult, as large-scale DMs tend to be controversial (Labella et al. 2018; Xu et al. 2019).

Consensus can be defined as “a state of mutual agreement among members of a group, where all legitimate concerns of individuals have been addressed to the satisfaction of the group’’ (Saint and Lawson 1994). Obtaining a complete consensus is often unnecessary in practice and, as a consequence, the notion of soft consensus is introduced, which requires most (but perhaps not all) of the DMs to agree on the most important alternatives (Kacprzyk and Fedrizzi 1988; Palomares et al. 2014). Reaching a consensus is an iterative group discussion process in which some DMs must modify their opinions to bring them closer to the group opinion (Wu and Xu 2018). Two types of consensus measures have commonly been used: the first is based on the distance to the group opinion, and the second is based on the distances between individual opinions (Du et al. 2020; Labella et al. 2018; Wu and Xu 2016). A general scheme for a CRP is described as follows:

  • Problem framework configuration: A GDM problem is established, including a finite set of alternatives, a set of DMs, and some pre-set important parameters (such as the consensus threshold).

  • Opinion gathering: The opinions provided by DMs are gathered.

  • Consensus measure: The group consensus index is calculated based on distance measures, and it reflects the level of agreement in the group.

  • Consensus control: If the obtained consensus index is greater than the consensus threshold, this indicates that the desired consensus has been achieved and the selection process is concluded; otherwise, more consensus iterations should be carried out.

  • Consensus progress: A procedure is adopted to identify the opinions that contribute less to the consensus and to guide the opinion adjustment process to increase the consensus index in the subsequent iterations. Then, another iteration starts by gathering opinions again.

To overcome the scalability challenge in LGDM problems, we consider that there will be DMs with similar opinions. In this study, a clustering algorithm is implemented to divide a large group and to weight subgroups before calculating the consensus measures. A hierarchical punishment-driven consensus model is then employed, by which the opinions of clusters are adjusted in the direction of the increasing consensus index.

3 Hierarchical Clustering Method for PL-LGDM Problems

In Sect. 3.1, we describe the establishment of a PL-LGDM problem and the preprocessing of decision information. To tackle the scalability challenge in PL-LGDM, Sect. 3.2 adopts a hierarchical clustering algorithm to classify the large group.

3.1 Problem Framework Configuration and Preprocessing of Decision Information

We first provide a description of the PL-LGDM scenario studied in this study. Let \(X = \left\{ {x_{1} ,x_{2} , \ldots ,x_{m} } \right\}\) be a finite set of alternatives, \(A = \left\{ {a_{1} ,a_{2} , \ldots ,a_{n} } \right\}\) be a set of criteria, \(E = \left\{ {e_{1} ,e_{2} , \ldots ,e_{q} } \right\}\left( {q \ge 20} \right)\) be a set of DMs, and S be a linguistic term set as before. This study considers a group to be a large group when the number of DMs exceeds 20. Let \(V_{l} = \left( {v_{l,ij} } \right)_{m \times n}\) be the individual decision matrix provided by DM \(e_{l} \left( {l = 1,2, \ldots ,q} \right)\), where \(v_{l,ij} = \left\{ {\left. {L_{l,ij}^{\left( u \right)} \left( {p_{l,ij}^{\left( u \right)} } \right)} \right|L_{l,ij}^{\left( u \right)} \in S,p_{l,ij}^{\left( u \right)} \ge 0, \, u = 1,2, \ldots ,\# v_{l.ij} ,\sum\nolimits_{u = 1}^{{\# v_{l,ij} }} {p_{l,ij}^{\left( u \right)} } \le 1} \right\}\). \(\# v_{l,ij}\) is the number of different linguistic terms in \(v_{l,ij}\). Generally, there are benefit types and cost types in the criteria. By using the linguistic negation operator (Xu 2005), we convert the cost types to benefit types (for convenience, the converted result is still represented by \(v_{l,ij}\)) so that \(L_{l,ij}^{\left( u \right)} = neg\left( {L_{l,ij}^{\left( u \right)} } \right)\) when \(a_{j}\) is a cost criterion.

To obtain the aggregation and distance measure of a PLTS, normalization is necessary. Generally, discrete linguistic terms are used to evaluate alternatives, while virtual linguistic terms only appear in operations and rankings (Xu 2009). Intuitively, in CRPs, discrete linguistic terms enable DMs to better understand the differences between individual opinions, which is conducive to making more reasonable adjustments. Therefore, we believe that before the selection process, the normalization and aggregation of PLTSs should be expressed in terms of discrete linguistic terms associated with probability values. Here, we redefine the normalization and operational laws of PLTSs. Normalization includes two steps:

  1. 1

    Granularity normalization: All PLTSs contain the same linguistic terms;

  2. 2

    Probability normalization: The sum of the probability values of all linguistic terms in each PLTS is 1.

Definition 2

(Granularity normalization) Let \(L\left( p \right)_{1}\) and \(L\left( p \right)_{2}\) be any two PLTSs, where \(L\left( p \right)_{1} = \left\{ {\left. {L_{1}^{\left( u \right)} \left( {p_{1}^{\left( u \right)} } \right)} \right|u = 1,2, \ldots ,\# L\left( p \right)_{1} } \right\}\) and \(L\left( p \right)_{2} = \left\{ {\left. {L_{2}^{\left( u \right)} \left( {p_{2}^{\left( u \right)} } \right)} \right|u = 1,2, \ldots ,\# L\left( p \right)_{2} } \right\}\). If there is a linguistic term \(L_{1}^{\left( u \right)}\) in \(L\left( p \right)_{1}\) that does not appear in \(L\left( p \right)_{2}\), then add \(L_{1}^{\left( u \right)}\) to \(L\left( p \right)_{2}\). By this way, a new probabilistic linguistic element is obtained as \(L_{2}^{\left( u \right)} \left( {p_{2}^{\left( u \right)} } \right)\), where \(L_{2}^{\left( u \right)} = L_{1}^{\left( u \right)}\) and \(p_{2}^{\left( u \right)} = 0\). Repeat the above steps until the two PLTSs have the same linguistic terms.

Definition 3

(Probability normalization) Given any PLTS \(L\left( p \right)\) with \(\sum\nolimits_{u = 1}^{\# L\left( p \right)} {p^{\left( u \right)} } < 1\), the associated PLTS \(\dot{L}\left( p \right)\) is obtained as

$$\dot{L}\left( p \right) = \left\{ {\left. {L^{\left( u \right)} \left( {\dot{p}^{\left( u \right)} } \right)} \right|L^{\left( u \right)} \in S, \, u = 1,2, \ldots ,\# L\left( p \right)} \right\},$$
(2)

where \(\dot{p}^{\left( u \right)} = {{p^{\left( u \right)} } \mathord{\left/ {\vphantom {{p^{\left( u \right)} } {\sum\nolimits_{u = 1}^{\# L\left( p \right)} {p^{\left( u \right)} } }}} \right. \kern-0pt} {\sum\nolimits_{u = 1}^{\# L\left( p \right)} {p^{\left( u \right)} } }}\), \(u = 1,2, \ldots ,\# L\left( p \right)\). \(\dot{L}\left( p \right)\) is called the normalization form of \(L\left( p \right)\).

For simplicity, the normalized PLTS (NPLTS) is still written as \(L\left( p \right)\). The granularity of the NPLTS \(L\left( p \right)\) is denoted as \(\# \# L\left( p \right)\), where \(\# \# L\left( p \right) \ge \# L\left( p \right)\).

Example 1

Given two PLTSs \(L\left( p \right)_{1} = \left\{ {s_{ - 1} \left( {0.1} \right),s_{1} \left( {0.4} \right)} \right\}\) and \(L\left( p \right)_{2} = \left\{ {s_{0} \left( {0.2} \right),s_{1} \left( {0.3} \right),s_{2} \left( {0.5} \right)} \right\}\), where the linguistic terms of both are drawn from \(S_{1} = \left\{ {s_{ - 2} ,s_{ - 1} ,s_{0} ,s_{1} ,s_{2} } \right\}\), normalization is divided into two steps:

Step 1:

Granularity normalization: \(L\left( p \right)_{1} = \left\{ {s_{ - 1} \left( {0.1} \right),s_{0} \left( 0 \right),s_{1} \left( {0.4} \right),s_{2} \left( 0 \right)} \right\}\) and \(L\left( p \right)_{2} = \left\{ {s_{ - 1} \left( 0 \right),s_{0} \left( {0.2} \right),s_{1} \left( {0.3} \right),s_{ 2} \left( {0.5} \right)} \right\}\). Thus, we have \(\# \# L\left( p \right)_{1} = \# \# L\left( p \right)_{2}\)

Step 2:

Probability normalization: \(L\left( p \right)_{1} = \left\{ {s_{ - 1} \left( {0.2} \right),s_{0} \left( 0 \right),s_{1} \left( {0.8} \right),s_{2} \left( 0 \right)} \right\}\) and \(L\left( p \right)_{2} = \left\{ {s_{ - 1} \left( 0 \right),s_{0} \left( {0.2} \right),s_{1} \left( {0.3} \right),s_{ 2} \left( {0.5} \right)} \right\}\)

Next, we describe some basic operations.

Definition 4

Given any two NPLTSs, \(L\left( p \right)_{1} = \left\{ {\left. {L_{1}^{\left( u \right)} \left( {p_{1}^{\left( u \right)} } \right)} \right|L_{1}^{\left( u \right)} \in S, \, u = 1,2, \ldots ,\# \# L\left( p \right)_{1} } \right\}\) and \(L\left( p \right)_{2} = \left\{ {\left. {L_{2}^{\left( u \right)} \left( {p_{2}^{\left( u \right)} } \right)} \right|L_{2}^{\left( u \right)} \in S, \, u = 1,2, \ldots ,\# \# L\left( p \right)_{2} } \right\}\), \(0 \le \lambda_{1} ,\lambda_{2} \le 1\). Then, we have

  1. 1

    \(L\left( p \right)_{1} \oplus L\left( p \right)_{2} = \left\{ {\left. {L_{3}^{\left( u \right)} \left( {p_{3}^{\left( u \right)} } \right)} \right|L_{3}^{\left( u \right)} \in S, \, u = 1,2, \ldots ,\# \# L\left( p \right)_{1} } \right\}\),

  2. 2

    \(\lambda L\left( p \right)_{1} = \left\{ {\left. {L_{1}^{\left( u \right)} \left( {\lambda p_{1}^{\left( u \right)} } \right)} \right|u = 1,2, \ldots ,\# \# L\left( p \right)_{1} } \right\}\),

  3. 3

    \(\lambda_{1} L\left( p \right)_{1} \oplus \lambda_{2} L\left( p \right)_{2} = \left\{ {\left. {L_{4}^{\left( u \right)} \left( {p_{4}^{\left( u \right)} } \right)} \right|L_{4}^{\left( u \right)} \in S, \, u = 1,2, \ldots ,\# \# L\left( p \right)_{1} } \right\}\),

where \(L_{3}^{\left( u \right)} = L_{1}^{\left( u \right)}\), \(p_{3}^{\left( u \right)} = p_{1}^{\left( u \right)} + p_{2}^{\left( u \right)} - p_{1}^{\left( u \right)} p_{2}^{\left( u \right)}\), \(L_{4}^{\left( u \right)} = L_{1}^{\left( u \right)}\), \(p_{4}^{\left( u \right)} = \lambda_{1} p_{1}^{\left( u \right)} + \lambda_{2} p_{2}^{\left( u \right)} - \lambda_{1} \lambda_{2} p_{1}^{\left( u \right)} p_{2}^{\left( u \right)}\).

Remark 1

Notice that sometimes the sum of the probability values of the PLTS obtained by Definition 4 is less than 1. In this case, we need to redo the process of probability normalization.

Example 2

(Continued with Example 1) Let \(\lambda_{1} = 0.4\) and \(\lambda_{2} = 0.6\); then, we have the following results:

Pang et al. (2016)’s operational law yields:

$$\begin{aligned} \lambda_{1} L\left( p \right)_{1} \oplus \lambda_{2} L\left( p \right)_{2} & = 0.4\left\{ {s_{1} \left( {0.8} \right),s_{ - 1} \left( 0 \right),s_{ - 1} \left( {0.2} \right)} \right\} \oplus 0.6\left\{ {s_{2} \left( {0.5} \right),s_{1} \left( {0.3} \right),s_{0} \left( {0.2} \right)} \right\} \\ & = \left\{ {0.4 \times 0.8s_{1} \oplus 0.6 \times 0.5s_{2} ,0.4 \times 0s_{ - 1} \oplus 0.6 \times 0.3s_{1} ,0.4 \times 0.2s_{ - 1} \oplus 0.6 \times 0.2s_{0} } \right\} \\ & = \left\{ {0.32s_{1} \oplus 0.3s_{2} ,0s_{1} \oplus 0.18s_{1} ,0.08s_{ - 1} \oplus 0.12s_{0} } \right\} = \left\{ {s_{0.92} ,s_{0.18} ,s_{ - 0.08} } \right\}, \\ \end{aligned}$$

and our operational law yields:

$$\begin{aligned} \lambda_{1} L\left( p \right)_{1} \oplus \lambda_{2} L\left( p \right)_{2} & = 0.4\left\{ {s_{ - 1} \left( {0.2} \right),s_{0} \left( 0 \right),s_{1} \left( {0.8} \right),s_{2} \left( 0 \right)} \right\} \oplus 0.6\left\{ {s_{ - 1} \left( 0 \right),s_{0} \left( {0.2} \right),s_{1} \left( {0.3} \right),s_{2} \left( {0.5} \right)} \right\} \\ & = \left\{ {s_{ - 1} \left( {0.0849} \right),s_{0} \left( {0.1273} \right),s_{1} \left( {0.4694} \right),s_{2} \left( {0.3184} \right)} \right\}. \\ \end{aligned}$$

From Example 2, we find that the probabilities of linguistic terms are missing if Pang et al.’s (2016) operational law is used. In this case, the unique feature of PLTSs compared with ordinary linguistic term sets disappears. More importantly, the virtual linguistic terms make it difficult for DMs to understand and determine the adjustments in the CRP. In our operational law, the ordinary linguistic terms remain unchanged, yet the associated probabilities always change.

To better apply PLTSs to GDM problems, the following aggregation operators are defined:

Definition 5

Let \(L\left( p \right)_{i} = \left\{ {\left. {L_{i}^{\left( u \right)} \left( {p_{i}^{\left( u \right)} } \right)} \right|L_{i}^{\left( u \right)} \in S, \, k_{i} = 1,2, \ldots ,\# \# L\left( p \right)_{i} } \right\}\; \, \left( {i = 1,2, \ldots ,m} \right)\) be a set of m NPLTSs and \(\theta_{i}\) be the weight of \(L\left( p \right)_{i}\), such that \(\theta_{i} \ge 0\), \(i = 1,2, \ldots ,m\), and \(\sum\nolimits_{i = 1}^{m} {\theta_{i} } = 1\); then, the probabilistic linguistic weighted averaging (PLWA) operator is defined as

$$L\left( p \right) = PLWA\left( {L\left( p \right)_{1} ,L\left( p \right)_{2} , \ldots ,L\left( p \right)_{m} } \right) = \theta_{1} L\left( p \right)_{1} \oplus \theta_{2} L\left( p \right)_{2} \oplus \cdots \oplus \theta_{m} L\left( p \right)_{m} ,$$
(3)

where \(L\left( p \right) = \left\{ {\left. {L^{\left( u \right)} \left( {p^{\left( u \right)} } \right)} \right|L^{\left( u \right)} = L_{1}^{\left( u \right)} , \, u = 1,2, \ldots ,\# \# L\left( p \right)} \right\}\), such that

$$\begin{aligned} p^{\left( u \right)} & = \sum\limits_{i = 1}^{m} {\theta_{i} p_{i}^{\left( u \right)} } - \sum\limits_{1 \le i \le j \le m} {\theta_{i} p_{i}^{\left( u \right)} \theta_{j} p_{j}^{\left( u \right)} } \\ & \quad + \,\sum\limits_{1 \le i \le j \le l \le m} {\theta_{i} p_{i}^{\left( u \right)} \theta_{j} p_{j}^{\left( u \right)} \theta_{l} p_{l}^{\left( u \right)} } + \cdots + \left( { - 1} \right)^{n - 1} \prod\limits_{i = 1}^{m} {\theta_{i} p_{i}^{\left( u \right)} } . \\ \end{aligned}$$
(4)

Definition 6

Let \(L\left( p \right)_{1}\) and \(L\left( p \right)_{2}\) be any two NPLTSs; then, the distance between \(L\left( p \right)_{1}\) and \(L\left( p \right)_{2}\) is calculated by

$$d\left( {L\left( p \right)_{1} ,L\left( p \right)_{2} } \right) = \frac{{\left( {\sum\nolimits_{u = 1}^{{\# \# L\left( p \right)_{1} }} {\left( {\left| {I_{1}^{\left( u \right)} } \right|\left| {p_{1}^{\left( u \right)} - p_{2}^{\left( u \right)} } \right|} \right)^{\lambda } } } \right)^{{{1 \mathord{\left/ {\vphantom {1 \lambda }} \right. \kern-0pt} \lambda }}} }}{{\# \# L\left( p \right)_{1} }},$$
(5)

where \(\lambda > 0\) and \(I_{1}^{\left( u \right)}\) is the subscript of the linguistic term \(L_{1}^{\left( u \right)}\). In this study, we adopt the Euclidean distance, i.e., \(\lambda = 2\).

It is clear that the distance measures have the following properties:

  1. 1

    \(0 \le d\left( {L\left( p \right)_{1} ,L\left( p \right)_{2} } \right) \le 1\),

  2. 2

    \(d\left( {L\left( p \right)_{1} ,L\left( p \right)_{2} } \right) = 0\), if \(L\left( p \right)_{1} = L\left( p \right)_{2}\),

  3. 3

    \(d\left( {L\left( p \right)_{1} ,L\left( p \right)_{2} } \right) = d\left( {L\left( p \right)_{2} ,L\left( p \right)_{1} } \right)\).

3.2 Hierarchical Clustering Algorithm

Clustering is a widely used methodology for analyzing and processing large-scales DMs, and it is considered to be effective in addressing the scalability challenge in LGDM problems. In this study, we adopt a hierarchical clustering method to classify the large-scale DMs into \(K\left( {1 \le K \le q} \right)\) subgroups. Hierarchical clustering treats each data object as a separate cluster and then finds clusters with small distances to merge in each iteration. This process repeats until a designated number of clusters is reached or there is only one cluster. The procedure for the hierarchical clustering method is described in Algorithm 1.

figure a

Theorem 1

The time complexity of Algorithm 1 is \(O\left( {q^{3} } \right)\).

Proof

First, it will take \(O\left( {q^{2} } \right)\) time to calculate the distances between each pair of clusters. Once a new cluster is determined, the distance between the new cluster and each of the other clusters needs to be calculated. It can be seen that Algorithm 1 must be carried out for q iterations and that each iteration updates the distances, so the time complexity is the cube of the number of original clusters, i.e., \(O\left( {q^{3} } \right)\).□

Calculating the weights of the clusters is a necessary step because it is closely related to the generation of group opinions. Inspired by Rodríguez et al. (2018), this study takes into account the size and cohesion. The size refers to the number of DMs in a cluster. Based on the majority principle, the larger the size of a cluster is, the more weight it should be given. However, the representation of the size should be adjusted according to the number of DMs involved in the LGDM problem. Based on computing with words (Quesada et al. 2015), Rodríguez et al. (2018) modeled the size with the fuzzy membership function \(\mu_{\text{size}}\) shown in Fig. 1, where the universe of discourse is the number of DMs in a cluster and the membership degree reflects the cluster’s influence on all the DMs in terms of the number of DMs. The points a and b of the membership function depend on the numbers of alternatives and DMs in the LGDM scenario. The highest membership degree is for values above b, the lowest membership degree is for values below a, and varying importance is assigned between a and b.

Fig. 1
figure 1

Membership function for the cluster size

The cohesion represents the consensus index of the DMs’ opinions in a cluster. The more coherent the individual opinions within a cluster are, the more weight we give the cluster.

Definition 7

For cluster \(c_{k}\) including \(n_{k}\) DMs, the cohesion of the cluster is defined as

$$cohesion\left( {c_{k} } \right) = 1 - \frac{1}{{{{n_{k} \left( {n_{k} - 1} \right)} \mathord{\left/ {\vphantom {{n_{k} \left( {n_{k} - 1} \right)} 2}} \right. \kern-0pt} 2}}}\sum\limits_{h = 1}^{{n_{k} - 1}} {\sum\limits_{l = h + 1}^{{n_{k} }} {d\left( {R_{l} ,R_{h} } \right)} } ,$$
(7)

where \(e_{l} , \, e_{h} \in c_{k}\). Clearly, \(0 \le cohesion\left( {c_{k} } \right) \le 1\).

By integrating size and cohesion, Rodríguez et al. (2018) defined a function to calculate the weights of clusters, which could be flexibly adapted to a specific LGDM problem.

Definition 8

(Rodríguez et al. 2018) Let \(Y_{{c_{k} }} = \left\{ {y_{1} ,y_{2} } \right\}\) be the values obtained for cohesion and size, respectively, where \(y_{1} ,y_{2} \in \left[ {0,1} \right]\); then, the fusion value of the cohesion and size is calculated by

$$\varphi \left( {Y_{{c_{k} }} } \right) = \left( {1 + y_{2} } \right)^{{y_{1\beta } }} ,$$
(8)

where the parameter \(\beta \left( {\beta > 0} \right)\) is used to adjust the effect of cohesion in calculating the weight of the cluster.

The aggregated value \(\varphi \left( {Y_{{c_{k} }} } \right)\) reflects the relevance of cluster \(c_{k}\). Therefore, the weight vector of the clusters, \(\theta = \left( {\theta_{1} ,\theta_{2} , \ldots ,\theta_{K} } \right)^{T}\), can be obtained from

$$\theta_{k} = \frac{{\varphi \left( {Y_{k} } \right)}}{{\sum\nolimits_{k = 1}^{K} {\varphi \left( {Y_{k} } \right)} }}.$$
(9)

Clearly, \(0 \le \theta_{k} \le 1\) \(\left( {k = 1,2, \ldots ,K} \right)\) and \(\sum\nolimits_{k = 1}^{K} {\theta_{k} } = 1\). By using Eq. (3), the cluster opinion is obtained; i.e., \(G_{k} = \left( {g_{k,ij} } \right)_{m \times n}\), where \(g_{k,ij} = PLWA_{\theta } \left( {r_{1,ij} , \ldots ,r_{{n_{k} ,ij}} } \right)\).

We use the case in Sect. 5 to illustrate the weight calculation. In this study, we consider 10% of DMs to define the point a and the number of DMs divided by the number of alternatives to define the point b: \(a = round\left( {q \cdot 0.1} \right) = 2\), \(b = round\left( {{q \mathord{\left/ {\vphantom {q m}} \right. \kern-0pt} m}} \right) = 4\), where \(round\left( \cdot \right)\) is the round function. Table 2 shows the values of size, membership degree, and cohesion for each cluster. Figure 2 shows the weights of clusters when different values of the parameter β are used to solve Eq. (8). We find that although clusters \(c_{ 2}\) and \(c_{3}\) have the same membership degree in terms of size, they are assigned different weights due to the different values of cohesion among their members. As shown in Fig. 2, when the value of β increases, the weight of cluster \(c_{2}\) increases more than that of cluster \(c_{3}\). This means that the parameter β can adjust the degree of the influence of cohesion on the weight.

Table 2 The values of size, membership degree, and cohesion for each cluster
Fig. 2
figure 2

Weights of the clusters as \(\beta\) varies. Note: set \(a = 2\), \(b = 4\)

4 Hierarchical Punishment-Driven Consensus Model in PL-LGDM Problems

We first describe the consensus measure. Section 4.2 presents punishment-driven consensus iterations. In Sect. 4.3, a hierarchical punishment-driven consensus model is given.

4.1 Consensus Measure

The consensus measure is designed to compute the differences among DMs’ opinions. As previously discussed, the calculation of the consensus measure can be based on the distance to the group opinion or the distances among individual opinions. Xu et al. (2019) gave a detailed comparison of these two measures. The former measure is heavily influenced by the weights of the clusters. In this case, the CRP tends to target the opinions of low-weight clusters because their opinions contribute less to the group opinion. So that our model uses the latter measure, which is completely based on the differences among cluster’s opinions, without considering the influence of the clusters’ weights. We present the three-level consensus index measure below.

Level 1

Individual consensus index at the matrix element level. The consensus index of a cluster with respect to the others on alternative \(x_{i}\) under criterion \(a_{j}\) is

$$ICI_{k,ij}^{t} = 1 - \frac{1}{K - 1}\sum\limits_{h = 1,h \ne k}^{K} {d\left( {g_{k,ij}^{t} ,g_{h,ij}^{t} } \right)} ,$$
(10)

where \(d\left( {g_{k,ij}^{t} ,g_{h,ij}^{t} } \right)\) is the distance between \(g_{k,ij}^{t}\) and \(g_{h,ij}^{t}\). Clearly, \(0 \le ICI_{k,ij}^{t} \le 1\). The greater the value of \(ICI_{k,ij}^{t}\) is, the higher the consensus index of the matrix element \(g_{k,ij}^{t}\).

Level 2

Individual consensus index at the matrix level. The consensus index of a cluster opinion with respect to the others is

$$ICI_{k}^{t} = \frac{1}{m \times n}\sum\limits_{i = 1}^{m} {\sum\limits_{j = 1}^{n} {ICI_{k,ij}^{t} } } .$$
(11)

Clearly, \(0 \le ICI_{k}^{t} \le 1\). The greater the value of \(ICI_{k}^{t}\) is, the higher the consensus index of cluster \(c_{k}\).

Level 3

Consensus index at the group level. The group consensus index is computed by

$$GCI^{t} = \frac{1}{K}\sum\limits_{k = 1}^{K} {ICI_{k}^{t} } .$$
(12)

Clearly, \(0 \le GCI^{t} \le 1\). The higher the value of \(GCI^{t}\) is, the higher the consensus among the clusters.

Usually, we need to define the consensus threshold (denoted as \(\overline{GCI}\), where \(0 \le \overline{GCI} \le 1\)), which is used to determine whether the CRP can terminate. Calculating the threshold is an important matter, and many related studies have been published (see Labella et al. 2018; Xu et al. 2015; Zhang et al. 2017).

Remark 2

The three-level consensus index measure provides inspiration for constructing a hierarchical consensus model. We can either adjust all elements of the opinion (based on the individual consensus indexes at the matrix level) in each iteration or adjust only some elements of the opinion (based on the individual consensus index at the matrix element level).

4.2 Punishment-Driven Consensus Iterations

There are usually two ways to adjust opinions: the cluster opinion that has the largest distance from the group opinion is adjusted in each iteration (e.g. Wu and Liao 2019); or the cluster opinions that contribute less to the consensus are adjusted (e.g. Rodríguez et al. 2018). The former way sorts the individual consensus indexes and adjusts the opinions of clusters accordingly until the current group consensus index meets the consensus threshold. Compared with the latter way, the former has the advantage in terms of reducing the number of adjusted opinions. This paper selects the first method. Since this study takes the cluster as the basic decision-making unit, the opinions of DMs have been merged into the opinions of clusters before the CRP begins. This means that the next step for DMs is to focus on discussing the adjustment of the clusters’ opinions. Suppose cluster \(c_{{\dot{k}}}\) has the lowest consensus index in the t-th iteration, i.e., \(ICI_{{\dot{k}}}^{t} = \hbox{min} \left\{ {ICI_{k}^{t} } \right\}\). Cluster \(c_{{\dot{k}}}\) provides the adjustment coefficient \(\delta_{{\dot{k}}}^{t} \, \left( {0 \le \delta_{{\dot{k}}}^{t} \le 1} \right)\) to make its opinion closer to the others. To provide guidance to the cluster regarding the adjustment coefficient and soften the human supervision of the CRP, we propose the concept of the punishment coefficient, which is the degree to which a cluster opinion is required to move towards the collective opinion.

Definition 9

Given the consensus threshold \(\overline{GCI}\) and identified cluster consensus index \(ICI_{{\dot{k}}}^{t}\) in the t-th iteration, the punishment coefficient is defined as

$$pc_{{\dot{k}}}^{t} = \left( {\frac{{\overline{GCI} - ICI_{{\dot{k}}}^{t} }}{{\overline{GCI} }}} \right)^{\sigma } ,$$
(13)

where \(\sigma \left( {0 \le \sigma \le 1} \right)\) is the power of the punishment coefficient and is used to indicate the urgency of reaching a consensus. Clearly, \(0 \le pc_{{\dot{k}}}^{t} \le 1\).

Figure 3 shows the distribution of the punishment coefficient for different values of σ. We set \(\sigma = 1\) as the benchmark. We find that the larger the value of σ is, the smaller the punishment coefficient \(pc_{{\dot{k}}}^{t}\). σ = 0 is an extreme case, indicating that the identified cluster opinion is completely replaced by the collective opinion. Another important function of the punishment coefficient is that it can be used as the lower limit of the adjustment coefficient. If the given adjustment coefficient is less than the punishment coefficient, the cluster is required to use the punishment coefficient to adjust its opinion; i.e., if \(\delta_{{\dot{k}}}^{t} < pc_{{\dot{k}}}^{t}\), then set \(\delta_{{\dot{k}}}^{t} = pc_{{\dot{k}}}^{t}\). This is why our model is called a punishment-driven consensus model. In this way, the model not only provides guidance for opinion adjustment, but also designs a semi-automatic mandatory adjustment mechanism, which can soften the human supervision of the CRP.

Fig. 3
figure 3

Distribution of the punishment coefficient for different values of σ

Based on the operational laws proposed in this paper, the clusters’ opinions are always expressed in the form of PLTSs before the selection process begins, which include discrete linguistic terms and their corresponding probability values. We define the adjusted opinion of the CRP as follow.

Definition 10

Let \(G_{{\dot{k}}}^{t}\) be the identified opinion with the lowest consensus index in the t-th iteration, \(\delta_{{\dot{k}}}^{t}\) be the given adjustment coefficient, and \(pc_{{\dot{k}}}^{t}\) be the punishment coefficient. If \(GCI^{t} < \overline{GCI}\), then the opinion \(G_{{\dot{k}}}^{t}\) should be adjusted as \(G_{{\dot{k}}}^{t + 1} = \left( {g_{{\dot{k},ij}}^{t + 1} } \right)_{m \times n}\), such that

$$g_{{\dot{k},ij}}^{t + 1} = \left\{ {\begin{array}{*{20}l} {\delta_{{\dot{k}}}^{t} g_{{c \to \dot{k},ij}}^{t} \oplus \left( {1 - \delta_{{\dot{k}}}^{t} } \right)g_{{\dot{k},ij}}^{t} } \hfill & {\delta_{{\dot{k}}}^{t} \ge pc_{{\dot{k}}}^{t} } \hfill \\ {pc_{{\dot{k}}}^{t} g_{{c \to \dot{k},ij}}^{t} \oplus \left( {1 - pc_{{\dot{k}}}^{t} } \right)g_{{\dot{k},ij}}^{t} } \hfill & {\delta_{{\dot{k}}}^{t} < pc_{{\dot{k}}}^{t} } \hfill \\ \end{array} } \right.,$$
(14)

where \(g_{{c \to \dot{k},ij}}^{t} = \oplus_{{h = 1,h \ne \dot{k}}}^{K} g_{k,ij}^{t} \bar{\theta }_{h}^{t}\) is the collective opinion used to guide the adjustment of \(G_{{\dot{k}}}^{t}\) and \(\bar{\theta }_{h}^{t}\) is the remaining cluster \(c_{h}\)’s weight, which satisfies \(\bar{\theta }_{h}^{t} = {{\theta_{h}^{t} } \mathord{\left/ {\vphantom {{\theta_{h}^{t} } {\sum\nolimits_{{k = 1,k \ne \dot{k}}}^{K} {\theta_{k}^{t} } }}} \right. \kern-0pt} {\sum\nolimits_{{k = 1,k \ne \dot{k}}}^{K} {\theta_{k}^{t} } }}\). The addition operation “\(\oplus\)” is interpreted as

$$L_{{\dot{k},ij}}^{\left( u \right),t + 1} = \left\{ {\begin{array}{*{20}l} {\delta_{{\dot{k}}}^{t} L_{{c \to \dot{k},ij}}^{\left( u \right),t} \oplus \left( {1 - \delta_{{\dot{k}}}^{t} } \right)L_{{\dot{k},ij}}^{\left( u \right),t} } \hfill & {\delta_{{\dot{k}}}^{t} \ge pc_{{\dot{k}}}^{t} } \hfill \\ {pc_{{\dot{k}}}^{t} L_{{c \to \dot{k},ij}}^{\left( u \right),t} \oplus \left( {1 - pc_{{\dot{k}}}^{t} } \right)L_{{\dot{k},ij}}^{\left( u \right),t} } \hfill & {\delta_{{\dot{k}}}^{t} < pc_{{\dot{k}}}^{t} } \hfill \\ \end{array} } \right.,\quad u = 1, \ldots ,\# \# g_{{\dot{k},ij}}^{t} ,$$
(15)
$$p_{{\dot{k},ij}}^{\left( u \right),t + 1} = \left\{ {\begin{array}{*{20}l} {\delta_{{\dot{k}}}^{t} p_{{c \to \dot{k},ij}}^{\left( u \right),t} + \left( {1 - \delta_{{\dot{k}}}^{t} } \right)p_{{\dot{k},ij}}^{\left( u \right),t} } \hfill & {\delta_{{\dot{k}}}^{t} \ge pc_{{\dot{k}}}^{t} } \hfill \\ {pc_{{\dot{k}}}^{t} p_{{c \to \dot{k},ij}}^{\left( u \right),t} + \left( {1 - pc_{{\dot{k}}}^{t} } \right)p_{{\dot{k},ij}}^{\left( u \right),t} } \hfill & {\delta_{{\dot{k}}}^{t} < pc_{{\dot{k}}}^{t} } \hfill \\ \end{array} } \right.,\quad u = 1, \ldots ,\# \# g_{{\dot{k},ij}}^{t} ,$$
(16)

where \(p_{{c \to \dot{k},ij}}^{\left( u \right),t} = \sum\nolimits_{{h = 1,h \ne \dot{k}}}^{K} {p_{h,ij}^{\left( u \right),t} \bar{\theta }_{h}^{t} }\), \(L_{{c \to \dot{k},ij}}^{\left( u \right),t} = L_{{\dot{k},ij}}^{\left( u \right),t}\), \(u = 1, \ldots ,\# \# g_{{\dot{k},ij}}^{t}\), \(i = 1,2, \ldots ,m\), \(j = 1,2, \ldots ,n\).

4.3 Hierarchical Punishment-Driven Consensus Iterations

In Sect. 4.1, we divide the consensus measures into three levels. The group consensus index is used to determine whether the current consensus satisfies the threshold. The other consensus indexes can be used to determine which clusters’ opinions or matrix elements need to be adjusted. Different matrix elements in a cluster opinion usually have different consensus indexes. Therefore, it is reasonable to use different adjustment coefficients (or punishment coefficients).

Definition 11

For the consensus threshold \(\overline{GCI}\) and the identified cluster \(c_{{\dot{k}}}\) in the t-th iteration, the punishment coefficient at the matrix element level is defined as

$$pc_{{\dot{k},ij}}^{t} = \left\{ {\begin{array}{*{20}c} {\left( {\frac{{\overline{GCI} - ICI_{{\dot{k},ij}}^{t} }}{{\overline{GCI} }}} \right)^{\sigma } } & {ICI_{{\dot{k},ij}}^{t} < \overline{GCI} } \\ 0 & {ICI_{{\dot{k},ij}}^{t} \ge \overline{GCI} } \\ \end{array} } \right..$$
(17)

Let σ be the same as in Definition 9. Clearly, \(0 \le pc_{{\dot{k},ij}}^{t} \le 1\), \(i = 1,2, \ldots ,m\), \(j = 1,2, \ldots ,n\).

Based on Eq. (14), we use the following formula to adjust the matrix elements in the opinion \(G_{{\dot{k}}}^{t}\):

$$g_{{\dot{k},ij}}^{t + 1} = \left\{ {\begin{array}{*{20}l} {\delta_{{\dot{k},ij}}^{t} g_{{c \to \dot{k},ij}}^{t} \oplus \left( {1 - \delta_{{\dot{k},ij}}^{t} } \right)g_{{\dot{k},ij}}^{t} } \hfill & {\delta_{{\dot{k},ij}}^{t} \ge pc_{{\dot{k},ij}}^{t} } \hfill \\ {pc_{{\dot{k},ij}}^{t} g_{{c \to \dot{k},ij}}^{t} \oplus \left( {1 - pc_{{\dot{k},ij}}^{t} } \right)g_{{\dot{k},ij}}^{t} } \hfill & {\delta_{{\dot{k},ij}}^{t} < pc_{{\dot{k},ij}}^{t} } \hfill \\ \end{array} } \right.,$$
(18)

where \(\delta_{{\dot{k},ij}}^{t} \left( {0 \le \delta_{{\dot{k},ij}}^{t} \le 1} \right)\) is the adjustment coefficient given by cluster \(c_{{\dot{k}}}\) that is used to modify the probability value of the matrix element of position \(\left( {i,j} \right)\) in \(G_{{\dot{k}}}^{t}\).

In general, although the identified cluster has the lowest consensus index at the individual matrix level, it does not necessarily indicate that the consensus indexes of all elements in the cluster opinion do not satisfy the threshold. Using the same coefficient to adjust all the matrix elements may result in adjusting elements that should not be adjusted, or should not be adjusted by such a large amount, which will result in opinion distortion. We consider opinion distortion to have two parts: the adjustment amount of the opinion and the proportions of the adjusted matrix elements in the opinion. The latter part reflects the number of matrix elements that were adjusted but should not have been.

Definition 12

Based on Eq. (5), the adjustment amount of the identified opinion \(G_{{\dot{k}}}^{t}\) after the t-th iteration is calculated by

$$\overline{AA}_{{\dot{k}}}^{t + 1} = d\left( {G_{{\dot{k}}}^{t + 1} ,G_{{\dot{k}}}^{t} } \right).$$
(19)

Clearly, \(0 \le \overline{AA}_{{\dot{k}}}^{t + 1} \le 1\). A larger value of \(\overline{AA}_{{\dot{k}}}^{t + 1}\) indicates a greater adjustment in going from \(G_{{\dot{k}}}^{t}\) to \(G_{{\dot{k}}}^{t + 1}\).

Definition 13

Let \(IX\left( {G_{{\dot{k}}}^{t} ,G_{{\dot{k}}}^{t + 1} } \right)\) be the number of matrix elements that have been adjusted in the t-th iteration. The proportion of adjusted matrix elements is obtained by

$$\widehat{AA}_{{\dot{k}}}^{t + 1} = \frac{{IX\left( {G_{{\dot{k}}}^{t} ,G_{{\dot{k}}}^{t + 1} } \right)}}{m \times n}.$$
(20)

Definition 14

Given the adjustment amount and the proportion of adjusted matrix elements, the distortion degree of the opinion \(G_{{\dot{k}}}^{t}\) is defined as

$$DD_{{\dot{k}}}^{t + 1} = \frac{{\overline{AA}_{{\dot{k}}}^{t + 1} + \widehat{AA}_{{\dot{k}}}^{t + 1} }}{2} .$$
(21)

Clearly, \(0 \le DD_{{\dot{k}}}^{t + 1} \le 1\). A larger value of \(DD_{{\dot{k}}}^{t + 1}\) represents a higher distortion degree in going from \(G_{{\dot{k}}}^{t}\) to \(G_{{\dot{k}}}^{t + 1}\).

We propose the soft adjustment strategy, which allows different punishment coefficients to be used to adjust the matrix elements. Conversely, hard adjustment strategy requires that all matrix elements be adjusted with the same coefficient. Employing hard adjustment often leads to adjusting a larger number of matrix elements than employing soft adjustment in each consensus iteration. However, when the CRP begins, there are usually a large number of matrix elements in the identified opinion whose consensus indexes do not meet the requirements. Too much time will be spent reviewing the consensus index for each element and providing the adjustment coefficient if soft adjustment is used. Therefore, we consider that the hard adjustment strategy can be adopted first to improve the consensus indexes at different levels. When the group consensus index approaches the consensus threshold closely, the soft adjustment strategy is used to balance the relationship between opinion distortion and consensus improvement. To determine which adjustment strategy to adopt, we introduce a new parameter, the attainment rate of the consensus index, denoted as \(ARoCI\left( {0 \le ARoCI \le 1} \right)\). This parameter elevates the punishment-driven consensus model to a hierarchical consensus model.

Definition 15

Given the group consensus index \(GCI^{t}\) and consensus threshold \(\overline{GCI}\), the attainment rate of the consensus index is defined as

$$ARoCI^{t} = \frac{{GCI^{t} }}{{\overline{GCI} }}.$$
(22)

Clearly, \({{GCI^{0} } \mathord{\left/ {\vphantom {{GCI^{0} } {\overline{GCI} }}} \right. \kern-0pt} {\overline{GCI} }} \le ARoCI^{t} \le 1\). \(ARoCI^{t}\) reflects the percentage of the current consensus index that satisfies the consensus threshold. The higher the group consensus index, the greater the value of \(ARoCI^{t}\). DMs often set the threshold \(\overline{ARoCI} \in \left[ {{{GCI^{0} } \mathord{\left/ {\vphantom {{GCI^{0} } {\overline{GCI} }}} \right. \kern-0pt} {\overline{GCI} }},1} \right]\) for \(ARoCI^{t}\).

The essential difference between hard adjustment and soft adjustment is whether a matrix element that satisfies the consensus threshold should be adjusted when the consensus index at the matrix level does not meet the consensus threshold. We define the following rule.

Definition 16

Let \(IX\left( {G_{{\dot{k}}}^{t} } \right)\) be the number of matrix elements that contribute less to the consensus. When the conditions \(ARoCI^{t} \ge \overline{ARoCI}\) and \({{IX\left( {G_{{\dot{k}}}^{t} } \right)} \mathord{\left/ {\vphantom {{IX\left( {G_{{\dot{k}}}^{t} } \right)} {m \times n}}} \right. \kern-0pt} {m \times n}} \le \overline{IX}\) are met, the soft adjustment strategy is adopted; otherwise, the hard adjustment strategy is used. \(\overline{IX} \left( {0 \le \overline{IX} \le 1} \right)\) is the threshold of \({{IX\left( {G_{{\dot{k}}}^{t} } \right)} \mathord{\left/ {\vphantom {{IX\left( {G_{{\dot{k}}}^{t} } \right)} {m \times n}}} \right. \kern-0pt} {m \times n}}\).

If \(ARoCI^{t} \ge \overline{ARoCI}\) and \({{IX\left( {G_{{\dot{k}}}^{t} } \right)} \mathord{\left/ {\vphantom {{IX\left( {G_{{\dot{k}}}^{t} } \right)} {m \times n}}} \right. \kern-0pt} {m \times n}} \le \overline{IX}\), this indicates that the group consensus index is high enough, and the number of elements in the identified opinion that need to be adjusted drops significantly. In this case, the soft adjustment strategy is used. Clearly, when we set \(\overline{ARoCI} = 1\) or \(\overline{IX} = 0\), the hard adjustment strategy will be used throughout the CRP. By combining the hard adjustment and soft adjustment, the hierarchical punishment-driven consensus model (HPDCM) is obtained. Algorithm 2 presents the implementation process for the model.

figure b

5 Case Study

In this section, we apply the HPDCM to a PL-LGDM problem concerning global supplier selection.

5.1 Problem Description

ABC, an electronics manufacturer located in Quanzhou, China, is reassessing the suppliers it currently works with. Considering the rising cost of domestic labor and the policy advantages of the Belt and Road cooperation initiative, the company has decided to select a material supplier located along this route and add it to the supplier list. Five alternatives \(X = \left\{ {x_{1} ,x_{2} ,x_{3} ,x_{4} ,x_{5} } \right\}\) have been preselected for further evaluation, and their geographical locations are Alma-Ata, Ho Chi Minh City, Manila, Calcutta and Minsk, as shown in Fig. 4. A special committee of 20 relevant department heads and experts is formed to evaluate the alternative suppliers. Drawing from the research (Awasthi et al. 2018; Reefke and Sundaram 2018; Viswanadham and Samvedi 2013; Yücenur et al. 2011) and business practices, three comprehensive criteria are used: commercial factors (\(c_{1}\)), such as quality, price, quantity and delivery time; sustainability dimensions (\(c_{2}\)), including economic, environmental, and social bottom lines; and global risks (\(c_{3}\)), such as currency fluctuations, political instability, terrorism, and cultural incompatibility. The weight vector of the criteria is set as \(\omega = \left( {0.4,0.3,0.3} \right)^{T}\). The DMs provide their evaluations by means of PLTSs, in which the linguistic evaluation scale

$$S = \left\{ \begin{array}{l} s_{ - 3} = Extremely \, poor, \, s_{ - 2} = Poor, \, s_{ - 1} = Somewhat \, poor, \, s_{0} = Neutral, \, \hfill \\ s_{1} = Somewhat \, good, \, s_{2} = Good, \, s_{3} = Extremely \, good \hfill \\ \end{array} \right\}$$

is used. The original individual opinions are given in Online Resource.

Fig. 4
figure 4

Geographical distribution of preselected suppliers along the Belt and Road. Note: The original source of this map is http://www.sohu.com/a/111613685_472018

5.2 Decision Process

Here, the HPDCM is used to manage the processes of consensus and selection. The following steps are involved. To save space, only the main results are listed. A detailed textual description of the decision process is given in Online Resource.

Input::

The initial individual opinions \(V_{l}^{0} \left( {l = 1,2, \ldots ,20} \right)\), the weight vector of the criteria ω, the consensus threshold \(\overline{GCI} = 0.88\), and other parameters \(\overline{ARoCI} = 0.95\), \(\overline{IX} = 1\), and \(\sigma = {1 \mathord{\left/ {\vphantom {1 3}} \right. \kern-0pt} 3}\)

Output::

The final ranking of the alternatives

Process 1

Preprocessing of the decision information.

Normalize the individual opinions. Here, we just give the normalized opinion of cluster \(c_{1}\):

$$R_{1}^{0} = \left( {\begin{array}{*{20}l} {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( 0 \right),s_{0} \left( {0.5} \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( {0.4} \right),s_{3} \left( {0.1} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 1 \right),s_{ - 1} \left( 0 \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.1} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( {0.9} \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.5} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( {0.5} \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( 0 \right),s_{0} \left( {0.4} \right), \hfill \\ s_{1} \left( {0.6} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.5} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.5} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.3} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.7} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( {0.2} \right),s_{ - 1} \left( {0.3} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.5} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.7} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.3} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.2} \right),s_{0} \left( {0.8} \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( 0 \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( 1 \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.6} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.4} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( {0.5} \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( {0.5} \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( {0.2} \right),s_{ - 1} \left( 0 \right),s_{0} \left( {0.8} \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( 0 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( 0 \right),s_{ - 1} \left( 0 \right),s_{0} \left( 0 \right), \hfill \\ s_{1} \left( 0 \right),s_{2} \left( 1 \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ \end{array} } \right)$$

Process 2

Clustering process.

Use Algorithm 1 to divide the large group into six clusters by setting the minimum distance to 0.23, as shown in Fig. 5. As previously discussed, we set \(\beta = 1\), \(a = round\left( {q \cdot 0.1} \right) = 2\), and \(b = round\left( {{q \mathord{\left/ {\vphantom {q m}} \right. \kern-0pt} m}} \right) = 4\). Table 3 presents the clustering results.

Fig. 5
figure 5

Visual clustering processes using Algorithm 1

Table 3 Clustering results by using Algorithm 1

Process 3

Consensus reaching process.

Use the HPDCM to manage opinion differences. Table 4 shows the consensus iterations.

Table 4 Consensus iterations using the HPDCM

In conclusion, after three consensus iterations (including one hard adjustment and two soft adjustments), the group consensus index satisfies the consensus threshold. Let \(t^{ * } = 3\), and the final opinions of the clusters can be obtained as \(G_{k}^{3} \left( {k = 1,2, \ldots ,6} \right)\).

Process 4

Selection process.

Step 4.1:

Aggregate all the clusters’ opinions to obtain the group opinion via Eq. (3):

$$R_{c} = \left( {\begin{array}{*{20}l} {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.1715} \right),s_{ - 2} \left( {0.148} \right),s_{ - 1} \left( {0.2116} \right),s_{0} \left( {0.0937} \right), \hfill \\ s_{1} \left( {0.168} \right),s_{2} \left( {0.1334} \right),s_{3} \left( {0.0739} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0461} \right),s_{ - 2} \left( {0.2101} \right),s_{ - 1} \left( {0.2349} \right),s_{0} \left( {0.1238} \right), \hfill \\ s_{1} \left( {0.2604} \right),s_{2} \left( {0.0547} \right),s_{3} \left( {0.0701} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.086} \right),s_{ - 2} \left( {0.2399} \right),s_{ - 1} \left( {0.1439} \right),s_{0} \left( {0.269} \right), \hfill \\ s_{1} \left( {0.1584} \right),s_{2} \left( {0.1028} \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0406} \right),s_{ - 2} \left( {0.0515} \right),s_{ - 1} \left( {0.1458} \right),s_{0} \left( {0.2173} \right), \hfill \\ s_{1} \left( {0.1082} \right),s_{2} \left( {0.2195} \right),s_{3} \left( {0.2172} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( {0.0504} \right),s_{ - 1} \left( {0.1659} \right),s_{0} \left( {0.2464} \right), \hfill \\ s_{1} \left( {0.2903} \right),s_{2} \left( {0.2094} \right),s_{3} \left( {0.0375} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0226} \right),s_{ - 2} \left( {0.1518} \right),s_{ - 1} \left( {0.1864} \right),s_{0} \left( {0.307} \right), \hfill \\ s_{1} \left( {0.1746} \right),s_{2} \left( {0.0961} \right),s_{3} \left( {0.0615} \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.045} \right),s_{ - 2} \left( {0.0311} \right),s_{ - 1} \left( {0.3031} \right),s_{0} \left( {0.1539} \right), \hfill \\ s_{1} \left( {0.2198} \right),s_{2} \left( {0.0711} \right),s_{3} \left( {0.1761} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0926} \right),s_{ - 2} \left( {0.0839} \right),s_{ - 1} \left( {0.187} \right),s_{0} \left( {0.3122} \right), \hfill \\ s_{1} \left( {0.1895} \right),s_{2} \left( {0.1129} \right),s_{3} \left( {0.0219} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0057} \right),s_{ - 2} \left( {0.062} \right),s_{ - 1} \left( {0.2705} \right),s_{0} \left( {0.2153} \right), \hfill \\ s_{1} \left( {0.2141} \right),s_{2} \left( {0.1555} \right),s_{3} \left( {0.0768} \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0216} \right),s_{ - 2} \left( {0.1011} \right),s_{ - 1} \left( {0.157} \right),s_{0} \left( {0.2611} \right), \hfill \\ s_{1} \left( {0.0519} \right),s_{2} \left( {0.4072} \right),s_{3} \left( 0 \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.196} \right),s_{ - 2} \left( {0.0667} \right),s_{ - 1} \left( {0.3346} \right),s_{0} \left( {0.0992} \right), \hfill \\ s_{1} \left( {0.1843} \right),s_{2} \left( {0.0692} \right),s_{3} \left( {0.0501} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0242} \right),s_{ - 2} \left( {0.1133} \right),s_{ - 1} \left( {0.1819} \right),s_{0} \left( {0.2911} \right), \hfill \\ s_{1} \left( {0.178} \right),s_{2} \left( {0.1897} \right),s_{3} \left( {0.022} \right) \hfill \\ \end{array} \right\}} \hfill \\ {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.0856} \right),s_{ - 2} \left( {0.1248} \right),s_{ - 1} \left( {0.1129} \right),s_{0} \left( {0.1323} \right), \hfill \\ s_{1} \left( {0.2006} \right),s_{2} \left( {0.2018} \right),s_{3} \left( {0.142} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( {0.1194} \right),s_{ - 2} \left( {0.1515} \right),s_{ - 1} \left( {0.1622} \right),s_{0} \left( {0.1663} \right), \hfill \\ s_{1} \left( {0.1912} \right),s_{2} \left( {0.1152} \right),s_{3} \left( {0.0943} \right) \hfill \\ \end{array} \right\}} \hfill & {\left\{ \begin{array}{*{20}l} s_{ - 3} \left( 0 \right),s_{ - 2} \left( {0.1183} \right),s_{ - 1} \left( {0.1965} \right),s_{0} \left( {0.2072} \right), \hfill \\ s_{1} \left( {0.1236} \right),s_{2} \left( {0.2806} \right),s_{3} \left( {0.0738} \right) \hfill \\ \end{array} \right\}} \hfill \\ \end{array} } \right)$$
Step 4.2:

Calculate the overall criterion values \(Z\left( {x_{i} } \right)\left( {i = 1,2, \ldots ,m} \right)\) based on Eq. (3), as follows:

$$\begin{aligned} & Z\left( {x_{1} } \right) = \left\{ {s_{ - 3} \left( {0.1051} \right),s_{ - 2} \left( {0.182} \right),s_{ - 1} \left( {0.1858} \right),s_{0} \left( {0.148} \right),s_{1} \left( {0.181} \right),s_{2} \left( {0.0976} \right),s_{3} \left( {0.0499} \right)} \right\}, \\ & Z\left( {x_{2} } \right) = \left\{ {s_{ - 3} \left( {0.0229} \right),s_{ - 2} \left( {0.0793} \right),s_{ - 1} \left( {0.1552} \right),s_{0} \left( {0.2322} \right),s_{1} \left( {0.1726} \right),s_{2} \left( {0.1697} \right),s_{3} \left( {0.1138} \right)} \right\}, \\ & Z\left( {x_{3} } \right) = \left\{ {s_{ - 3} \left( {0.0469} \right),s_{ - 2} \left( {0.0552} \right),s_{ - 1} \left( {0.2376} \right),s_{0} \left( {0.2046} \right),s_{1} \left( {0.1949} \right),s_{2} \left( {0.1052} \right),s_{3} \left( {0.0978} \right)} \right\}, \\ & Z\left( {x_{4} } \right) = \left\{ {s_{ - 3} \left( {0.0737} \right),s_{ - 2} \left( {0.0916} \right),s_{ - 1} \left( {0.2031} \right),s_{0} \left( {0.2068} \right),s_{1} \left( {0.1244} \right),s_{2} \left( {0.2267} \right),s_{3} \left( {0.0215} \right)} \right\}, \\ & Z\left( {x_{5} } \right) = \left\{ {s_{ - 3} \left( {0.688} \right),s_{ - 2} \left( {0.1253} \right),s_{ - 1} \left( {0.1452} \right),s_{0} \left( {0.1561} \right),s_{1} \left( {0.1651} \right),s_{2} \left( {0.1871} \right),s_{3} \left( {0.1037} \right)} \right\} \\ \end{aligned}$$
Step 4.3:

Based on the score function of PLTSs proposed by Pang et al. (2016), we can calculate the scores of the overall criterion values:

\(E\left( {Z\left( {x_{1} } \right)} \right) = s_{ - 0. 3 3 9 2}\), \(E\left( {Z\left( {x_{2} } \right)} \right) = s_{0.4 7 0 9}\), \(E\left( {Z\left( {x_{3} } \right)} \right) = s_{0. 2 1}\), \(E\left( {Z\left( {x_{4} } \right)} \right) = s_{0. 0 3 5}\), \(E\left( {Z\left( {x_{5} } \right)} \right) = s_{0. 2 4 8 2}\)

Therefore, the ranking is \(x_{2} \succ x_{5} \succ x_{3} \succ x_{4} \succ x_{1}\), and the best solution is \(x_{2}\). That is, the DMs are more likely to include the supplier in Ho Chi Minh City on the supplier list.

6 Comparative Analysis and Discussion

We discuss some additional important issues about the HPDCM in Sect. 6.1, including a comparison of hard adjustment and soft adjustment and the determination of relevant parameters. Our proposal is compared with other linguistic models in Sect. 6.2. Section 6.3 presents the managerial implications involved in the practical application of the proposed model. Note that the data used in the comparative analysis and discussion originate from the case study in Sect. 5.

6.1 Further Discussion of the Hierarchical Punishment-Driven Consensus Model

This section analyzes two important issues regarding the HPDCM: the comparison between hard adjustment and soft adjustment and the calculation of the threshold for the attainment rate of the consensus index.

6.1.1 Hard Adjustment and Soft Adjustment

To eliminate DMs’ subjective attitudes towards adjustment, we adopt the punishment coefficient in implementing the CRP. Figure 6 shows the consensus results obtained by hard adjustment and soft adjustment. First, regardless of which type of adjustment strategy is utilized, the group consensus index continues to rise as the number of consensus iterations increases. This indicates that the two adjustment strategies are effective in promoting consensus improvement. However, different adjustment strategies lead to different degrees of opinion distortion. For example, as shown in Fig. 6, the distortion degree is 0.5464 when using hard adjustment after the first iteration, while the soft adjustment results in only a 0.4813 distortion degree. This is because adopting a hard adjustment strategy causes some matrix elements that should not have been adjusted to be adjusted.

Fig. 6
figure 6

Consensus results obtained by hard adjustment and soft adjustment. Note: GCI and DD represent the group consensus index and the distortion degree, respectively

Although, as noted above, soft adjustment performs better than hard adjustment in terms of opinion distortion, this does not mean that soft adjustment can completely replace hard adjustment. Hard adjustment requires only one adjustment coefficient, which is easy to obtain, especially when no punishment coefficient is used as a reference. The hard adjustment may lead to undue distortion, because even if a certain matrix element already satisfies the consensus threshold, it will still be adjusted. Soft adjustment refines the adjustment range to the matrix element level, which helps reduce unnecessary distortion. However, for the DMs, too much time is spent providing adjustment coefficients according to the different consensus indexes of matrix elements. We conclude that hard adjustment is better suited to situations where (1) the decision process has just begun, (2) the identified matrix is low-dimensional, or (3) the decision is time-critical. When the group consensus index has increased significantly and most elements in the identified opinion have already met the consensus threshold, soft adjustment takes precedence.

6.1.2 Determination of the Relevant Parameters

In this section, we discuss how to assign the following two parameters: the threshold of the attainment rate of the consensus index \(\overline{ARoCI}\) and the power of the punishment coefficient σ.

The attainment rate of the consensus index plays a decisive role in whether to adopt hard adjustment or soft adjustment. A high threshold can delay the use of soft adjustment strategy, which may increase opinion distortion. A low threshold will lead to premature use of soft adjustment strategy; this may make it difficult for the cluster to provide the adjustment coefficient at the matrix element level because there may be multiple elements in the identified opinion that need to be adjusted. We suggest that determining the threshold should follow two principles:

  1. 1

    Computer software is used to simulate the CRP so as to provide reference for practical operation.

  2. 2

    There are two factors to consider: the number of consensus iterations and the degree of opinion distortion.

Assume that the following requirements are met: the maximum number of iterations is 5 and the opinion distortion is no more than 0.5 in each iteration. We implement punishment-driven consensus iterations using MATLAB software. Figure 7 shows the simulation results of the number of iterations and the distortion degree when σ increases from 0 to 0.5. When \(a = 0.5\), the number of iterations reaches 8 by using the hard adjustment strategy (or 6 by using the soft adjustment strategy). According to the function characteristic of punishment coefficient, the setting \(a > 0.5\) will lead to a greater number of iterations. Therefore, Fig. 7 only depicts the simulation results under the condition of \(\sigma \in \left[ {0,0.5} \right]\). We find that as σ increases, the number of iterations and the total distortion degree increase, but from the overall trend, the average distortion degree decreases. The average distortion degree is defined as the total distortion degree divided by the number of iterations. Determining the parameter σ includes the following steps.

Fig. 7
figure 7

Consensus results when \(\sigma\) increases from 0 to 0.5. Note: TDD represents the total distortion degree, which is calculated by \(\sum\nolimits_{t} {DD_{k}^{t} }\). ADD is the average distortion degree, which is calculated by \({1 \mathord{\left/ {\vphantom {1 t}} \right. \kern-0pt} t}\sum\nolimits_{t} {DD_{k}^{t} }\)

Step 1:

Observe the simulation of the number of iterations and determine the membership interval of σ

If adopting the hard adjustment strategy, then we can set \(\sigma \in \left[ {0,0.4} \right]\); If adopting the soft adjustment strategy, then we can set \(\sigma \in \left[ {0,0.5} \right)\). By taking the intersection of the above two intervals, the membership interval is temporarily determined as \(\sigma \in \left[ {0,0.4} \right]\).

Step 2:

Modify the temporary interval based on the simulation of opinion distortion

From Fig. 7, the average distortion will always be greater than 0.5 if the hard adjustment strategy is adopted, while the opposite result will be obtained if the soft adjustment strategy is adopted. Therefore, σ should be assigned a larger value so that the average distortion is as small as possible. Based on the obtained membership interval, we can set \(\sigma = 0.4\). In this case, the soft adjustment strategy must be adopted at least once.

Without loss of generality, we assume that the hard adjustment strategy is allowed to be used once. In fact, this case can only happen at the beginning of the CRP. Figure 8 shows the changes of some parameters in the CRP when setting \(\sigma = 0.4\). From Fig. 8, the following results can be obtained \(ARoCI^{0} = 0.9437\), \(ARoCI^{1} = 0.9698\), and \({{IX\left( {G_{{\dot{k}}}^{1} } \right)} \mathord{\left/ {\vphantom {{IX\left( {G_{{\dot{k}}}^{1} } \right)} {m * n}}} \right. \kern-0pt} {m * n}} = 0.7333\). To ensure that the hard adjustment strategy is adopted in the first iteration, we set \(\overline{ARoCI} > ARoCI^{0}\). In the second iteration the soft adjustment strategy can be adopted when setting \(\overline{ARoCI} \le ARoCI^{1} = 0.9698\) and \(\overline{IX} \ge 0.7333\). In conclusion, we have that \(\overline{ARoCI} \in \left( {0.9437,0.9698} \right]\) and \(\overline{IX} \ge 0.7333\). Specifically, the parameters can be set as \(\overline{ARoCI} = 0.95\) and \(\overline{IX} = 1\).

Fig. 8
figure 8

Changes in some parameters in the CRP. Note: Set \(\sigma = 0.4\). \(IX\) represents the proportion of adjusted matrix elements in the identified opinion. \(ARoCI\) is the attainment rate of the consensus index

6.2 Comparison to Other Linguistic Models

This section presents a comparison of our proposal with Pang et al.’s (2016) research and HFLTSs in the decision-making process.

An HFLTS represents hesitation among multiple linguistic terms but cannot reflect the importance of these linguistic terms. In the kind of case we are considering, if the decision information is expressed by HFLTSs, this can be interpreted as indicating that all possible linguistic terms in an HFLTS have the same importance degree. Accordingly, an HFLTS can be written in the form of a PLTS, in which all the linguistic terms have the same possibility. For instance, if the opinion \(v_{1,13}^{0} = \left\{ {s_{ - 1} \left( {0.1} \right),s_{2} \left( {0.9} \right)} \right\}\) is represented using an HFLTS, it is changed to \(\left\{ {s_{ - 1} \left( {0.5} \right),s_{2} \left( {0.5} \right)} \right\}\).

On the other hand, since the concept of PLTSs was proposed by Pang et al. (2016), research on aggregation and distance measurement for PLTSs has attracted much attention. In this section, we compare our proposal with Pang et al.’s (2016) model in terms of the effect that processing PLTSs has on the decision results. In general, there are two models for addressing PLTSs: (1) Probability values are incorporated into the linguistic terms in the process of information aggregation (see Pang et al. (2016)). In this model, the probabilities of linguistic terms are not reflected in the result, and the original discrete linguistic terms are transformed into virtual ones. (2) Similar to our proposal, before entering the selection process, the information aggregation only involves operations on probability values corresponding to the same linguistic terms and does not merge probability values and linguistic terms.

To ensure comparability, we set the same number of clusters, i.e., \(K = 6\). From the visual clustering processes shown in Figs. 5 and 9, the most significant differences are the order in which new clusters are formed and the distances between DMs. Table 5 depicts the results of the first three clusters formed by using different linguistic models. Our proposal differs from the other two models in forming the second and third clusters. This is due to the distance measure. As \(d\left( {R_{5}^{0} ,R_{8}^{0} } \right) > d\left( {R_{4}^{0} ,R_{9}^{0} } \right)\) in our proposal, the second cluster is composed of \(e_{4} , \, e_{9}\); however, in the other two models, since \(d\left( {R_{5}^{0} ,R_{8}^{0} } \right) < d\left( {R_{4}^{0} ,R_{9}^{0} } \right)\), DMs \(e_{5} , \, e_{8}\) form the second cluster. It is important to note that if HFLTSs are used, the distance between \(R_{2}^{0}\) and \(R_{17}^{2}\) is 0, which is different from the result obtained by using our proposal or Pang et al.’s (2016) model. This is due to the fact that HFLTSs do not reflect the weights of the linguistic terms.

Fig. 9
figure 9

Visualization of the clustering process when taking different approaches to decision information

Table 5 Clustering results of the first three clusters formed by using different linguistic models

Owing to different clustering results, the quantitative comparison of the CRP seems unfair. Nonetheless, we consider that it is difficult for DMs to understand the meaning of virtual linguistic terms, as opposed to the discrete linguistic terms with corresponding probabilities. Additionally, the alternative rankings prior to the CRP are depicted in Fig. 10. All three linguistic models yield the same optimal solution (i.e., \(x_{2}\)) and the same ranking (i.e., \(x_{2} \succ x_{5} \succ x_{3} \succ x_{4} \succ x_{1}\)). However, there are large differences in the scores of the alternatives, although this does not affect the selection of the best alternative.

Fig. 10
figure 10

Initial rankings of the alternatives obtained by using different linguistic models

Above all, our proposal overcomes the shortcomings of the other two models in processing linguistic information. Unlike HFLTSs, which represent hesitations regarding multiple linguistic terms, our proposal based on PLTSs considers the importance of the linguistic terms, which is more in line with actual circumstances. On the other hand, using Pang et al.’s (2016) model to handle PLTSs leads to the probabilities of linguistic terms not being reflected in the result. In this way, the unique feature of PLTSs compared with ordinary linguistic term sets is lost. Furthermore, according to the statement “the ordinal discrete linguistic terms are used to evaluate alternatives, while the virtual linguistic terms can only appear in the operation and ranking” (Xu 2009), virtual linguistic terms should be avoided in the CRP. Therefore, our proposal insists on preserving the original discrete linguistic term structure of PLTSs in the processes of clustering, aggregating and consensus reaching, and only after beginning the selection process are virtual linguistic terms introduced to calculate the scores of PLTSs.

6.3 Managerial Implications

The proposed HPDCM can be applied to the practical case of supplier selection in the manufacturing and service industries, and the following important issues are raised:

  1. 1

    Supplier selection is a strategic and complex decision with risks that is closely related to the survival and sustainable development of enterprises. It is therefore necessary to measure the group consensus index and make any useful adjustments to ensure that the result is accepted by the majority. This arrangement can reduce the risks and losses of decision-making errors, but more importantly, only when a result is approved by the majority can it obtain the maximum practical support from the majority of departments.

  2. 2

    The increase in the consensus index is due to some DMs adjusting their own views, whether voluntarily or involuntarily. In this way, the adjusted opinion may be somewhat distorted from the original opinion. Therefore, it is important to measure the degree of opinion distortion and use it as an indicator to evaluate the performance of the CRP. In business practice, some DMs insist on their opinions and are unwilling to compromise for the sake of consensus. Even if the adjustments are forced by the punishment-driven consensus model, the reasonableness of the adjusted opinion needs to be examined. The distortion degree is an important index to test rationality. If the distortion of a DM’s opinion is high in an iteration (close to 1), the DM’s original opinion is likely to deviate significantly from the majority opinion.

7 Conclusions

In this paper, a hierarchical punishment-driven consensus model for probabilistic linguistic large-group decision-making was developed, and an application of the model to global supplier selection was presented. The obtained results show that the new model can overcome the scalability challenge and soften the human supervision of the CRP. The major contributions of this study are as follows:

  1. 1

    The PL-LGDM was defined, and its characteristics were detailed. To improve the performance of PLTSs in the CRP, the rules governing their normalization and operations were redefined.

  2. 2

    A hierarchical punishment-driven consensus model for PL-LGDM was developed, including hard adjustment and soft adjustment, which aimed to reduce the distortion of opinions with the goal of achieving a consensus within a specified number of iterations. Before the CRP begins, the large group was classified into several subgroups by hierarchical clustering, and the opinions of clusters were generated. This operation can overcome the scalability challenge when the number of DMs involved is sufficiently large. When the necessary parameters are set, the model can automatically provide guidance for the adjustment of opinions, thereby reducing the human supervision cost of the CRP.

  3. 3

    The model was applied to a case study of global supplier selection, and the application implications were discussed.

It is worth noting that there are always social relationships between DMs (Ding et al. 2019; Wu et al. 2018). The influence of social relationships on the CRP should be considered. As further research, we will study the theoretical framework of consensus building in LGDM problems in the context of social networks.