Keywords

1 Introduction

Nowadays, selecting the optimal solution/alternative based on multiple criteria for a given decision problem is a major task for decision makers due to the availability of criteria with different scales/ranges. Multi-Criteria Decision Making (MCDM) is used by most decision makers to deal with the decision problems. They carry out each MCDM method by defining a decision matrix that consists of a set of alternatives Ai (i = 1, …, m), criteria Cj (j = 1,…, n), the relative importance of the criteria (or weights) Wj, and rij, corresponding the rating of alternative i with respect to criteria j [1]. Most of the criteria are measured on different scales (e.g., weight, height, temperature, etc.) while they need to be defined in the same scale/range to enable decision makers to make a valid comparison and selection of the best solution. In other words, decision makers need some preprocessing to produce comparable and dimensionless data from heterogeneous input data sets that is called normalization process. Generally, MCDM methods have four main steps as [2]:

  1. 1)

    Determine the decision matrix

  2. 2)

    Normalize the criteria values in the decision matrix

  3. 3)

    Calculate weighted normalized values

  4. 4)

    Aggregation process (which differs for each MCDM method)

  5. 5)

    Sort alternatives in a decreasing ordering.

There are some research papers about the importance of normalization techniques in MCDM methods [39]. They discuss the fact that different normalization techniques may address different ranking of alternatives. So, selecting an improper normalization technique may cause deviation from the original solution. Therefore, normalization has an essential role in most MCDM methods and using the most proper technique will help decision makers to improve the accuracy of the final solution.

Furthermore, new developments in big data and Artificial Intelligence (AI) and the integration with MCDM methods provide novel aspects for handling collected data from heterogeneous sensors. These big data from sensors need a normalization process to produce dimensionless input data sets to be used in decision problems. So, normalization techniques have important role in these new points of view of science.

Several normalization techniques are introduced in the literature along with different MCDM methods [1]. In this study, we analyze the effects of different normalization techniques on the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) method (to get more details about TOPSIS please see [7]). Thus, the main research question related to this study is: Which normalization technique is the most appropriate for usage with the TOPSIS method?

This paper shows the applicability of the assessment framework proposed by [8] and compares the suitability of different normalization techniques using the TOPSIS method. In order to test the mentioned framework, a case study was borrowed from [10], which analyzed the effects of five normalization techniques as Max, Max-Min, Sum, Vector, and Fuzzification (Gaussian) with the TOPSIS method. The obtained results are compared with the results of initial work from [10].

2 Contribution to Applied Artificial Intelligence Systems

In recent years, with the advent of big data, organizations deal with complex decision problems consisting of large amounts of data. So, the integration of MCDM and Artificial Intelligence (AI) techniques is considered and reached success in handling real-world problems [11]. This integration enables decision makers to better structure complex decision problems in static and distributed environments. Doumpos and Grigoroudisby [11] mention that these help as “handling of massive data sets, the modelling of ill-structured information, the construction of advanced decision models, and the development of efficient computational optimization algorithms for problem solving”.

Several AI techniques are introduced in the literature which are used with integration MCDM, such as Fuzzy Logic (FL), Genetic Algorithm (GA), Neural Network (NN), Heuristic or meta-heuristics, Knowledge-Based (KB), Expert Systems (ES), Tabu-Search (TS), Simulated-annealing (SA), Dampster Shafer (DS), and Self-Organizing-Map (SOM). However, FL is the most popular technique to be used with MCDM methods (e.g., Fuzzy-AHP, Fuzzy-ANP, and Fuzzy-TOPSIS, etc.) [12].

There are some researches on implementing AI techniques with MCDM methods such as Ho [13] that uses 8 meta-heuristics along with AHP, Pan [14] applies Fuzzy-AHP for bridge construction methods selection, Sheu [15] uses Fuzzy-AHP, Fuzzy-TOPSIS, and Fuzzy-MCDM for global logistic operational model, Kulturel-Konak et al. [16] apply TS for system redundancy allocation problem, Efendigil et al. [17] implement ANN and Fuzzy-AHP for third-party logistics providers selection, and Wu et al. [18] use Fuzzy-ANP for site selection problem.

3 Assessment Framework for Evaluation of Normalization Techniques

As mentioned above, normalization is a vital step of most MCDM decision problems that “transfers” all criteria to the same scale and provides an effective comparison between alternatives to select the optimal solution. Numerous normalization techniques have been introduced in the literature. For instance, Jahan and Edwards [1] listed 31 techniques and addressed the pros and cons of them and pointed the influence effects of normalization on the results of MCDM methods in material selection decision problems. The authors of [1] also discuss the different features that should be considered to evaluate a normalization technique namely ranking reversal, symmetry, handling negative values, and capability in removing scales. Furthermore, Vafaei et al. [8] added data type and topology to the initial features from [1] as an important aspect for considering in the evaluation process.

Among several introduced normalization techniques in the literature [1], some of them are more well-known and implemented for the specific mission or MCDM methods. For example, Vector normalization is often applied with TOPSIS method, Target base normalization technique is utilized for material selection decision problems, and Sum is used with AHP (Analytical Hierarchy Process) method [1]. However, there are no clear reasons in the literature for these specific usages, which motivated us to evaluate different normalization techniques in MCDM decision problems.

Some research papers investigated the evaluation of the effects of normalization techniques in MCDM decision problems using various metrics. For instance, Celen [3] analyzed the effect of four normalization techniques (Max, Max-Min, Sum, and Vector) in TOPSIS method using consistency conditions metrics and showed that Vector is the most appropriate normalization technique for the used case study [3]. Also, Charaborty and Yeh [4, 5] used Ranking Consistency Index (RCI) for evaluating the same normalization techniques (Max, Max-Min, Sum, and Vector) for TOPSIS [4] and SAW [5] methods. In another study, Lakshmi and Venkatesan [5] assessed five normalization techniques namely Max, Max-Min, Sum, Vector, Fuzzification (Gaussian membership function) with TOPSIS method using time and space complexity in MATLAB software and selected Sum normalization technique as the best for the case study. In addition, Mathew et al. [6] implemented Spearman correlation to evaluate 6 normalization techniques for the weighted aggregated sum product assessment (WASPAS) method and suggested Max-Min technique as the most suitable normalization for the case study. Recently, Vafaei et al. [79, 19, 20] developed an assessment framework to evaluate different normalization techniques for MCDM methods. The proposed assessment framework is consisted of several metrics that are presented in Fig. 1.

Fig. 1.
figure 1

Three level of the evaluation framework (adopted from [8]).

In this study, we show the applicability of the evolved framework that is proposed by [8]. Figure 1 shows that the first level of this framework is related to distinguishing the data type and topology of the input data sets. Proceeding to the second level enables decision makers to select the normalization techniques. Following, the third level that is the main part of the framework, includes several metrics to assess the used normalization techniques in the decision problems and recommend the most proper normalization techniques for the related case study (for more detail about the assessment framework please see [8]).

In the next section, we show the usage of this assessment framework using a case study that is borrowed from [10]. The related case study from [10] assessed five normalization techniques i.e. Max, Max-Min, Sum, Vector, Fuzzification (Gaussian membership function). So, in order to provide the possibility of benchmarking between our results and the initial results from the author we use the same normalization techniques. The formulas of the mentioned normalization techniques are presented in Table 1. It is noticeable that benefit formulas refer to the higher values the better and cost formulas represent the lower value is desirable.

Table 1. Well-known normalization techniques (adapted from [10]).

4 Comparison of Normalization Techniques with an Illustrative Example for TOPSIS Method

To test the applicability of the proposed assessment framework, we applied metrics from the framework to the small illustrative example borrowed from [10]. This case study analyzed and recommended normalization techniques for TOPSIS method in a car selection problem. It consists of 4 alternatives (A1, …, A4) regarding different car brands (Civic Coupe, Saturn Coupe, Ford Escort, and Mazda Miata) and 4 criteria (C1, …, C4) related to the car characteristics (style, reliability, fuel-eco, and cost). The decision matrix with the input data is depicted in Table 2.

Table 2. Decision matrix input data (borrowed from [10])

The authors of [10] used five normalization techniques (Max, Max-Min, Sum, Vector, and Fuzzification (using Gaussian membership function)) to analyze the effects of using different normalization techniques on the TOPSIS method. They calculated the relative closeness and ranking of alternatives, with TOPSIS, using the selected normalization techniques for this case study (Table 3). For more details about the TOPSIS method please see [7].

Table 3. Relative closeness (RC) and Ranking of alternatives (R) the case study from [10]

As Table 3 depicts, there is no consensus between the five selected normalization techniques for ranking alternatives. The work described in [10] calculated time complexity and space complexity for each normalization technique with the help of MATLAB and recommended the Sum normalization technique as the best one for the case study, using the TOPSIS method.

Now, we apply the proposed assessment framework and recommend the most appropriate normalization technique for the related case study and compare our results with the obtained results by the authors of [10].

In the first level of the assessment framework the topology of input data is considered. However, in this case study the decision matrix contains integer numbers between [6,7,8,9] without any outliers or decimal numbers, and zero. So, we skip the first level and proceed to the second level for selecting the normalization techniques to be assessed.

As mentioned before, the authors of [10] analyzed the effect of five normalization techniques (Max, Max-Min, Sum, Vector, and Fuzzification (using Gaussian membership function)) i.e. at least one normalization technique from three categories (Linear, Semi-linear, and Non-linear) that are mentioned in the second level of the assessment framework (Fig. 1) are selected.

The third level of the proposed assessment framework consists of several metrics from four classes (outlier detection; measures of location; measures of scale; and comparison methods). As mentioned above, the decision matrix includes integer numbers in the interval [6,7,8,9] without any outliers, decimal numbers and zero. Therefore, metrics for determining outliers are useless and we omitted these class of metrics for this case study. For the rest of the metrics, we focused on the comparison metrics namely Minkowski distances (Euclidean distance), Standard Deviation (STD), Mean Ks values (from Spearman/Pearson Correlation), Ranking consistency Index (RCI), and Mean squared error (MSE). For details about calculating Euclidean distance and STD please see [19] and for calculation of Mean Ks value and RCI please see [7].

For calculating MSE for aggregated data sets, it is obtained from the average of mean squared error for each normalization technique with other normalization techniques using the ranking of alternatives [21, 22]. It should be noticed that the favorable result for MSE is the lower value the better because it is desirable to have less error when comparing different normalization techniques [8]. For the rest of the metrics (Euclidean distance, STD, RCI, and Mean Ks) the higher values are better and more desirable [8]. The results of MSE are depicted in Table 4.

Table 4. MSE for the borrowed case study from [10] for TOPSIS method

Implementing different metrics from the third level of the proposed assessment framework as Euclidean distance, STD, RCI, and Mean Ks plus the results of MSE are provided comparable results (Table 5).

Table 5. Results of different metrics of assessment framework using TOPSIS method for the borrowed case study from [10].

Table 6 shows the ordering of normalization techniques with respect to each used metrics. The different metrics produced different ranks for the chosen techniques and still recommending the most appropriate normalization technique is impossible just by looking at the results. Thus, plurality voting (PV) is used to sum up the obtained results of metrics [8]. PV counts the number of times that a normalization technique being the first ranking/ordering with respect to the applied metrics. So, the normalization technique that has the large number of times being the first ranked would be recommended as a more appropriate technique [8].

Table 6. Ordering of normalization techniques with respect to the metrics and using plurality voting

From the obtained results (Table 6) using the proposed assessment framework and plurality voting, the Fuzzification (Gaussian) normalization technique is the best one due to having the large number of times being the first order considering the implemented metrics, while the approach by [10], recommended the Sum normalization technique.

Comparing the results of both approaches, we believe our framework provides more robust and reliable results than the ones obtained by [10], because we cover a wide range of metrics (STD, Euclidean distance, Mean Ks, RCI, and MSE). On the other hand, implementing PV enabled us to aggregate the obtained results from different used metrics. In this case study, the authors of [10] just calculated time and space complexity with MATLAB, and these results are highly dependent on the style of MATLAB users/programmers. For instance, someone can code Sum normalization technique in a manner that obtains time and space complexity twice higher than someone else. Therefore, our proposed framework ensures more accurate and reliable results to support decision makers.

5 Conclusions

This paper addressed the applicability and robustness of the assessment framework to evaluate different normalization techniques and recommend the most appropriate technique for the TOPSIS method. A case study borrowed from [10] was used to analyze the effects of five chosen normalization techniques namely Max, Max-Min, Sum, Vector, and Fuzzification (Gaussian) and recommended the Fuzzification (Gaussian) as the best technique to be used in the related case study. Moreover, the obtained results were compared with the initial results by the authors of [10] and showed that our results are more accurate and robust than the initial results due to the implementation of several metrics from the assessment framework.

There are several normalization techniques introduced in the literature [1] which are not evaluated yet. In future we plan to evaluate further normalization techniques for different MCDM methods (e.g. PROMETHEE, MOORA, COPRAS, and etc.) using the proposed framework. Also, the validation of the assessment framework has not been done with real-world case studies, which is planned as future work.