1 Introduction

Cloud computing has been gaining much attention by a wide variety of business and academic organizations due to its overwhelming benefits such as broad network access, measured service, resource pooling, on-demand service, and rapid elasticity (Sosinsky 2010). Being an open standard model, cloud provides ‘XaaS’—anything as a service (software, platform, infrastructure, storage, security, data, database, etc.) to its users through various deployment models (private cloud, community cloud, public cloud, and hybrid cloud) (Sosinsky 2010; Mell and Grance 2011). The giants of IT (Amazon, IBM, Google, Rackspace, etc.) have become the pioneers of cloud service offerings in almost all cloud service delivery models (XaaS), while the rest have begun to migrate their business applications to cloud, in order to bring down the infrastructure and operational costs (Ding et al. 2014a).

This scenario has resulted in two major research challenges with respect to the selection of appropriate and trustworthy cloud service providers which can be stated as,

1.1 Problem 1 How to identify the trustworthy cloud service providers?

The phenomenal increase in the number of CSPs offering functionally similar cloud services at different feature, cost, and quality makes the selection of trustworthy CSPs, a challenging task (Thampi and Bhargava 2013). The application of trust assessment mechanism to the cloud service selection problem helps the user to find the trustworthy service provider (Somu et al. 2017b). Trustworthiness is a most important comprehensive quality metric of the cloud service, which can be assessed either in an objective or subjective manner with respect to the several trust measure parameters (TMPs) such as usability, assurance, security, availability, performance, and cost (Marudhadevi et al. 2014; Qu 2016; Tang et al. 2017). The importance of trust metric and their relation to the quality of the cloud services has resulted in the development of several trust-based service evaluation and service selection models for the identification of suitable and trustworthy CSPs (Liao et al. 2007; Ding et al. 2014a, b, Sun et al. 2014, 2016; Wang et al. 2015; Qu 2016; Ma et al. 2016). However, these models evaluate the trustworthiness of the CSPs based on a pre-defined set of TMPs which affects the accuracy and complexity of the service selection model.

1.2 Problem 2 What are the major requirements to build an efficient trust-based cloud service selection model?

In general, a trust-based cloud service selection model identifies the trustworthy CSPs using various trust evaluation mechanisms based on objective (QoS values) and subjective (user feedbacks) assessment data. An efficient cloud service selection model should be adaptive to the dynamic nature of cloud services which further complicates the process of trust-based service evaluation. For example, it is hard to evaluate the trustworthiness of the existing cloud service and a new cloud service due to the context (time and location)-based performance variation and lack of assessment data, respectively (Qu et al. 2015). Trust prediction, a classification, and a multi-criteria decision-making (MCDM) problem is one plausible solution for the above-said challenge through the accurate trust evaluation of cloud service providers, especially when the trustworthiness of new CSPs needs to be evaluated (Akshya Kaveri et al. 2017; Obulaporam et al. 2018). The significance of trust prediction models in service-oriented environments has resulted in several research contributions in web service selection, cloud service selection, recommender systems, pervasive environments, and social networks (Ding et al. 2017b; Mao et al. 2017; Su et al. 2017).

According to Vimal Kumar et al. ‘A good TMP would be easily measurable from the service provider’s side and would be contextually related to the service the end user wants to use on the cloud’ (Thampi and Bhargava 2013). For example, let us consider a cloud ecosystem where \( 'n' \) cloud service users need to find a suitable and trustworthy cloud service provider from \( 'm' \) functionally similar cloud service providers. In a real-world scenario, the severity of the cloud service selection gets complicated with the ever-growing nature of the cloud user requests and the cloud service providers. In general, the performance assessment of each cloud service provider is based on the historical QoS and feedback information provided by the historical users across the QoS attributes. Further, the importance of the QoS attributes varies with respect to the cloud service type and user preferences. Hence, a generic solution to the above-said challenges (Problems 1 and 2) can be presented through the design of an efficient feature selection technique which can identify the optimal TMP reduct with respect to the service type requested by the users or their own preferences (Somu et al. 2017b). It also addresses ‘Lack of an advanced multi-criteria-based measurement of user preferences,’ an open issue in the present cloud service selection models and approaches (Sun et al. 2014). To summarize, the design of an efficient feature selection technique enhances the accuracy of the cloud service selection model in terms of reduced complexity (minimizes the search space) and better service selection (recommends the service providers that have maximum adherence with the users’ QoS requirements).

Feature selection is found to be one of the most challenging problems in the field of engineering science which aims at the identification of the minimal subset of features that have same information content as that of the original information system (Chen et al. 2015, 2017). In specific, feature selection techniques have their significance in cloud service selection problem through the identification of optimal service-specific TMPs, thereby reducing the complexity and enhancing the accuracy of the cloud service selection model (Somu et al. 2017b). Among the various computational intelligent models like principle component analysis (PCA), independent component analysis (ICA), instance-based feature selection, correlation-based feature selection, information gain-based feature selection, Bayesian feature selection, etc. rough set theory (RST) is found to be the major choice of researchers due to its capability of handling imprecise, uncertain and vague data (Pawlak 1982b; Chen et al. 2015; Raza and Qamar 2018; Zouache and Ben Abdelaziz 2018). In general, RST for feature selection operates over the positive region based on attribute dependency to find the subset of original features (reducts) from the given dataset. However, the dependency computation of RST is found to be computationally extensive since it involves the generation of equivalence class using both conditional and decision attribute followed by the generation of the positive region (Raza and Qamar 2018). One simple solution toward this problem is the generation of all possible reducts and choosing the one which has a higher dependency on the decision attribute and minimal cardinality. However, this approach is not applicable for the dataset with a large number of features and samples since the time incurred for the generation of all possible subset is in the order of exponential powers (Jiang et al. 2015). Several research works toward the integration of RST with the greedy approaches like hill climbing, forward selection, backward elimination, conditional entropy-based significant estimation, discernibility matrix-based methods, etc., failed to guarantee an optimal solution within a reasonable time (Chen et al. 2015; Das et al. 2018).

In general, finding all possible subset is an NP-hard problem (Gheyas and Smith 2010). Recent research contributions in the development of an efficient feature selection technique have revealed the significance of meta-heuristic techniques like genetic algorithm, particle swarm optimization, cuckoo search optimization, gray wolf optimization, firefly optimization, etc., in solving NP-hard problems (Table 1) (Gheyas and Smith 2010; Raman et al. 2016; Somu et al. 2016; Gauthama Raman et al. 2017a). Even though these meta-heuristic techniques guarantee the global optimal solution, they suffer from high time and computational complexity issues due to the distribution of computing and memory ability of various degrees to each individual in the generations (Chen et al. 2015).

Table 1 Related works

Fruit fly optimization algorithm (FFOA) is a new, simple, computationally effective swarm intelligent evolutionary algorithm proposed by Pan 2012 which is inspired by the food foraging behavior of fruit flies. Due to the inherent ability of FFOA in providing faster convergence and guaranteeing global optimal solution, it has been successfully applied for parameter optimization, combinatorial optimization and function estimation in various domains like image processing, bio-informatics, stock market analysis, etc. (Pan 2012; Dai et al. 2014; Shen et al. 2016). Hence, in this work, we have introduced a binary variant of FFOA (BFFO) for identifying the minimal rough set reducts. The major contribution in this work is as follows:

  1. 1.

    RST-HGBFFO exploits the hyperclique property of hypergraph to enhance the performance of BFFO through minimal time complexity and improved convergence rate.

  2. 2.

    HGBFFO integrated with RST (supervised quick reduct (SQR) and quick relative reduct (QRR)) to enhance the performance of RST for the identification of the optimal reduct in minimal time. This work is the extension of our previous work, where we have proposed a novel trust assessment model (TrustCom) and an efficient feature selection technique based on RST and hypergraph properties (Somu et al. 2017b).

  3. 3.

    This work uses Cloud Service Measurement Index Consortium—Service Measurement Index (CSMIC—SMI) metric as a standard for TMP (Valley 2011; Garg et al. 2013; Somu et al. 2017a). Further, chi-square test was applied to the considered datasets to assess the independent nature of the TMPs (Moore 1976).

  4. 4.

    Experiments were conducted using three different datasets, namely quality web service (QWS) dataset, Cloud Armor trust feedback dataset, and Centre for Information Super Highway (CISH)—SASTRA synthetic trust feedback dataset. The performance of RST-HGBFFO was validated using hypergraph-based computational model (HGCM) and WEKA tool in terms of reduct size, service ranking, classification accuracy, and time complexity (Øhrn 2001; Somu et al. 2017a).

The paper is structured as follows: Sect. 2 gives a brief introduction to the rough set theory, hypergraph, and fruit fly optimization. Section 3 discusses the proposed RST-HGBFFO feature selection technique. Section 4 highlights the experimentation and performance analysis of RST-HGBFFO over the state-of-the-art techniques based on the reduct size, service ranking, classification accuracy, and runtime analysis. Section 5 concludes the paper.

2 Materials and methods

2.1 Rough set theory

Rough set theory is an intelligent mathematical tool, proposed by Pawlak (1982b) for handling imprecise, uncertain and vague dataset. Unlike traditional soft computing techniques like fuzzy sets, Dempster-Shafer theory, etc., RST does not require any additional or prior information about the dataset. Therefore, RST has been extensively used in a wide range of applications for knowledge and data discovery. The fundamental assumptions in RST are as follows: (i) Objects in the universe of discourse are represented by the attributes with some information associated with them, and (ii) objects that consist of similar information are indiscernible. This section discusses a few basic concepts and definition of RST which are usually employed for the identification of the optimal reduct (Pawlak 1982a, b; Abraham 2009).

The data analysis process in the rough set theory (RST) initiates from the Information System \( \left( {I_{\text{S}} } \right) \) represented in the form of a table where rows correspond to the samples and columns correspond to the attributes (Table 2a). Mathematically, an information system can be viewed as four-tuple vector \( I_{\text{S}} = \left( {U,A,V,f} \right), \) where \( U \) is the universe with non-empty set of objects, \( A \) is the non-empty set of attributes, \( V \) is the union of attributes such that \( V = \mathop {\bigcup }\nolimits_{a \in A} V_{a} \) and \( f:U \times A \to V_{a} \) is the information function such that \( a \in A \) and \( x \in U, f\left( {x,a} \right) \in V_{a} \). For every \( M \) subset of \( A \), there exists an associated equivalence relation called indiscernibility relation as defined in Eq. (1) (Min and Liu 2009; Pawlak 1982a; Yao 2004).

$$ {\text{IND}}_{M} = \{ \left( {x,y} \right) \in U^{2} |\forall a \in M, f\left( {x,a} \right) = f\left( {y,a} \right)\} $$
(1)
Table 2 Information system (Is) and decision system (Ds)

The family of all equivalent classes of \( {\text{IND}}_{M} \) is denoted either by \( U/{\text{IND}}_{M} \) or \( U/M \) and an object \( x \) contained in the equivalent class of \( {\text{IND}}_{M} \) is represented as \( \left[ x \right]_{M} \). Basically, the class of attributes in \( I_{\text{s}} \) can be distinguished into conditional \( \left( C \right) \) and decisional \( \left( D \right) \) attributes \( \left( {A = C \cup D} \right), \) where such case is referred as decision table \( \left( {D_{\text{s}} } \right) \) (Table 2b). For any \( M \subseteq C \cup D \) and \( X \in U \), the \( M \)-lower approximation and \( M \)-upper approximation are defined in Eqs. (2) and (3) (Pawlak 1982a).

$$ Y_{M} = \Leftarrow \left\{ {\left[ x \right]_{M} \in U/{\text{IND}}\left( M \right):\left[ x \right]_{M} \subseteq X} \right\} $$
(2)
$$ Y^{M} = {\bigcup }\left\{ {\left[ x \right]_{M} \in U/{\text{IND}}\left( M \right):\left[ x \right]_{M} \cap X \ne 0} \right\} $$
(3)

The \( M - \) boundary region of set \( Y \) is defined as \( {\text{Bnd}}_{M} = Y_{M} - Y^{M} \). Generally, the roughness of the given decision table \( \left( {D_{T} } \right) \) of set \( Y \) is computed using Eq. (4).

$$ Y_{\delta } = \frac{{\left| {{\text{Bnd}}_{M} } \right|}}{{\left| {Y^{M} } \right|}} $$
(4)

The measure of roughness is found to be an important metric in RST since it reflects the degree of uncertainty of an approximation set. The value of \( Y_{\delta } \) lies in the range of [0,1], where 0 implies every information in the underlying set \( Y \) is known and 1 implies lack of information regarding the underlying set \( Y \). The M-positive region \( {\text{PoS}}\left( D \right) \) of \( U/{\text{IND}}\left( D \right) \) with respect to \( Y \) is defined as a set of all objects in \( U \) that can be certainly classified to the classes of \( U/{\text{IND}}\left( D \right) \) by the means of \( M \) (Eq. 5)(Pawlak 1982a).

$$ {\text{PoS}}_{M} \left( D \right) = \mathop {\bigcup }\limits_{{Y \in U/{\text{IND}}\left( D \right)}} Y_{M} $$
(5)

The degree of dependency among the set of attributes \( \left( M \right) \) and the decisional attribute \( \left( D \right) \) can be computed through the positive region using Eq. (6).

$$ \gamma_{M} \left( D \right) = \left| {{\text{PoS}}_{B} \left( D \right)} \right|/\left| U \right| $$
(6)

Obviously, \( \gamma_{M} \left( D \right) \in \left\{ {0,1} \right\}, \) where 0 represents \( M \) is independent of \( D; \) 1 represents \( M \) is completely dependent on \( D; \), and \( 0 < \gamma_{M} \left( D \right) < 1 \) represents the partial dependency between \( M \) and \( D \). The main objective of RST is the removal of redundant and irrelevant features, and therefore, the reduced set of attributes (reducts) consists of the same information content as that of the original dataset. Mathematically, the set of all reducts can be denoted as in Eq. (7).

$$ R\left( C \right) = \left\{ {R \subseteq C |\gamma_{R} \left( D \right)} \right\} = \gamma_{C} \left( D \right)\forall M \subseteq R,\gamma_{R} \left( D \right) \ne \gamma_{C} \left( D \right) $$
(7)

Since the dataset consists of multiple reducts, the optimal reduct is defined as in Eq. (8).

$$ R_{\text{Opt}} = \{ R^{\prime} \in R|\forall R^{\prime\prime} \in R,\left| {R^{\prime}} \right| < \left| {R^{\prime\prime}} \right|\} $$
(8)

2.2 Hypergraph

Hypergraph is a generalization of the traditional graph theory in which the edges (hyperedges) correspond to the collection of two or more vertices. Mathematically, hypergraph \( \left( {H_{G} } \right) \) is represented as an order 2-tuple \( H_{G} = \left\{ {V,E} \right\}, \) where \( V = \left\{ {V_{1} ,V_{2} , \ldots ,V_{n} } \right\} \) is the finite set of vertices and \( E = \left\{ {e_{1} ,e_{2} , \ldots ,e_{m} } \right\} \) corresponds to the hyperedges which is a non-empty subset of \( V, \) such that \( E_{i} \ne \emptyset \) and \( {\bigcup }E_{i} = V ; \forall i = \left( {1,2, \ldots ,n} \right) \). In recent years, the integration of hypergraph with various computational models has proven its efficiency in many applications in terms of minimal complexity since they are modeled based on the algebraic and duality concepts, i.e., geometry and topology structure (Raman et al. 2016, 2017, Somu et al. 2016, 2017a, b, 2018a, b; Gauthama Raman et al. 2017a, b). In this subsection, we discuss few basic concepts and definitions of hypergraph and clique property to improve the performance of binary fruit fly optimization algorithm (Berge 1973).

Definition 1

Hypergraph Let \( H_{G} = \left\{ {V,E} \right\} \), be a hypergraph, then the representative graph \( \left( {R_{G} = \left\{ {V_{r} ,E_{r} } \right\}} \right) \) of hypergraph is defined as follows (Fig. 1):

Fig. 1
figure 1

\( H = \left\{ {V,E} \right\};V = \left\{ {v_{1} ,v_{2} ,v_{3} ,v_{4} ,v_{5} ,v_{6} } \right\};e = \left\{ {e_{1} ,e_{2} ,e_{3} ,e_{4} } \right\};e_{1} = \left\{ {v_{1} ,v_{2} } \right\};e_{2} = \left\{ {v_{2} ,v_{3} } \right\};e_{3} = \left\{ {v_{3} ,v_{4} ,v_{5} } \right\};e_{4} = \left\{ {v_{4} ,v_{5} } \right\} \)

  1. (a)

    \( V_{r} = n\, or\, V_{r} = E \), when hypergraph \( H \) has no repeated hyperedges

  2. (b)

    \( \left\{ {a,b} \right\} \in E|a \ne b \), if and only if \( e_{a} \cap e_{b} \ne \emptyset \)

Definition 2

Neighborhood hypergraph Let \( R_{G} = \left\{ {V_{r} ,E_{r} } \right\} \) be a graph and \( m \subseteq V_{r} \). Then, the closed neighbor \( L \) in \( R_{G} \) is defined in Eq. (9).

$$ C_{L} \left( m \right) = \left\{ {L \in V_{r} |L \,{\text{is}} \,{\text{adjacent}}\, {\text{of}}\, m \,{\text{in}} R_{G} } \right\} \cup \left\{ m \right\} $$
(9)

Definition 3

Complete graph: A graph \( \left( {R_{G} = \left\{ {V_{r} ,E_{r} } \right\}} \right) \) is said to be complete, if there exists edge \( e \in E_{r} \) between the vertices \( \left( {V_{i} ,V_{j} } \right) \in V_{r} \forall i,j = \left\{ {1,2, \ldots ,n^{\prime}} \right\}|i \ne j \) where \( n^{,} \) is the number of vertices in \( R_{G} \).

Definition 4

\( K \)-clique Given a graph \( R_{G} = \left\{ {V_{r} ,E_{r} } \right\} \), \( K - \) clique of \( R_{G} \) is a subset \( D \subset V_{r} \) and \( K = \left| D \right| \), such that \( D \) obeys definition (Fig. 2).

Fig. 2
figure 2

a 3-clique, b 4-clique

Proposition 1

Consider a\( R_{G} = \left\{ {V_{r} ,E_{r} } \right\} \)and clique\( D \subset V_{r} \), then\( D \)is the clique of\( R_{G} \), if and only if\( \mathop {\bigcap }\nolimits_{i = 1}^{m} D_{L} \left( {V_{i} } \right) \).

Proof

Let \( D \) be a clique of graph \( R_{G} \) and \( q \subseteq V_{r} \), then \( q \in D_{L} \left( {V_{i} } \right), \forall i = \left( {1,2, \ldots ,m} \right) \& D \subset \mathop {\bigcap }\nolimits_{i = 1}^{m} D_{L} \left( {V_{i} } \right). \) Conversely, consider \( D \subset \mathop {\bigcap }\nolimits_{i = 1}^{m} D_{L} \left( {V_{i} } \right) \). Let \( g \) and \( h \) be two distinct elements of \( D \), then from the hypothesis \( q \in D_{L} \left( {V_{i} } \right), \forall i = \left( {1,2, \ldots ,m} \right) \), \( g \) and \( h \) are adjacent. Hence, \( D \) is the clique of \( R_{G} \) (Fig. 3).

Fig. 3
figure 3

a A graph (GRep) with 6 vertices and 8 edges; b, c, and d possible cliques of GRep is D1={VR1, VR2, VR5}, D2={VR2, VR3, VR4,VR5}, and D3={VR6, VR5, VR4}

2.3 Fruit fly optimization algorithm

Fruit fly optimization algorithm (FFOA) is a novel meta-heuristic technique based on the food foraging behavior of fruit flies to identify the global best solution for optimization problems (Pan 2012; Dai et al. 2014; Shen et al. 2016). The fruit flies have a superior sense of smell (osphresis) and perception (vision) compared to other species. The osphresis organ of a fruit fly can collect all kinds of smell in the environment. It can smell the food source at a distance of 40 km. Once a fly reaches a location nearer to the food source, it makes use of its sensitive vision and the company’s flocking location to locate the food. Figure 4 depicts the food foraging behavior of the fruit flies (particles) in the search space. Initially, the particles begin to traverse the search space in search for an optimum solution in a randomized manner. During each iteration, the fitness of each particle was assessed using the fitness evaluation function in terms of particle’s current location. The smell concentration of each particle represents their fitness value. On to the subsequent iterations, the particles fly based on the location of the particle with maximum fitness value. Based on the food foraging behavior of the fruit flies, the steps in FFOA can be categorized as follows:

Fig. 4
figure 4

Fruit fly’s food foraging behavior

Step 1 The major parameters of the FFOA, such as maximum number of generations \( \left( {Max_{Gen} } \right) \), population size \( \left( {PoP_{Size} } \right) \), number of population \( \left( {PoP_{Num} } \right) \), fitness value \( \left( {Fitness} \right) \), local best smell concentration \( \left( {Best_{Smell} } \right) \), global best smell concentration \( \left( {G_{Best_{Smell}} } \right) \), best position \( \left( {Best_{Pos} } \right) \), and random position of the particles were initialized with appropriate values as defined in Eqs. (10) and (11).

$$ X_{\text{Axis}} = {\text{Rand}}() $$
(10)
$$ Y_{\text{Axis}} = {\text{Rand(}}) $$
(11)

Step 2 The particles were stimulated to construct the population and traverse the search space using the osphresis gland for its food foraging process using Eqs. (12) and (13).

$$ X_{i} = X_{\text{Axis}} + {\text{Random}}\, {\text{valu}}e $$
(12)
$$ Y_{i} = Y_{\text{Axis}} + {\text{Random }}\,{\text{Value}} $$
(13)

Step 3 For each iteration, the fitness of each particle was assessed by substituting its current position in the fitness evaluation function (Eqs. (14) and (15)).

$$ S_{i} = \frac{1}{{D_{i} }}; D_{i} = \sqrt {X_{i}^{2} + Y_{i}^{2} } $$
(14)
$$ Smell_{i} = Fitness\left( {S_{i} } \right) $$
(15)

Step 4 The smell concentration and position coordinates of the fittest particle (a particle with maximum fitness value) were retained. For subsequent iterations, the particles traverse the search space with respect to the position of the fittest particle (Eqs. 1619).

$$ Best_{Smell} = Max\left( {Smell_{i} } \right) $$
(16)
$$ \left( {G_{Best_{Smell}} } \right) = Best_{Smell} $$
(17)
$$ X_{\text{Axis}} = X\left( {Best \;index} \right) $$
(18)
$$ Y_{\text{Axis}} = Y\left( {Best\; index} \right) $$
(19)

Step 5 Steps 2 and 4 were repeated in an iterative manner for a maximum number of generations or until an optimal solution is attained.

3 RST-HGBFFO: the proposed trust measure parameter selection technique

Feature selection or dimensionality reduction is a primary and indispensable phase in many data analytic applications. It is the process of finding the minimal set of attributes or reducts that reveals the maximum information content of the original dataset (Deogun et al. 1998). The major objective behind the design of any feature selection technique is to increase the accuracy and reduce the computation burden of the learning model. In general, feature selection techniques evolve from the decision system \( \left( {D_{S} } \right) \) which consists of \( 'S' \) observations \( O = \left\{ {O_{1} , O_{2} , \ldots , O_{S} } \right\} \), \( 'A' \) conditional attributes \( A = \left\{ {A_{1} , A_{2} , \ldots , A_{C} } \right\} \), and a decisional attribute \( \left( D \right) \) (Table 3) (Jia-Yuarn 2003; Hu 2012). Since all the conditional attributes in the decision table may not be important at all times, the importance of these attributes can be determined by the measure of relevancy and redundancy (Nina 2007). An attribute or feature is relevant, when it is highly predictive of the decisional attributes \( \left( D \right) \), and is redundant, when it is highly correlated with the neighboring conditional attributes \( \left( D \right) \). Hence, an optimal reduct must consist of features that are independent on each other \( \left( A \right) \) and closely related to the decisional attribute \( \left( D \right) \). Further, the performance of any learning model can be improved in terms of enhanced accuracy and reduced complexity due to minimized search space through the identification of optimal reducts (Jensen and Shen 2003).

Table 3 Decision system

The prominence of RST over other statistical and machine learning techniques has motivated the researchers and academicians toward the exploitation of RST for various research problems in medical diagnosis (Somu et al. 2016), intrusion detection systems (Raman et al. 2016; Gauthama Raman et al. 2017a), cloud service selection (Liu et al. 2016; Somu et al. 2017a), etc. A more generic way to obtain the reducts using RST is to generate all possible combinations of the feature subsets and identify the subset that maximizes the dependency function (Zhong et al. 2001; Jensen and Shen 2004; Nina 2007; Wang et al. 2007). However, the above-said approach holds good for small volume datasets, while in the case of massive and high-dimensional datasets, the complexity of the learning model increases exponentially (Somu et al. 2016). Therefore, for large-scale datasets, SQR and QRR were used to obtain the reducts (Velayutham and Thangavel 2011; Chen et al. 2015). The execution of SQR begins with an empty set, followed by the iterative addition of substantial features that maximize the dependency function. The major problem with this approach is the lack of global optima (traps at local minima) due to the instability in the dependency metric experienced by the iterative inclusion and exclusion of features. To avoid this, QRR replaces the dependency metric of RST by relative dependency function (Inbarani et al. 2014).

This work presents RST-HGBFFO, a cooperative bio-inspired algorithm which hybridizes the benefits of hypergraph and binary fruit fly optimization technique with RST (SQR and QRR) for the identification of optimal service–specific trust measure parameters or conditional attributes with minimum time complexity (Algorithm 1). The main objective of the proposed approach is to identify the optimal reducts, i.e., service-specific trust measure parameters thereby minimizing the computational and time complexity of the cloud service selection model. RST-HGBFFO operates over a \( m \)-dimensional binary search space with \( PoP_{Num} \) fruit flies or particles which are represented by a \( 'm' \) tuple vector (particle’s location). Each tuple vector varies in its length with respect to the number of features \( \left( A \right) \). For example, a four-tuple vector \( \left\langle {0 0 1 1} \right\rangle \) corresponds to the presence of \( A_{3} , A_{4} \) and the absence of \( A_{1} ,A_{2} \) (Table 4). Figure 5 presents a generic data flow of RST-HGBFFO.

Table 4 n-tuple vector and feature subset
Fig. 5
figure 5

Rough set theory-based hypergraph-binary fruit fly optimization trust measure parameter selection technique

figure a
figure b

The detailed explanation of the working of RST-HGBFFO is given as follows:

Phase 1Initialization phase It initializes the population size \( \left( {PoP_{Size} } \right) \), maximum generation \( \left( {Max_{Gen} } \right) \), fitness function \( \left( {Fitness} \right) \), local best smell concentration \( \left( {Best_{Smell} } \right) \), global best smell concentration \( \left( {G_{Best_{Smell}} } \right) \), best position \( \left( {Best_{PoS} } \right) \), and number of particles \( \left( {PoP_{Num} } \right) \).

Phase 2Generation of initial population In general, the position of each particle in the traditional BFFO was initialized in a random fashion. Similarly, the initial population of RST-HGBFFO was obtained through a step-by-step procedure: (i) construct a two-dimensional matrix \( \left[ M \right]_{{PoP_{Num} \times PoP_{Size} }} \) with randomly generated binary string (Fig. 6(a)), (ii) construct a hypergraph \( \left( {H_{G} } \right) \) using \( \left[ M \right]_{{PoP_{Num} \times PoP_{Size} }} \)(Fig. 6b), and (iii) application of hyperclique property on \( H_{G} \) (Definition 4) to obtain the refined population (Gauthama Raman et al. 2017b) (Fig. 6c).

Fig. 6
figure 6

Generation of initial population: a random population, b construction of hypergraph and application of hyperclique property, c initial population

Phase 3Fitness evaluation Each initial set of particles or population obtained from phase 1 was evaluated using the fitness function defined based on the attribute dependency measure of RST. The fitness evaluation function of RST-HGBFFO was designed with respect to the dependency metric \( \left( {Fitness_{SQR} } \right) \) and relative dependency metric \( \left( {Fitness_{QRR} } \right) \) of RST given in Eqs. (20) and (21), respectively.

$$ Fitness_{SQR} = \gamma_{R} \left( D \right) = \frac{{\left| {PoS_{R} \left( D \right)} \right|}}{\left| U \right|} $$
(20)
$$ Fitness_{SRR} = \gamma_{R} \left( D \right) = \frac{{\left| {U/IND\left( R \right)} \right|}}{{\left| {U/IND\left( {R \cup D} \right)} \right|}} $$
(21)

Phase 4Updation Once the fitness of each population was computed using the fitness evaluation function [Eqs. (20) and (21)], the value of the \( Smell \) function \( \left( {Best_{Smell} } \right) \) was updated with the fitness value of the fittest particle, i.e., the population with high fitness value (dependency metric). For each iteration, the global best smell value \( \left( {G_{Best_{Smell}} } \right) \) was updated based on the local best smell \( \left( {Best_{Smell} } \right) \), i.e., the smell concentration value of the \( \left( {G_{Best_{Smell}} } \right) \) and \( Best_{Smell} \) was compared against each other. Further, the \( G_{{Best_{Smell} }} \) is updated with the high smell concentration value \( (Best_{Smell} > G_{{Best_{Smell} }} ;G_{{Best_{Smell} }} \leftarrow Best_{Smell} ) \). The position of the fittest particle is stored in \( Best_{PoS} \) and set as a seed for the subsequent iteration.

Phase 5Generation of new population Check for the termination condition (optimal solution \( \left( {Fitness \cong 1} \right) \) or \( Max_{Gen} \)). If the optimal solution was found, return the \( Best_{PoS} \), i.e., the optimal TMP. Else, repeat phase 3 and 4.

3.1 Example

Let us consider a decision table (Table 5) with 20 observations represented in the form of seven conditional attributes \( \left( {A_{1} , A_{2} , \ldots ,A_{7} } \right) \) and a decisional attribute \( \left( D \right). \) The execution of RST-HGBFFO TMP selection technique begins with the initialization of \( Max_{Gen} = 3; PoP_{Size} = 7;PoP_{Num} = 20, Fitness = 0;Best_{Smell} = 0; G_{Best_{Smell}} = 0 \) and \( Best_{PoS} = \left\langle {0,0,0,0,0,0,0} \right\rangle \) (Table 6). The hyperclique property of hypergraph was exploited for the generation of initial population of BFFO (Fig. 7). The particles fly over the search space from their initial random position in search for the optimal reduct based on their sense of smell and vision. Over a certain period, the particles get displaced to a new position \( \left( {P_{1} = 0, 1, 0, 1,1,1,0; P_{2} = 1, 0, 1, 0,1,1,1;P_{3} = 1, 0, 1, 1,0,1,1;P_{4} = \left\langle {0,0,1,1,1,1} \right\rangle } \right) \), where \( 0 \) and \( 1 \) represent the absence and presence of the feature, respectively. The fitness of each particle was evaluated based on the particle’s current location using the fitness evaluation function of SQR and QRR given in Algorithm 1 \( \left( {P_{1} = 0.34; P_{2} = 0.60;P_{3} = 0.21;P_{4} = 0.53} \right) \) (Table 7).

Table 5 Decision table: an example
Table 6 Initialization phase
Fig. 7
figure 7

Construction of hypergraph and application of hyperclique property

Table 7 Iteration 1—optimal fitness value and reduct

The smell concentration value of the particle with high fitness value was stored in \(Best_{Smell} \) and compared against \( G_{Best_{Smell}} \left( {G_{Best_{Smell}} = 0} \right) \). The comparison resulted in the replacement of the smell concentration of \( G_{Best_{Smell}} \) with higher smell concentration value \( \left( {Best_{Smell} = 0.60; G_{Best_{Smell}} = 0; Best_{Smell} > G_{Best_{Smell}} ; G_{Best_{Smell}} = 0.60} \right) \). The position of the particle with high fitness value was stored in \( Best_{Pos} \left( {Best_{Pos} = 1, 0, 1,0,1,1, 1} \right) \) (Table 8).

Table 8 Fitness and position update

In the subsequent iteration, the random position of the particles was initialized based on \( Best_{Pos} = 1, 0, 1,0,1,1, 1 \). The displaced position of the particles \( P_{11} , P_{12} ,P_{13} \) and \( P_{14} \) after time interval \( t \) was \( 1,1,1, 0,1,1,1, \)\( 1,0,0,1,1,1,0, \left\langle {1,1,1,1,0,0,1} \right\rangle \), and \( 0,1,1,0,1, 1, 1 \) (Table 9). The fitness value of each particle based on their current position was computed as \( P_{11} = 0.17, P_{12} = 1, P_{13} = 0.21 \), and \( P_{14} = 0.52 \) (Table 9). An optimality check on the fitness value resulted in the attainment of the optimal solution \( \left( {P_{13} = 1,0,0,1,1,1,0 = 1} \right) \). The smell concentration value of \( G_{Best_{Smell}} \) was updated with the optimal fitness value \( \left( {Best_{Smell} = 1; G_{Best_{Smell}} = 0.60; Best_{Smell} > G_{Best_{Smell}} ; G_{Best_{Smell}} = 1} \right) \). The position of the fittest particle stored in \( Best_{Pos} \)\( \left( {Best_{Pos} = 1,0,0,1,1,1,0} \right) \) represents the optimal reduct \( \left( {TMP} \right) \), which can be decoded as \( A_{1} ,A_{4} ,A_{5} ,A_{6} \).

Table 9 Iteration 1—optimal fitness value and reduct

4 Experimental results and discussion

The implementation of RST-HGBFFO was carried out using Python 3.6 (Scikit-rough sets package) on an Intel® Core™ i5 processor @ 2.40 GHz system running Windows 7 OS with 16 GB RAM. The performance evaluation of RST-HGBFFO was validated using two benchmark datasets, namely QWS dataset, and Cloud Armor trust feedback dataset & a synthetic trust feedback dataset obtained from the Centre for Information Super Highway (CISH), SASTRA Deemed University (Sheng; Al-Masri and Mahmoud 2007).

4.1 Dataset description

  1. (i)

    QWS dataset QWS is a public QoS dataset owned and maintained by E. Al-Masri and Q.H Mahmoud, University of Guelph, Canada (Al-Masri and Mahmoud 2007). This QoS dataset is widely used in several research works and evaluation studies on QoS in service-oriented environments. It contains QoS records of 365 real-world web services which were collected using Web Service Crawler Engine (WSCE) developed by Al-Masri and Mahmoud. These services were tested over a period of 10 min for three continuous days. The trust rate or service classification of each service was assessed using the web service relevancy function (WsRF) with respect to various quality metrics like availability, response time, security, etc. (Table 10). The web services were classified into four categories, namely platinum (high quality), gold, silver, and bronze (low quality) based on the rating provided by WsRF. The service offering qualities of the web services were represented in numerical form (1 to 4).

    Table 10 Trust measure parameters in QWS dataset (\( TMP_{1} - TMP_{12} : \) conditional attributes; service classification: decisional attribute)
  2. (ii)

    Cloud Armor trust feedback dataset It was obtained from Cloud Armor, a research project at University of Adelaide which deals with the development of a robust and scalable trust management model for cloud environments. The Cloud Armor trust feedback dataset comprises 10,080 QoS feedbacks given by approximately 7,000 customers over various time stamps (9 years) for 114 real-world cloud services (Table 11).

    Table 11 Trust measure parameters in Cloud Armor and synthetic trust feedback dataset (\( TMP_{1} - TMP_{10} : \) conditional attributes; trust result: decisional attribute)
  3. (iii)

    Synthetic trust feedback dataset at CISH, SASTRA University It comprises the trust feedbacks obtained from the students for a diverse set of cloud services (academic and student) provided by the private cloud at CISH, SASTRA University. This dataset consists of users’ feedbacks in the range of \( \left[ {0, 5} \right] \) for 12 trust measure parameters (availability, security, response time, etc.) and corresponding trust result (Table 11). Since the significance of the TMPs varies with respect to the service type requested by the users and their preferences, the cloud services in the considered datasets were grouped based on their service type. The TMPs or quality service attributes given in Tables 10 and 11 represent the KPIs in the CSMIC-SMI metrics (Valley 2011; Somu et al. 2017b).

RST-HGBFFO employs CSMIC-SMI as a standard metric for the QoS attributes or TMPs (Valley 2011). The CSMIC-SMI metric is generally viewed as a hierarchy of categories, attributes, and sub-attributes or key performance indicators (KPIs). At the first level of the CSMIC-SMI hierarchy, the entire metric space is divided into seven categories. Further, each category is partitioned into four or more attributes. At the extremity, the attributes are further divided into one or more sub-attributes or KPIs, whose actual count varies with the nature of cloud business solutions. The CSMIC-SMI hierarchy reveals the high-dimensional nature of the cloud objective (monitoring QoS value) and subjective (user feedbacks) assessment data with respect to the QoS attributes and QoS records (samples). However, due to the unavailability of appropriate benchmark datasets, we have used two benchmark datasets (QWS dataset and Cloud Armor synthetic feedback dataset) and one synthetic dataset (CISH—SASTRA trust feedback dataset) to prove the efficiency of RST-HGBFFO over the state-of-the-art rough set-based feature selection techniques. Further, the relation between the QoS attributes in the considered datasets and CSMIC-SMI metrics is given in Tables 10 and 11.

4.2 Results and discussions

The experiments carried out can be divided into three phases, namely data preprocessing, generation of reducts, and performance validation. At the initial phase, the considered datasets were processed using appropriate data preprocessing techniques and the training and testing datasets were generated in 80:20 ratio. In the subsequent phase, the reducts were obtained from the existing and the proposed (RST-HGBFFO) feature selection techniques. Finally, the performance of RST-HGBFFO was evaluated using Weka tool and HGCM in terms of reduct size, service ranking, classification accuracy, and runtime analysis (Øhrn 2001; Somu et al. 2017a).

4.2.1 Phase 1—data preprocessing

At the initial stage of the experiment, each sample in the considered dataset was transformed into a compatible format supported by the classifiers. Further, data normalization was used to minimize the impact of the features with high value. Each feature in the sample is normalized in such a way that all the values lie in the range of [0,1] [Eq. (22)].

$$ f_{ab} = \frac{{f_{ab} }}{{\mathop \sum \nolimits_{b = 1}^{n} f_{ab} }}; \forall a = \left( {1,2, \ldots ,S} \right);b = \left( {1,2, \ldots ,n} \right) $$
(22)

where \( S \) and \( n \) are total number of samples and features in the dataset, respectively.

The discrete and continuous values in the synthetic trust feedback datasets were converted into a compatible format by data mapping technique, i.e., mapping the discrete and continuous values to numerical values ranging from 1 to \( A_{C} \), where \( A_{C} \) represents the total number of conditional attributes. Further, the considered dataset was subjected to chi-square test with \( \left( {r - 1, c - 1} \right) \) degrees of freedom, where \( r \) and \( c \) represent the number of rows and columns, respectively, to check the degree of dependency among the QoS attributes. During the experiments, it was noted that a minimum of 80% of the TMPs were found to be dependent on each other, even at a level of 10% significance. Hence, the inter-relation between the QoS attributes has to be exploited for identification of the optimal feature subset. Finally, the training and testing datasets for each considered dataset were generated in 80:20 ratio.

4.2.2 Phase 2: generation of reducts

In this phase, the optimal TMP reduct was generated for RST-HGBFFO and the state-of-the-art feature selection techniques. The reducts for the existing rough set theory-based feature selection techniques like RSARSubsetEval_Genetic Search \( \left( {RST_{GS} } \right) \), RSARSubsetEval_PSO Search \( \left( {RST_{PSO} } \right) \), RSARSubsetEval_Cuckoo Search \( \left( {RST_{CU} } \right) \), RSARSubsetEval_ANT Search \( \left( {RST_{AN} } \right) \), RSARSubsetEval_Harmony Search \( \left( {RST_{HS} } \right) \), RSARSubsetEval_Firefly Search \( \left( {RST_{FF} } \right) \), and RSARSubsetEval_Bat Search \( \left( {RST_{BS} } \right) \) were generated by the integrating attribute evaluator (RSARSubsetEval) with the search techniques (PSO Search, Cuckoo Search, Genetic Search, etc.) (Witten et al. 2016). In addition, we have used few state-of-the-art classifiers like Bayes Net, SVM, K Star, BF Tree, J48, and Random forest in WEKA tool to validate the reducts obtained from the proposed and the existing rough set-based feature selection techniques in terms of classification accuracy. As mentioned earlier, the identification of all possible subsets in a high-dimensional and massive dataset is an NP-hard problem. Further, from the extensive survey, it is evident that meta-heuristic techniques outperform the statistical and graph-based techniques in solving NP-hard problems. Hence, the comparative analysis of RST-HGBFFO (SQR and QRR) was carried out using the state-of-the-art RST-based feature selection techniques rather than our previous work, rough set hypergraph-based feature selection technique (RSHT) (Somu et al. 2017b). The critical parameter in RST-HGBFFO was set as \( Max_{Gen} = 100 \), and \( PoP_{Num} \) varies from 10 to 50 in the scale of 5; similarly, the default parameter values were used for the existing rough set-based feature selection techniques.

4.2.3 Phase 3: performance validation

The performance of RST-HGBFFO over the other RST-based feature selection techniques was evaluated using Weka tool and HGCM in terms of various quality metrics described as follows:

  1. (i)

    Reduct size

Figures 8 and 9 present the impact of the number of population \( \left( {Num_{Pop} } \right) \) on the fitness value of RST-HGBFFO (SQR and QRR) for the considered datasets. From Figs. 8 and 9, it can be observed that RST-HGBFFO achieves better dependency metric, i.e., the optimal fitness value (0.999) at \( Num_{Pop} = 50 \), i.e., \( Best_{Pos} \), the position of the fittest particle at \( Num_{Pop} = 50 \) is presented as optimal reducts for RST-HGBFFO in Figs. 10, 11, and 12.

Fig. 8
figure 8

Impact of population on the fitness value—SQR

Fig. 9
figure 9

Impact of population on the fitness value—QRR

Fig. 10
figure 10

Validation—reduct size (QWS dataset)

Fig. 11
figure 11

Validation—reduct size (Cloud Armor trust feedback dataset)

Fig. 12
figure 12

Validation—reduct size (synthetic trust feedback dataset—CISH, SASTRA)

The reducts obtained from the existing and proposed (RST-HGBFFO) feature selection techniques (phase 2) were evaluated based on the reduct size. Based on the experiments, the following observations were made,

  1. 1.

    For QWS dataset, the optimal reduct provided by RST-HGBFFO (SQR and QRR) was similar to that of \( RST_{PSO} , RST_{AN} , RST_{HS} , \), and \( RST_{FF} \) (Fig. 10).

  2. 2.

    For Cloud Armor trust feedback dataset, the optimal reduct provided by RST-HGBFFO (SQR and QRR) was similar to that of \( RST_{CU} , RST_{HS} , \), and \( RST_{BS} \) (Fig. 11).

  3. 3.

    For CISH-SASTRA trust feedback dataset, the optimal reduct provided by RST-HGBFFO (SQR and QRR) was similar to that of \( RST_{PSO} \) and \( RST_{CU} \) (Fig. 12).

  4. (ii)

    Classification accuracy

For further validation, the reducts obtained from the existing and proposed (RST-HGBFFO) feature selection techniques (phase 2) were evaluated based on the classification accuracy. Based on the experimental evaluations, it can be observed that the available classifiers (Bayes Net, SVM, K Star, BF Tree, J48, and Random Forest) in Weka tool exhibited high classification accuracy when trained with the optimal reducts obtained from RST-HGBFFO (SQR and QRR) for the considered datasets (Figs. 13, 14, and 15). A noteworthy point is that RST-HGBFFO has similar performance to \( RST_{CU} \) when SVM was trained with the reducts obtained from QWS dataset.

Fig. 13
figure 13

Validation—classification accuracy (QWS dataset)

Fig. 14
figure 14

Validation—classification accuracy (Cloud Armor trust feedback dataset)

Fig. 15
figure 15

Validation—classification accuracy (CISH—SASTRA trust feedback dataset)

The validation based on the reduct size, i.e., the cardinality of the reduct, shows that the performance of RST-HGBFFO was similar to the state-of-the-art rough set-based feature selection techniques (Figs. 10, 11, and 12). The similarity in the cardinality of the feature subset does not imply the optimality of the features in the optimal feature subset. Even though RST-HGBFFO and few existing RST-based feature selection techniques have similar reduct size, the reducts obtained from RST-HGBFFO provide the optimal reduct that maximize the fitness function, i.e., dependency metric. Further, from Figs. 13, 14, and 15, it was evident that the classifiers exhibit high classification accuracy when trained with the optimal feature subset obtained from RST-HGBFFO.

  1. (iii)

    Service ranking

The optimal TMP reduct obtained from phase 2 was fed as input to the hypergraph-based computational model (Fig. 5). HGCM ranks the CSPs based on the adherence between the CSPs’ provision and users’ unique requirements for various CSMIC-SMI KPIs. For validation purposes, we have considered case study 2, which comprises 5 CSPs provisions (Amazon EC2 \( \left( {S_{1} } \right) \), Windows Azure \( \left( {S_{2} } \right) \), Rackspace \( \left( {S_{3} } \right) \), cloud services from private cloud set up at SASTRA University \( \left( {S_{4} {\text{and }}S_{5} } \right) \)) and one user requirement. HGCM processes the CSP provisions and user’s requirements for the entire set of CSMIC-SMI metric and ranks the cloud service providers as \( S_{3} , S_{4} , S_{5} ,S_{1} , \) and \( S_{2} \). During the validation phase, HGCM processes the QoS data of CSP provisions and CU’s requirements based on the optimal reduct obtained from phase 2 and ranks the CSPs as \( S_{3} , S_{4} , S_{5} ,S_{1} , \) and \( S_{2} \). From Table 12, it was more clear that the service ranking of RST-HGBFFO (SQR and QRR) was similar to \( RST_{GS} , RST_{PSO} , RST_{AN} , \) and \( RST_{FF} \). However, RST-HGBFFO outperforms \( RST_{GS} , RST_{PSO} , RST_{AN} , \) and \( RST_{FF} \) in terms of minimal reduct and run time.

Table 12 Validation of RST-HGBFFO TMP selection technique—service ranking
  1. (iv)

    Run time analysis

In addition to above quality metrics, runtime analysis has its own significance in evaluating the performance of the feature selection technique. With respect to feature selection technique, time complexity can be redefined as the total time taken by the technique to identify the optimal reduct. From the experiments carried out, the following can be observed:

  1. 1.

    For QWS dataset, the run time of RST-HGBFFO (SQR and QRR) was slightly higher than \( RST_{CU} , RST_{AN} , \), and \( RST_{BS} \) (Fig. 8).

  2. 2.

    For Cloud Armor trust feedback dataset, the run time of RST-HGBFFO (SQR and QRR) was similar to \( RST_{BS} \), but slightly higher than \( RST_{FF} \) (Fig. 9).

  3. 3.

    For CISH-SASTRA trust feedback dataset, the run time of RST-HGBFFO (SQR and QRR) to find the optimal subset was found to be minimal (Fig. 10).

To summarize, even though the reduct size of RST-HGBFFO (SQR and QRR) was found to be similar to some of the state-of-the-art feature selection techniques (RSARSubsetEval_PSO Search \( \left( {RST_{PSO} } \right) \), RSARSubsetEval_Cuckoo Search \( \left( {RST_{CU} } \right) \), RSARSubsetEval_ANT Search \( \left( {RST_{AN} } \right) \), RSARSubsetEval_Firefly Search \( \left( {RST_{FF} } \right) \), RSARSubsetEval_Harmony Search \( \left( {RST_{HS} } \right) \), and RSARSubsetEval_Bat Search \( \left( {RST_{BS} } \right) \)), it substantiates its efficiency in terms of classification accuracy and run time analysis. Even though the run time of RSARSubsetEval_Firefly Search \( \left( {RST_{FF} } \right) \), RSARSubsetEval_ANT Search \( \left( {RST_{AN} } \right) \), RSARSubsetEval_Cuckoo Search \( \left( {RST_{CU} } \right) \), and RSARSubsetEval_Bat Search \( \left( {RST_{BS} } \right) \) was found to be minimal, RST-HGBFFO outperforms the state-of-the-art rough set-based feature selection techniques in terms of classification accuracy and reduct size (Fig. 10-12).

The overwhelming performance of RST-HGBFFO was due to the search strategy of HGBFFO which involves the application of hyperclique property of hypergraph for the optimal number of 1’s in the population of BFFO. Further, the search strategy of HGBFFO optimizes the relevance between the features (TMPs) and the target class (trust value) and reduces the redundancy among the selected TMPs which resulted in an optimal feature subset that maximizes the fitness function, i.e., dependency metric.

5 Conclusions

The immense benefits provided by the cloud computing ecosystem have enticed diverse organizations to become a part of this phenomenal technology. This scenario has led to the substantial increase in the number of cloud service providers offering a wide variety of cloud services, which makes it difficult for the users to find suitable and trustworthy CSPs who comply with their QoS requirements. Tireless efforts by the research community have yielded trust-based service selection models as a prominent solution for the above-said challenge. However, the intrinsic relation between the accuracy of the trust-based service selection models and the optimality of the service-specific TMP subset makes the identification of the optimal service-specific TMP subset an open research challenge.

Hence, this paper put forth an efficient rough set theory-based hypergraph-binary fruit fly optimization (RST-HGBFFO), a cooperative bio-inspired technique for the identification of the optimal TMPs based on the service type and user preferences. The hyperclique property of hypergraph was exploited for the generation of initial population to enhance the performance of the binary fruit fly optimization algorithm in terms of reduced time complexity and by avoiding premature convergence. Further, HGBFFO was integrated with rough set theory for the identification of optimal TMP with minimal time complexity, i.e., the fitness of each particle of HGBFFO was evaluated based on the RST-specific fitness function. The CSMIC-SMI metric was used as a standard for TMPs. The experiments were conducted using QWS dataset, Cloud Armor and synthetic trust feedback dataset obtained from the private cloud at CISH, SASTRA University. The performance of RST-HGBFFO over the state-of-the-art feature selection technique was evaluated using HGCM and Weka tool in terms of reduct size, service ranking, classification accuracy, and time complexity. Further, RST-HGBFFO was found to be computationally attractive, scalable, and applicable in the field of intrusion detection, stock market analysis, service ranking, weather forecasting, DNA sequencing, metadata validation in geographical information system (GIS), etc.