Keywords

1 Introduction

CC has transformed the operational landscape for businesses by offering convenient and immediate access to a diverse array of computational resources for example servers, storage, and applications. This innovation empowers businesses to store and process data remotely, eliminating the requirement for extensive physical infrastructure and dedicated IT personnel [1]. CC offers numerous benefits, including scalability, cost-effectiveness and flexibility. Moreover, the exponential rise of CC has also brought about several challenges, one of which is host underload.

Host underload denotes to a situation in cloud computing where a physical server or virtual machine (VM) is operating with a workload significantly lower than its capacity. It means that the resources allocated to the host are not fully utilized, resulting in inefficient resource allocation and potential wastage as show in Fig. 1. Underload hosts may consume unnecessary power, leading to increased energy costs and decreased operational efficiency [1, 2]. Addressing host underload involves strategies such as load balancing, dynamic resource allocation, and workload consolidation to optimize resource utilization, improve performance, and achieve cost-efficiency in cloud computing environments.

Fig. 1.
figure 1

Underutilized Host.

To solve this problem we proposed the algorithm uses linear regression technique for underload host detection in cloud computing. Underload host detection can significantly impact the SLA, Energy, and Performance. Linear regression is a widely utilized machine learning methodology i.e. commonly employed to forecast the value of a variable based on another variable [3]. Within this particular framework, the variable that is to be anticipated is commonly referred to as the dependent variable (y), whereas the variable employed for the purpose of prediction is denoted as the independent variable (x).

2 Literature Review

This section presents different strategies to overcome the underload host problem, reduce energy consumption and SLA violation and improve performance.

Youssef Saadi proposed an algorithm that focuses on energy efficiency for the consolidation of VM in cloud data centers. The aim is to minimize energy consumption while considering the utilization of hosts and ensuring the data center operates at optimal throughput. The authors compare their suggested strategy with onset algorithms, namely IQR and LR, and demonstrate its hopped-up [1]. The scheme successfully meets SLA requirements, as indicated by the simulation results.

Minarolli proposed a methodology for local resource allocation that involves modifying the CPU allocation assigned to virtual machines (VMs) in response to the current workload. Additionally, global resource allocation is achieved by VM migration, which aims to balance the distribution of resources among hosts that are either overloaded or underloaded. In order to forecast resource utilization, the authors utilize Gaussian processes as a machine learning methodology for time series prediction, taking into account long-term tendencies [2].

Abbas Horri, examines the trade-off that exists between energy usage and SLA Violation inside cloud systems. The objective of cloud service providers is to enhance their revenue by employing energy-efficient resource management strategies, which include the consolidation of virtual machines (VMs) and converting inactive servers into sleep modes. Nevertheless, inadequate consolidation may lead to the development of Sleep-Related Arousal Variants (SLAV). The author proposed the implementation of consolidation algorithms that aim to optimize energy usage while simultaneously minimizing Service Level Agreement Violations (SLAVs), in order to achieve a harmonious equilibrium between these two aims [3]. The simulation results indicate that the proposed methods effectively reduce the quantity of virtual machine migrations, SLAV (Service Level Agreement Violations), and total transferred data in comparison to existing strategies.

Nimisha introduced a novel algorithm, known as the Host Utilization Aware (HUA) Techniques, which aims to detect underloaded hosts. The algorithm presented in this study aims to estimate the upper limit of hosts that can be made available by taking into account the overall utilization of the data center. The primary emphasis of the Author’s work is on the crucial stage of identifying underloaded hosts throughout the process of workload consolidation. The algorithm predicts the maximum number of hosts that may be made available by taking into account the overall utilization of the data center. The experimental results illustrate the effectiveness of the HUA Algorithm in accurately identifying hosts with low workload and thus freeing up a larger number of hosts. This leads to a reduction in energy consumption while still ensuring compliance with Service Level Agreement (SLA) requirements [4].

Weichao Ding revolves around the optimization of resource allocation and utilization in Cloud data centers. The overarching objective is to achieve a reduction in energy consumption while simultaneously upholding a high degree of compliance with service-level agreements (SLAs). The suggested framework by the author introduces a novel approach to dynamic virtual machine (VM) consolidation [5]. This approach is based on the prediction of resource use and the PPR (performance-to-power ratio) of heterogeneous hosts. The framework has four distinct stages, including host overload detection, VM selection for migration, host underload detection, and VM allocation with a modified power-aware best-fit reducing method. The proposed methodology effectively reconciles the trade-off between energy usage and performance. The suggested strategy has been validated for its effectiveness and scalability through experimental evaluations.

Bernardi Pranggono examines the matter of VM consolidation in cloud data centers and presents a novel approach for classifying host load inside an energy-performance VMC framework. The primary aim is to decrease energy usage while simultaneously guaranteeing the fulfillment of quality of service (QoS) criteria. The proposed approach entails the classification of hosts experiencing underload into three distinct states: underloaded, normal, and critical. This classification is achieved by the utilization of an underload detection algorithm. Additionally, the study presents the introduction of overload detection as well as the VM selection strategies. The overload detection policy, referred to as Mean (Mn), utilizes the mean to forecast the upper threshold [6]. On the other hand, the VM selection policy, known as Maximum Requested Bandwidth (MBW), relies on the maximum requested bandwidth.

Nirmal Kr. Biswas revolves around the resolution of two key issues in Smart Cities: the optimization of energy consumption as well as the mitigation of SLA violation. To tackle these challenges, Biswas proposes the utilization of Cloud of Things (CoT) technology. The growing processing capabilities inside cloud computing have necessitated the identification of an optimal balance between energy consumption and service level agreement violations (SLAV). The proposed methodology put forth by the author involves the integration of a novel New Linear Regression prediction model, host underload/overload detection as well as a VM placement policy. The objective of this technique is to effectively mitigate both energy consumption and service level agreement violations. The new linear regression prediction model has been developed with the purpose of forecasting forthcoming CPU use by employing a linear regression line and a mean point (MNP) [7]. The proposed algorithms are assessed using an expanded version of the CloudSim Simulator.

The suggested techniques are evaluated using CloudSim simulation with different types of random workloads as well as PlanetLab real. MadMCHD algorithm shows significant improvements compared to commonly used algorithms such as thr, mad, iqr, lr, and lrr, in minimizing SLA violation rates and VM migrations. The combination of MadMCHD and MPABFD algorithms further reduces SLA violations overall.

3 Proposed Work

In this section, we first briefly introduce Linear Regression, then show the detail of our suggested Linear Regression algorithm for underload host detection. With the help of this algorithm, we decrease energy consumption and improve the performance matrix.

3.1 Linear Regression

Linear regression is a fundamental technique in the field of machine learning, commonly employed for the purpose of predicting the value of a variable based on the value of another variable. The variable to be predicted is commonly referred to as the dependent variable (y). The variable employed for predicting the value of other variables is commonly referred to as the independent variable (x). The objective of the linear regression algorithm is to determine the optimal values for B0 and B1 in order to identify the most suitable line of best fit. The best fit line is characterized by its ability to minimize the error between projected values and actual values.

To calculate linear regression best fit line.

$$ Yi\, = \,\beta 0\, + \,\beta 1Xi $$
(1)

where Xi = Independent variable, Yi = Dependent variable, β1 = Slope/Intercept, β0 = constant/Intercept (Fig. 2).

.

Fig. 2.
figure 2

Linear Regression.

Here’s a step-by-step explanation of how to use linear regression for dynamic underload prediction.

  • Data Collection

  • Data Preprocessing

  • Split the Data

  • Model Building

  • Model Training

  • Prediction

  • Monitoring and Updating

3.2 Energy Efficient Linear Regression Algorithm for Underload Host Detection

The proposed Energy efficient linear regression algorithm for the host underload detection can be formulated. In the proposed algorithm, we predict the upcoming load on the host on behave of the previous host load history and set the minimum threshold value. The host is considered underutilized whenever the host load value is down from the minimum threshold value [9]. This proposed algorithm is helpful in improving energy efficiency and also improving performance.

3.3 Proposed Algorithm

  1. 1.

    Begin method is Host Under Utilized(host)

  2. 2.

    Cast host to Power Host Utilization History and assign it to _host

  3. 3.

    Get utilization History from _host

  4. 4.

    Set length to 10 (adjustable parameter)

  5. 5.

    If utilization History length is less than length,

  6. 6.

    Return host.get Utilization Of Cpu()

  7. 7.

    Create an array utilization History Reversed with length elements

  8. 8.

    Reverse the utilization History and store it in utilization History Reversed

  9. 9.

    Declare estimates as an array and assign the result of getParameterEstimates(utilizationHistoryReversed)

  10. 10.

    Calculate migrationIntervals as ceil (getMaximumVmMigrationTime(_host) / getSchedulingInterval()).

  11. 11.

    Calculate predictedUtilization as (estimates[0] + estimates[1] * (length + migrationIntervals)) * getSafetyParameter()

  12. 12.

    Add a history entry with host and predictedUtilization

  13. 13.

    Return predictedUtilization

  14. 14.

    End method is HostUnderUtilized.

In this algorithm, we take the three-parameter for enhancing performance. The First parameter is SLA violation second parameter is Energy, and the last parameter is No of VM migration.

$$ {\text{PM}}\, = \,{\text{E }}*{\text{ SLA }}*{\text{ No Of VM Migration}} $$
(2)

where PM is Performance matrix, E is Energy, SLA is Service level agreement.

This formula is used for calculating overall performance, and after calculating overall performance, we quickly identify which technique gives the best result (Fig. 3).

Fig. 3.
figure 3

Workflow Diagram of the Proposed Algorithm.

4 Result

4.1 Simulation Setup

The presented technique is implemented using the CloudSim toolkit 3.0.3 simulator. CloudSim is specifically designed to simulate various components of a cloud system, including virtual machines (VMs) and data centers. It offers support for VM selection and allocation policies, power models, and diverse workload types. The hardware and software requirements for implementing the proposed method are outlined as follows.

Hardware Details. The hardware used in the implementation of the proposed method.

  • Processor: Intel(R) Core(TM) i3-6100U CPU @ 2.30 GHz

  • RAM: 8 G.B

  • Hard disk: 2 TB

Software Details.

  • Operating system: Windows 10 Pro 64-bit Version

  • IDE: IntelliJ IDEA Edition 2021.3

  • Programming Language used: JAVA 17.0.1

4.2 Performance Evaluations

To simulate the proposed method we have used the cloudsim. The performance of proposed technique is measured in the term of the SLA violation, energy consumption, number VM migration. To check the effectiveness and correctness of the presented method we have done comparison with the some existing algorithm like IQRMC, IQRMMT, IQRMU, LRMMT, LRMU, MadMc, MadMMT, MadMu.

Analysis of SLA Voilation.

SLA violation occurs when a service provider fails to meet the agreed-upon performance levels or standards specified in the SLA. Analyzing SLA violations is essential to identify the root causes, assess the impact on stakeholders, and take appropriate measures to prevent future occurrences [10, 11]. The Fig. 4 demonstrate the value of SLA violation and finding clearly shows that our proposed method outperformed the existing techniques IQRMC, IQRMMT, IQRMU, LRMMT, LRMU, MadMc, MadMMT, MadMu.

Analysis of Energy Consumption.

It refers to the amount of electrical power consumed by data centers and associated infrastructure to support the operation of cloud services. Cloud computing relies on large-scale data centers that house numerous servers, storage systems, networking equipment, and cooling systems, all of which consume significant amounts of energy [12]. Analyzing energy consumption in cloud computing is crucial for improving energy efficiency, reducing environmental impact, and optimizing resource utilization.

Energy consumption of data center is determine with the given formula.

$$ \left( {{\text{EC}}} \right) \, = E_{1} \, + E_{1} \, + - E_{n} $$
(3)

where \({E}_{i}\)= energy consumption of host. \({E}_{i}\) Can be calculate using linear interpolation method based on given utilization.

Figure 5 demonstrate the results of the energy consumption of presented approach compare with the other existing techniques and results clearly indicates that presented technique perform better in terms of energy saving.

Fig. 4.
figure 4

Analysis SLA Violation.

Fig. 5.
figure 5

Analysis of Energy Consumption.

Analysis of VM Migration. VM migration is a process of moving a running or idle virtual machine instance from one physical host or data center to another. This is done for various reasons, including load balancing, resource optimization, hardware maintenance, and disaster recovery. The number of VM metrics can be calculated by utilizing the below formula:

$$ \text{VM}\,\text{Mirgration}\left( {\text{F,t1,t2}} \right) = \sum\nolimits_{i = 1}^{N} {\int_{t1}^{t2} {Mig_{i} \left( F \right)} } $$
(3)

where, N is count of VM, \(Mi{g}_{i}(F)\) is migration count of host i in the time interval t1 to t2.

The Fig. 6, clearly shows that the proposed method outperformed the existing techniques.

Fig. 6.
figure 6

Analysis of VM Migrations.

5 Conclusion

A dynamic underload host detection approach has been proposed using linear regression. The objective of the presented approach is to minimize consumption of energy, SLA violation and the number of VM migrations. With the increasing demand of cloud computing the energy consumption is also rising sharply. Hence, energy saving becomes prime concern in cloud because it also helps in reduction of cost, reduction in carbon emission and it also increase the performance of the cloud. The efficiency and accuracy of the presented technique is evaluated in terms of SLA violations, consumption of energy, and the number of migrations as performance metrics. The experiments are conducted over cloudsim tool. The performance of the presented technique is compared with the standard existing techniques like IQMRC, IQRMMT, IQRMU, LRMMT, LRMU, MADMC, MADMMT, MADMU. The results shows that the presented approach outperforms existing techniques in terms of consumption of energy, SLA violation as well as number of VM migrations. In the future, we can leverage real-time data for host underload prediction to improve the accuracy and efficacy of the proposed algorithm in real-time scenarios.