1 Introduction

To advertise an item or service in a particular region, a business may use LBA (location-based advertising) [1, 2]. The term "location-based advertising," or "LBA" for short, describes a specific category of location-based services that make use of geolocation data gathered from mobile devices like smartphones. LBA might potentially be utilized to hire seasonal or full-time staff for a neighbourhood company. LBA allows merchants to send location-specific, timely communications to consumers. Advertisers may target consumers with various messages based on their physical location by using their GPS coordinates. Allowing marketers to contact consumers while they are physically close to their stores has broken down formerly insurmountable barriers between marketplaces and consumers [3].

Recommendation systems (RSS) use a variety of filtering approaches, all of which are described here. Collaborative filtering is the approach that is most often used. A large number of online service recommendation systems are being developed based on collaborative filtering (CF). The analysis of the quality of service-aware web service suggestions and customized QoS-aware web service recommendations is discussed in detail [4]. Many of the most common customer-facing tactics are based on user-based or item-based approaches. The Matrix Factorization (MF) method stands out as one of the most effective approaches [5]. It leverages hidden vector attributes to represent individuals or items, initially mapping them into a shared latent space [6]. While MF has demonstrated considerable success, numerous adaptations and variations have emerged [7, 8]. However, despite its effectiveness, MF-based strategies continue to grapple with challenges like cold-start and data sparsity problems [9, 10], significantly limiting the efficacy of current methodologies.

It is important to include the geographical data that is often a part of the customer database when clustering the data to generate customer segments, including, in response, selecting where the channels of distribution need to be positioned. In recent years, geo-marketing has seen an uptick in the practice of clustering geo-referenced (or spatial) data, although conventional clustering approaches have shown certain limits in light of the industry’s rising need for precision and consistency [11]. To maximize the effectiveness of customized targeting through the right channels, consumer segmentation must be both precise and homogeneous.

Current Challenges The existing environment of location-based advertising and recommendation systems is fraught with difficulties. The commonly used matrix factorization (MF) technique in recommendation systems encounters data sparsity and cold-start concerns. The drawbacks of this strategy impede system efficiency, particularly when dealing with sparse data or new user interactions. Furthermore, typical clustering algorithms [12, 13] are proven to have drawbacks in the realms of geo-marketing and clustering geo-referenced data for consumer segmentation. These difficulties come as a result of the industry’s increased demand for precision and consistency, making precise consumer segmentation harder to attain using traditional clustering methods. Furthermore, analytic methodologies for independent consumer data encounter difficulties when dealing with vast and complicated geographic datasets, reducing the depth and accuracy of insights derived from exploratory data analysis.

Proposed Solutions Proposed Solutions: The proposed solutions outlined in this study aim to address challenges within recommendation systems and geo-marketing by exploring various methodologies. Firstly, alternatives to Matrix Factorization and strategies to mitigate data sparsity and enhance cold-start handling are investigated [14, 15]. Additionally, the development of hybrid models merging collaborative and content-based filtering or incorporating contextual information is suggested for more robust recommendations [16,17,18]. In the realm of geo-marketing, the study proposes exploring advanced clustering techniques, potentially integrating machine learning algorithms to manage large geographic datasets effectively. Furthermore, advocating for the utilization of complex analytic approaches, such as machine learning models or deep learning methods, to extract meaningful insights from multidimensional datasets for independent customer data analysis is emphasized. Successful implementation of these recommended solutions in real-world scenarios would require collaboration among data scientists, domain experts, and industry stakeholders. This research proposes, different deep learning-based approaches that are applied to perform the prediction, and the names of these are the LSTM and CNN, with the employ of these models, a hybrid model is built to implement this research, and the name of this hybrid model is hybrid CNN-LSTM model the location-aware web service recommendation.

Customers are independent whenever it comes to exploratory data analysis, even though most commercially available GIS provides for substantial storage, editing, and display of geo-referenced data [19]. Due to the extensive size and high complexity of geographic data, as well as the need for sophisticated data integration and mining tools, this strategy is impractical. When a dataset has a lot of dimensions, it might be difficult for analytic techniques to work properly. It is a common issue that not all of the variables will have a significant relationship with one another. Most analyses assume a very basic pattern, which may be adjusted according to several criteria, and so restrict or compress the field of possible hypotheses [20]. It is recommended that the many artificially enabled location-based services (LBS) in mobile ads make use of deep learning methods such as long short-term memory (LSTM), convolutional neural networks (CNN), recurrent neural networks (RNN), and artificial neural networks (ANN) [21].

The following article structures the remaining article as follows: The background and related work on location-based advertising have been discussed in Sect. 2. Section 3 talks about the proposed work, including location-based advertising and service methods for an identified problem. In Sect. 4, we discuss the experimental results of the suggested approaches and make comparisons to other current methods; in Sect. 5, we conclude and provide directions for future study.

2 Related Work

Location-based services (LBS) is a broad and evolving field that encompasses various technologies and applications. In this section, we are giving a detailed survey on LBS. Related research on LBS is presented here.

Tan et al. [22] used a partial least squares structural method to analyze advertising on social media. In addition, an integrated framework based on interaction theory, individual variables, and the mobile technology acceptance model was used to comprehend customer preferences. Li and Xu [23] proposed a diversity-aware digital ad architecture to adequately meet their clients’ needs (D-AdFeed). The knapsack problem, with several possible solutions, was used to frame this issue. Both greedy and genetic algorithms were used to find a solution. Goh et al. [24] used Logit and Poisson count models to analyse the search patterns and ad clickthrough rates. By developing a tailored mobile advertising system, Li and Du [25] were able to provide shoppers and businesses with contextually aware advertisements.

Utilising an ensemble-based approach, Haider et al. [26] investigated fraudulent practices in mobile ads. Consumers’ familiarity with location-based advertising was measured using the persuasiveness needed for planning by Ryu and Park [27]. To examine why customers make repeat purchases, Lu et al. [28] introduced the theory of the planned model. Shin and Lin [29] studied consumer opinions and their propensity to ignore advertisements. Yang et al. [30] (2013) integrated an advertising model based on technological and emotional evaluations. Sharma et al. [31] used a neural network and partial least squares structure to increase consumers’ receptivity to mobile marketing.

According to [32] analyse the LBSs’s privacy-preserving strategies. They classify and provide an in-depth analysis of the current methods. They found the main ideas and most recent progress in several common works from each group and then analysed them in the past. For new study possibilities, we also talk about how privacy-preserving methods can be used in LBSs. This survey, which offers a current and thorough summary of previous research, may encourage further investigations into this promising field.

In [33] research examines the feasibility of using workers’ smartphones as a legitimate tool for employers to do presence control. In addition, they suggest a mobile location-aware information system that satisfies universal access requirements, uses only reactive location technologies based on terminals and allows for non-intrusive presence control while keeping costs down. Encouraging workers to feel comfortable and in control of when their location data is collected while meeting the employer’s control demands is the main emphasis. Using UAProf data processing at the origin server and A-GPS terminal-based/network-aided mobile positioning algorithms, the LAMS platform is a cutting-edge framework for synchronous mobile location-aware content personalisation.

This study [34] provides the creation of a set of fake query sequences to conceal mobile users’ query locations and query characteristics and safeguard their privacy in LBS. First, they provide a framework that is centred on the client and aims to safeguard user privacy in LBS. This framework does not need any modifications to the LBS algorithm used on the server side, and it also ensures that the accuracy of an LBS query remains unaffected. Second, they propose a privacy model within the framework to establish the criteria that ideal dummy query sequences should adhere to. These criteria include: (1) maintaining a similar feature distribution, which assesses the ability of the dummy query sequences to conceal the true user query sequence, and (2) ensuring a high level of user privacy protection, which evaluates the effectiveness of the dummy query sequences in safeguarding the location and query privacy of a mobile user. Finally, they provide an implementation approach that meets the privacy model’s requirements. Furthermore, both theoretical analysis and actual assessment reveal the usefulness of our proposed technique, demonstrating that the location and attribute privacy underlying LBS inquiries may be successfully safeguarded by the fake queries created by our approach.

The author [35] offers a wide range of scientific data for many study topics that are used by the associated scientific groups. Frequently, these websites’ architecture or design does not match the way their customers think. Consequently, the desired data is not easily accessible. Scientific online information services may benefit from the methods developed by Usability Engineering and User Experience to better understand and meet the needs of their users. An easy-to-implement method for evaluating and improving scientific online information services is detailed in this article. Combining personas, usability inspections, online questionnaires, Kano models, and web analytics with other techniques that have previously been effectively used in reality makes up this approach.

In [36] the research suggests an Explainable Food Recommendation system to support its suggestions using the visual content of food. In particular, a new similarity score is created and included in the suggestion. This score is based on a propensity measure that quantifies the degree to which the user community favours a certain cuisine category. To improve the recommendation outcome’s openness and interpretability, a rule-based explainability is finally added. Our results on a crawled dataset demonstrate that compared to other food recommendation methods, the suggested approach improves recommendation quality by 7.35%, 6.70%, 7.32%, and 14.38%, respectively, in terms of recall, precision, F1, and Normalised Dis-counted Cumulative Gain (NDCG). Additionally, they conduct ablation studies to prove that our recommendation system’s components are technically sound.

In this study, [37] the author proposes to provide a novel strategy called Recommendation Based on Embedding Spectral Clustering in Heterogeneous Networks (RESCHet). This approach makes use of the embedding spectral clustering method, whose similarity matrix is produced using a heterogeneous embedding technique. Subsequently, they used the notions of submit-paths and atomic meta-paths to identify the connections among people and things relevant to every cluster. Finally, by calculating the Hadamard product between the pertinent vectors, they produced user suggestions. Tests conducted on three publicly available benchmark datasets have shown that RESCHet performs noticeably better than the current methods.

In this study [38], a recommender system that relies on Facebook user behaviour was created, and it provides consumers with the option to purchase their favourite things in two stages. The consumers’ behaviour is examined in the first phase, and items are provided to them depending on their interests. In the second step, the recommender system uses data mining methods to provide consumers with offers that are related to their past purchases. The study’s data are accurate, and the findings are reliable. Furthermore, the findings show that the developed recommender system is quite accurate in presenting offers to consumers (Table 1).

Table 1 Review of existing models for location aware-base services

2.1 Research Gap

Certain gaps become apparent within the current corpus of research concerning location-based services and advertising. To begin with, although research has examined digital advertising architectures and customer preferences, more extensive inquiries are required to determine the efficacy and consequences of location-based advertising interventions for consumers. Present models predominantly concentrate on concerns about advertising, customer preferences, and recommendation systems. Consequently, there exists a knowledge vacuum concerning the complexities linked to location-based mobile advertisements, including the need to minimise intrusiveness while optimising openness. Furthermore, the investigation into energy-efficient and secure routing protocols for wireless sensor networks highlights the criticality of further research into the intersection of energy efficiency and security in location-based services. Moreover, a comprehensive examination of the integration of user-generated text and geographical data for spatial market segmentation is lacking. Finally, although recent research has examined privacy concerns in location-based advertising, there remains a dearth of knowledge regarding the continual creation of strategies for successfully addressing these concerns in light of the growing relationship between advertisers and mobile users. By recognising and rectifying these deficiencies, one can enhance the overall comprehension of the complexities and potentialities that lie within the dynamic domain of location-based advertising and services. In contrast to prior investigations, the present study employs GIS to acquire accurate location data by integrating geographical context into the word embedding procedure. Innovating the application of advanced deep learning techniques to location-based advertising is exemplified by the creation of a Bidirectional Optimised Hybrid Model (BLSTM-DNN-ASOA) and the implementation of a deep sparse autoencoder (DSAE) for feature extraction. The efficacy of a proposed model is demonstrated by a comprehensive set of evaluation metrics, which comprise F-measure, precision, accuracy, computational time, AUC, and recall. Through the attainment of enhanced performance compared to prior approaches, this article makes a significant contribution to the advancement of knowledge and implementation of deep learning in the domain of location-based advertising. As a result, it provides industry practitioners and academics with invaluable insights.

3 Methodology

3.1 Convolutional Neural Network (CNN)

Among the various applications for deep neural networks, image and video processing are among the most common. This is especially true with convolutional neural networks (CNNs). There has been a rise in the use of convolutional neural networks (CNNs) to analyze numerical data, including time series and sensor data, in recent years [48]. The main idea behind CNNs for numeric data is to apply the convolution operation to local temporal windows of the input data, allowing the network to learn temporal patterns and dependencies in the data. CNNs are effective at capturing both local and global patterns in time series data, making them suitable for various applications such as time series forecasting, anomaly detection, and signal processing. Despite their effectiveness, CNNs for numeric data still face some challenges, such as dealing with missing data and handling long-term dependencies. Recent advances in research have focused on developing more robust and efficient CNN architectures, such as the WaveNet and Temporal Convolutional Network (TCN) models, which have shown promising results in various applications. In addition to numeric data, CNNs have also been used for other types of data, such as text and graphs, demonstrating their versatility and potential to advance the field of deep learning. Overall, CNNs represent a powerful tool for processing various types of data, and their continued development and optimization have the potential to advance the field of artificial intelligence and its applications in various domains [49]. In Fig. 1, we see the basic structure of a convolutional neural network (CNN) that may be used to classify images.

A typical CNN will include the following layers: input, convolution, ReLU, pooling, and completely connected

  • Input The raw pixel values are stored in this layer, as the name implies. The original data for a photograph may be seen in its "raw" pixel values.

  • Convolution Being the primary processing node, this layer is an essential component of convolutional neural networks.

  • ReLU Activation-function-using layer: a layer that takes the output of the previous layer and utilizes it to activate its output (also called a rectified linear unit layer). RELU’s addition to the network’s non-linearity would take another form.

  • Pooling The pooling layer is another part of convolutional neural networks. Pooling operations refer to the act of combining the values of neighbouring features into a single one by using either an average or a histogram operator. The incorporation of pooling into the model has several goals, the primary ones being to make the model immune to local distortions and to reduce the total number of features [50].

  • Fully Connected Moreover, this layer might be referred to as the "output" layer or the fully linked layer. It is used in the determination of class score output, the outcome of which is a voluminous 1*1*L array, where L is the integer representing the class score.

Fig. 1
figure 1

Convolutional neural network (CNN)

3.2 Long Short Term Memory (LSTM)

LSTMs, like RNNs, are composed of many interconnected layers, but the interactions between these four levels occur differently. There are memory cells in the LSTM model, and gates control them. There are three distinct types of entrance gates (input gate, output gate, and forget gate). These gates, which control the flow of data like dials, are responsible for mixing the data. To alter the data in an LSTM, these gates are used. A fixed quantity of training data may be stored in the memory module. Cell state memory is the memory unit that gives LSTM the ability to recall long-term dependencies. There are three primary varieties of gates: the forget gate, the I/P gate, and the O/P gate [51]. Short-term and long-term memories are stored in different types of memory cells. It reminds me of a conveyor belt in certain ways. It permeates the whole sequence and has only minimal pointwise operations with the gates. LSTM is an excellent tool for classifying processes and making predictions based on time series over a given amount of time [52].

  • Hidden State An LSTM layer’s output, known as the hidden state, is used as input in the layer that follows it. To indicate how much of each piece should be sent, the sigmoid layer produces values between 0 and 1. The Tanh layer produces new state-enhancing vectors.

    $$\begin{aligned} h_t = o_t \cdot \tanh (c_t) \end{aligned}$$
    (1)
  • Forget Gate The data stored in a memory cell may be removed using the forget gate. If the situation changes, the forget gate will produce zeros, which the memory cell will pointwise multiply, erasing the associated data. The sigmoid layer then generates a vector with values between 0 and 1.

    $$\begin{aligned} f_t = \sigma (W_f \cdot [h_{t-1}, x_t] + b_f) \end{aligned}$$
    (2)
  • Input Gate The I/P gate of the state cell decides if the input needs to be looked at more closely to see if data needs to be entered (or changed). For example, the output from the previous iteration ot-1, the I/P tx, and the initial condition of iteration ct-1 are all examples of inputs that may need to be looked at more closely. This is put point-first into the memory cell.

    $$\begin{aligned} i_t = \sigma (W_i \cdot [h_{t-1}, x_t] + b_i) \end{aligned}$$
    (3)
  • Output Gate The final product will be an abstracted version of the current state of memory. The amount of information sent from the visible state of the cell to the hidden state is managed by the output gate. Then, we choose which features of the cell will go to the hidden layer, which is the sigmoid layer (Fig. 2). The cellular state is multiplied by the sigmoid gate’s output after a tanh-tanh transformation, producing values between -1 and 1.

    $$\begin{aligned} o_t = \sigma (W_o \cdot [h_{t-1}, x_t] + b_o) \end{aligned}$$
    (4)
Fig. 2
figure 2

Long Short Term Memory (LSTM)

4 Proposed Model

An in-depth description of the suggested model and its accompanying algorithm are presented here, together with a discussion of the research process used to develop it.

4.1 Problem Identification

Establishing a measure of service quality is the primary obstacle to service recommendation (QoS). Location-based advertising is a popular advertising strategy that targets users based on their geographical location. However, existing location-based advertising methods have limited accuracy and effectiveness due to several factors, such as inaccurate location data, a limited understanding of user behaviour, and a lack of personalized targeting. Deep learning techniques can provide a solution to these problems by leveraging complex models to analyze and interpret user data, which can be used to generate personalized and effective location-based advertising.

Therefore, the problem for location-based advertising using deep learning is to develop accurate and effective models that can leverage user data to generate personalized advertising recommendations based on their geographical location. This requires addressing the following challenges: data quality, user behaviour modeling, personalization, and privacy concerns. In solving these challenges, developing deep learning models is required to effectively analyse and interpret location data to generate personalized advertising recommendations that are relevant and effective for individual users while also addressing privacy concerns.

4.2 Research Methodology

In this study, we mix conventional and deep learning techniques to create a hybrid model for predicting the popularity of location-based services based on user recommendations. Python is used for the actual implementation. This is accomplished by using the freely available WS-Dream dataset. There are a total of 5,825 services and 19,74,675 QoS values from 339 users. Then the collected dataset was preprocessed using different data preprocessing techniques. In the data preprocessing, handling and filling in the missing values, and also using the Z-score normalization technique for data normalization. In addition, the min-max method is used for data scaling. In this work, features are identified and extracted in the feature extraction, and the Pearson correlation coefficient is also calculated in the correlation analysis. Then, split the dataset into two sets: train sets and test sets, and the ratio of training and testing is 80 and 20. After this process, different deep learning approaches are applied to perform the prediction, and the names of these are LSTM and CNN. With the use of these models, a hybrid model is built to implement this research, and the name of this hybrid model is the hybrid CNN-LSTM model. The correctness of the proposed model was evaluated using several statistical metrics, including the average absolute error, root mean squared error, coefficient of determination, etc. Finally, the proposed hybrid CNN-LSTM model may be used to forecast the QoS values of online services at indicated locations.

4.3 Hyperparameter Tunning

The design of the hybrid CNN-LSTM model for predicting location-based service popularity in this research included the use of many crucial hyperparameters. The architecture of the model was determined by setting the amount of LSTM units and CNN filters to 64, which influenced its ability to collect both temporal and spatial data efficiently. The learning rate of 0.001, an important hyperparameter, was carefully chosen to manage the step size of the optimization process, which influences the model’s convergence pace. Another critical parameter, the dropout rate of 0.3, was used to prevent overfitting, with its value controlling the proportion of neurons randomly destroyed during training. To balance training speed and memory limitations, the batch size of 64, which represents several samples processed in every iteration, was determined. A number of epochs, which represents iterations throughout the full training dataset, was a critical hyperparameter influencing training length and model convergence. Non-linearity was introduced by carefully selecting activation functions in the layers, such as ReLU for CNN and sigmoid or tanh for LSTM. The optimizer (Adam) and the loss function (typically Mean Squared Error (MSE) or others adapted to the prediction task) further influenced the model’s performance. These hyperparameters worked together to fine-tune the model’s prediction capabilities, with optimization carried out through extensive experimentation and validation on training and validation datasets.

4.4 Dataset Collection and Description

For this research, the WS-Dream dataset is utilized to assess how well the suggested method works. QoS values from several users across multiple services may be found in the WS-Dream dataset, a massive web services dataset. To evaluate the efficacy of our proposed algorithm, the dataset’s inclusion of location data for both users and services makes it a particularly suitable candidate. The WS-Dream dataset includes 19,74,675 quality of service metrics, collected from 339 users over 5,825 services. There is a wide range of possible quality of service metrics, including throughput, response time, reliability, and availability. QoS data was shown as a user-service matrix, with people in the rows and services in the columns, to make it easier to analyze (Table 2).

Table 2 Data description

4.5 Data Preprocessing

The suggested method would benefit from the addition of the Autonomous System Number (ASN). The utilized dataset included customers from 30 countries and 136 ASNs and services from 73 countries and 990 ASNs. To guarantee the algorithm’s success, we preprocessed the data in question. It included dealing with missing values, which may arise from some causes, including improper data input or inadequate data gathering. Imputation was used to manage missing values, which entails replacing them with approximated values based on the existing data. In this research, we imputed missing values using the attribute’s mean value using a process called mean imputation. Including CN and ASN attributes into the algorithm required using Sklearn’s category encoding transforms the labeled features into numerical embeddings. This allowed us to express each classified attribute as a nation code, which can be used to generate personalized ads for each user based on their location. Furthermore, we also performed data normalization to ensure that all the attributes had the same scale, which is important for many machine learning algorithms to function correctly. Z-score normalization, whereby the numbers of every feature are adjusted precisely to have a mean of zero and a standard deviation of one, was also utilized. Finally, we also performed feature scaling to ensure that all the features had similar ranges, which is necessary for some algorithms to function properly. All features are also scaled to fall inside the range [0, 1] using the Min-Max scaler approach. Overall, these pre-processing steps helped us to ensure that the data was in a suitable format for machine learning algorithms to be applied and improved the accuracy and effectiveness of the proposed model.

4.6 Feature Extraction

What we mean by "feature extraction" is the transformation of unstructured data into discrete characteristics that may be used in subsequent analyses without losing any of the original data’s context. In this study, we use feature extraction methods to zero in on the data points that will be most useful to our model. We employed a method called correlation analysis, which entails determining how closely each characteristic is linked to the desired outcome (here, the user’s choice for a certain service). The model places a premium on features that have a strong correlation with the dependent variable. The Pearson correlation coefficient between each characteristic and the outcome variable was determined for examination of correlation. The Pearson correlation coefficient, which may also take on ranges around -1 and 1, is used to measure the linear connection between two variables. There is a perfect positive correlation when the value is 1, no connection at all when the value is 0, and a perfect negative correlation when the value is -1.

Based on the correlation analysis, we identified the most relevant features for this proposed model. These features were then used as input variables for our algorithm, and with careful feature selection, we were able to boost the model’s accuracy and performance. Overall, feature extraction techniques such as correlation analysis can be useful in identifying the most important features of a machine learning model, which can improve its accuracy and efficiency.

4.7 Data Splitting

Machine learning often makes use of data splitting to separate data into distinct groups. There should ideally be three distinct sets of information: "train," "validation," and "test.". Dividing data is advised to improve the amount of training data in each dataset. In training and testing, data is often divided 80:20 or 70:30. For moderate-sized datasets, dividing the data into three equal parts–70% training, 20% validation, and 10% testing–is suggested. For this study, we created a training set with 80% of the data and a testing set with 20%. (20% of total information).

4.8 Proposed Hybrid CNN-LSTM Model

Here, we present a hybrid approach that utilizes LSTM and CNNs, and we evaluate it using the WS-Dream dataset. Quality of Service (QoS) measurements from many users across multiple providers may be found in the WS-Dream dataset. Due to the inclusion of user and service locations, this dataset is ideal for testing the efficacy of the suggested approach. The proposed CNN-LSTM model takes in user ratings of web service quality from the WS-Dream dataset. Quality-of-service (QoS) indicators are used to train a convolutional neural network (CNN) to extract high-level spatial features from an input dataset. The LSTM receives the CNN’s output and learns to successively process the features while capturing their temporal relationships. For the LSTM part, we employed cells that can remember and pass along information about their surroundings over very long periods. The WS-Dream dataset was used to teach the hybrid model, which has 19,74,675 QoS values across 339 users and 5,825 services. Response time, throughput, availability, and dependability are just a few examples of the many QoS variables that may be measured. This article describes the deep learning models and their architectures that were utilized to produce this hybrid model (Fig. 3).

Fig. 3
figure 3

Structure of Proposed Hybrid CNN-LSTM model

5 Experimental Results and Analysis

Here, we report on our experiments with Python-based location-based services applied to the WS-Dream dataset. The success of the proposed method may be measured in two ways.

5.1 Evaluation Metrics

RMSE and MAE are common statistical assessment metrics used to determine how well the provided method performs.

  1. 1.

    Mean Absolute Error (MAE) While calculating the mean absolute error, the errors are not squared. The absolute value of the mistakes is determined and then averaged. We only care about the magnitude of the difference between the estimated and real target values; hence, the MAE uses the absolute value. This prevents the MAE from being inaccurately calculated due to mistakes cancelling each other out. The formula for RMSE is as follows Eq. 5.

    $$\begin{aligned} MAE = \frac{1}{n} \sum _{i=1}^{n} |y_i - x_i| \end{aligned}$$
    (5)

    In this formula: MAE represents the Mean Absolute Error. n: is the number of data points or observations. \(y_i\): represents the actual or observed values. \(x_i\): represents the predicted or estimated values for the corresponding observations.

  2. 2.

    Root Mean Squared Error (RMSE) The Root Mean Squared Error (RMSE) is calculated by averaging the squared errors over all samples and then squaring the result. This is similar to the Mean Squared Error (MSE) but with a square root instead of a plus sign. This allows RMSE to produce an error metric that is consistent with the target variable’s measurement system. If next year’s sales are our objective y, then the RMSE will offer the error in dollars, whereas the MSE will give the error in dollars squared, which is considerably less comprehensible. The formula for RMSE is as follows Eq. 6.

    $$\begin{aligned} RMSE = \sqrt{\frac{1}{n}\Sigma _{i=1}^{n}{(Y_i -\hat{Y_i})^2}} \end{aligned}$$
    (6)

    In this formula: RMSE represents the Root Mean Square Error. n: is the number of data points or observations. \(y_i\): represents the actual or observed values. \(x_i\): represents the predicted or estimated values for the corresponding observations.

The selection of RMSE and MAE as evaluation metrics for this research is driven by their interpretability, sensitivity to outliers, and suitability for model optimisation. The measures used provide a fair assessment of both the average magnitude of mistakes and the consequences of bigger deviations. The combination of RMSE and MAE allows for a comprehensive evaluation that aligns with specific characteristics and priorities of a dataset and a recommendation system in the domain of location-aware recommendations, where the goal is to provide accurate and meaningful suggestions to users.

5.2 Result

This section provides visual representations of the different WS-Dream dataset-based trial results. The output for the WS-Dream dataset utilized is presented in the form of graphs and tables below.

Table 3, showcases a performance result of a proposed Hybrid CNN-LSTM model, evaluated employing the WS-Dream dataset. For the metric of Response Time, the MAE and RMSE exhibit a decreasing trend as the density increases from 0.5 to 0.25, with the lowest MAE and RMSE observed at a density of 0.25. Specifically, at a density of 0.25, the MAE is 0.359 and the RMSE is 0.1388. Similarly, for Throughput, both MAE and RMSE decrease as density increases, with the lowest values again observed at a density of 0.25. At this density, the MAE is 12.764, and the RMSE is 37.874. These results suggest that increasing density leads to improved performance of the Hybrid CNN-LSTM model in predicting both RT and Throughput metrics for location-aware web service recommendation tasks.

Table 3 Performance results of the proposed hybrid CNN-LSTM model

Table 4, showcases a performance result of a proposed Hybrid CNN-LSTM model, evaluated employing the WS-Dream dataset. For the metric of Response Time, the MAE and RMSE exhibit a decreasing trend as the density increases from 0.5 to 0.25, with the lowest MAE and RMSE observed at a density of 0.25. Specifically, at a density of 0.25, the MAE is 0.359 and the RMSE is 0.1388. Similarly, for Throughput, both MAE and RMSE decrease as density increases, with the lowest values again observed at a density of 0.25. At this density, the MAE is 12.764, and the RMSE is 37.874. These results suggest that increasing density leads to improved performance of the Hybrid CNN-LSTM model in predicting both RT and Throughput metrics for location-aware web service recommendation tasks.

Table 4 Response time experimental results

Table 5 presents a comprehensive comparison of experimental results for Response Time in Location-Aware Web Service Recommendation across various models and densities. The performance metrics, like MAE and RMSE, are depicted for densities ranging from 0.5 to 0.25. Among the models evaluated, Hybrid CNN-LSTM demonstrates competitive performance with decreasing MAE and RMSE values as density increases. Notably, at a density of 0.25, the Hybrid CNN-LSTM model achieves an MAE of 0.359 and an RMSE of 0.1388. For RLSD, at the same density, the MAE is 0.5214, and the RMSE is 0.1583. Comparison with other models reveals varying levels of effectiveness, with UMEAN, IMEAN, UPCC, and IPCC demonstrating relatively higher errors across different densities. Models like LDCF and NCF exhibit lower errors compared to others, while RLSD displays consistent performance across densities. Overall, the results underscore the effectiveness of the Hybrid CNN-LSTM model in accurately predicting Response Time, particularly at higher densities, showcasing its potential for enhancing location-aware web service recommendation systems.

Table 5 Throughput experimental results

Figures 4 and 5 show the results of an MAE and RMSE comparison of a proposed hybrid CNN-LSTM model with those of current models in terms of reaction time and throughput at varying densities (0.05 to 0.25). A more precise model could be possible with less loss (except if the model was overfitting to the data in the training set). During both training and testing, the loss value may be calculated to evaluate the model’s efficacy on the two datasets. Similar to accuracy, the loss does not have a proportional representation. It’s the sum of all the blunders in a sample’s training and testing data. In terms of reaction time and throughput, the suggested hybrid CNN-LSTM model was demonstrably superior to modern methods.

Fig. 4
figure 4

Loss values comparison of response time

Fig. 5
figure 5

Loss values comparison of throughput

Figures 4 and 5 shows the results of comparing the proposed hybrid CNN-LSTM model with current models in terms of reaction time and throughput at different densities (0.05 to 0.25) using the mean absolute error (MAE) and root mean squared error (RMSE). A more precise model could be possible with less loss (except if the model was overfitting to the data in the training set). During both training and testing, the loss value may be calculated to evaluate the model’s efficacy on the two datasets. Similar to accuracy, the loss does not have a proportional representation. It’s the sum of all the blunders in a sample’s training and testing data. In terms of reaction time and throughput, the suggested hybrid CNN-LSTM model was demonstrably superior to state-of-the-art methods.

6 Conclusion

This research offers a unique deep neural network-based Location-aware services recommendation model, in particular the Hybrid CNN-LSTM model, which has been suggested as an approach for location-aware web service recommendation, exhibits outstanding accuracy in the WS-Dream dataset at different densities. The assessment, illustrated through the utilization of tables and figures, demonstrates that the model’s precision in forecasting reaction time and throughput enhances as the density level decreases. Notably, the minimum values of MAE and RMSE are detected at density 0.25. As illustrated in Tables 3 and 4, the reduced MAE and RMSE values for reaction time and throughput indicate that a model consistently outperforms modern approaches. The model’s precision is further underscored by loss value comparisons in Figs. 4 and 5, which demonstrate reduced losses throughout the training and testing phases in comparison to existing models. The combined evidence highlights the effectiveness of the Hybrid CNN-LSTM model that has been proposed for improving the precision of location-aware web service recommendations, with a particular emphasis on its performance at lower density levels. Despite this, it does what we need it to do; nonetheless, more research and improvement in QoS prediction are required. We hope to modify the hyperparameters of the deep learning model used for location service recommendations shortly. This study presents an application survey and research scope to encourage researchers to solve service recommendation difficulties and to aid them in picking a more effective algorithm strategy for a suggestion based on the system’s requirements and input sets. This work will continue to progress in the area of recommendation systems and improve their ability to fulfil user requirements.