1 Introduction

The growth of Web 2.0 has contributed to the rise in popularity of user-generated content-based platforms. Community Question-Answering (CQA) websites like Stack Overflow,Footnote 1 Quora,Footnote 2 and Yahoo! AnswersFootnote 3 have grown in popularity in recent years [1]. People now frequently look for information on CQA. The community’s members are allowed to respond to queries posed by users and contribute their own. A CQA question usually consists of three parts. First, the question’s subject briefly describes the question. The subject helps experts scan the questions and find those that are of interest to them. Then comes a section that describes the specifics of a question. The body part typically contains information about the subject and serves as a supplement to the subject. Finally, when a question is proposed, the question askers assign the corresponding tags. The person who asks the question need to wait for the experts to respond. When multiple responses are received to a question, the asker can select the best response as the accepted answer. While CQA services are a great resource for anyone looking for information, their fast expansion presents some special difficulties [2]. These websites receive thousands of new questions daily in addition to the millions of questions that currently exist. Finding a question that matches a respondent’s area of expertise is, therefore, challenging. After posting their inquiries, users anticipate high-quality responses. Hence several recommendation approaches are used to get more relevant individuals to contribute their thoughts and solutions. As a result, recent research on CQA systems has preferred to use recommendation algorithms and certainly associated data mining approaches to bring more active and intelligent solutions to help identify experts automatically. With such suggestions, there is a greater likelihood that each query will be promptly assessed by the appropriate users and will receive more insightful responses.

By recommending personalised and pertinent material to users across a variety of sectors, including movies [3, 4], books [5], and online commerce [6], recommendation systems play a crucial role in improving the user experience. The classification of Recommender Systems (RS) includes three subcategories: Collaborative Filtering (CF) RS [7, 8], content-based RS, and hybrid RS [9]. The user profile and the similarity of the question description are taken into account when the content-based technique makes suggestions. Based on the views of other users, CF provides suggestions to users. Because it won’t be concerned with the question’s description when making recommendations, it may offer sophisticated recommendations. Because of this property, CF has become a popular filtering approach and is crucial in many applications. Combining CF with content-based systems is what is known as a hybrid system. The primary problem with the CF recommender method is data sparsity [10], which is compounded by the cold start issue [11]. In the absence of complete information, the CF model struggles to make effective recommendations. A sparsity problem occurs when a user interacts with a small percentage of the questions in a specific application domain. Another issue in collaborative filtering is the cold start problem. This is because there is a scarcity of information about new entities, such as new users. When a new user was added to the system, she or he had no voting score, and the system could not determine the relevance to the question.

In this study, a novel approach called SSC-CF is proposed in order to address the sparsity and cold-start problems in CF algorithms and thereby improve the performance of expert recommender systems. This study makes an effort to develop a new CF-based recommendation system using dimensionality reduction and semantic similarity techniques. The suggested system solves the cold-start issue of recommender systems by utilizing Singular Value Decomposition++(SVD++) [12] and semantic similarity to increase prediction accuracy. The proposed system, SSC-CF, is built on the user- and question-based CF. For the question- and user-based CF recommendation portion, SSC-CF utilizes SVD++, and for the question-based semantic similarity calculation, it utilizes Bidirectional Encoder Representations from Transformers(BERT) embeddings with cosine similarity. The contributions of this work can be summarized as follows:

  • In this work, we propose a novel expert recommendation system, SSC-CF, that relies on dimensionality reduction and semantic similarity methods.

  • The accuracy of expert recommendation systems will increase as a result of solving the sparsity problem in CF.

  • The cold-start problem of expert recommendation systems can be solved with dimension reduction strategies.

The article is structured as follows: The recent research on CQA recommendation algorithms is introduced in Sect. 2. The statement of the problem for the proposed work SSC-CF is provided in Sect. 3. Section 4 talks about the suggested SSC-CF approach. The experimental findings are presented in Sect. 5, and the final section summarises the suggested work and potential future directions.

2 Related work

In the information retrieval community, question routing has received a lot of interest recently [13]. This section covers earlier studies on question routing and categorizes the methods into different groups: classification method, language models, topic model, network-based method, and collaborative filtering.

Classification Method: When considering experts as a specific class of users among all users, the issue of recognizing the experts can quickly be changed into a classification issue that seeks to separate such a specific class of expert users from the other users. Comparatively speaking to the other methods, categorization methods are more flexible in how they can apply a variety of features to the expert recommendation problem from the user’s perspective, including questions, answers, and user-user interactions. The most popular classification approach for separating experts from non-experts is a Support Vector Machine (SVM). Zhou et al., [14] employed SVM, but they also specified local and global features on questions, user history, and question-user relationships, as well as taking Kullback-Leible divergence into consideration as a novel feature. Ji et al., [15] used text similarity as a feature when training RankingSVM, a version of SVM. For classifying the experts, additional techniques like random forest [16, 17] and Naive Bayes [18] are also used.

Language Models: By computing the word-based relevance of a user’s prior behaviors to the query, language models compute the likelihood that a user would provide an answer to the question [19]. It is anticipated in a language model that the individuals whose profiles are most likely to generate the given query will also be the individuals who are most likely to deliver an answer. Finally, the model offers a sorted list of users who are most likely to reply to the inquiry. Some variants of the language models are relevance-based language model [20], cluster-based language model [21], and hierarchical-based language model [22].

Topic Model: Later, topic models were developed that do not require the precise word to be in the user profile but instead measure relationships in the topic space rather than the word space [23]. Latent Dirichlet Allocation (LDA) [24] and Probabilistic Latent Semantic Analysis (PLSA) [25] are a couple of topic models that are frequently employed. In order to represent a document in a low-dimension space, Latent Semantic Indexing (LSI) [26], which is the foundation of PLSA, requires Singular Value Decomposition. The data creation process is modeled as a Bayesian network in PLSA using latent topics to represent documents [27, 28]. In the LDA approach, the topic mixture is derived from a conjugate Dirichlet prior that is constant across all users. First, topics based on historical user activity are extracted using LDA to demonstrate the relationship between knowledgeable people and fresh questions. These topics are used to calculate each user’s likelihood of providing an answer in the second phase, and users are then ranked according to this likelihood [29].

Network-Based Method: The users with the highest levels of authority are suggested as the subject matter experts for new questions by the network-based approaches, which assess users’ authoritativeness in a user-user network created by their asking-answering relationships. By employing the degree centrality measure InDegree [30], which values users who have answered more questions in the user-user network as better answerers, it is the most straightforward way to assess a user’s authority in the CQA community. There are three basic network-based methods: PageRank [31], Hyperlink Induced Topic Search (HITS) [32], and Expertise Rank [30].

Collaborative Filtering Method: Matrix Factorization (MF) techniques, which are well known to be favorable in terms of flexibility and scalability in the recommendation domain, were employed in a different area of study [33]. In order to determine the user’s level of competence on particular words, Zhao et al., [7] employed MF to represent the questions with the words that made up their content. The performance of the MF technique is, however, negatively impacted by the high-dimensional, sparse matrix that comes from this [34]. The semantic similarity between words is also disregarded because the matrix factorization approach treats each element as an independent entity.Following in this approach, Yang et al., [35] suggested performing the MF for question retrieval using tags, which condense the topic of the question rather than the text of the questions and responses. The study showed that tags are more useful and do a better job of summarising the subject of the query. The method also assesses the answerer’s proficiency in a particular question and associated tags based on the number of votes for the answer. However, even if using tags rather than words somewhat resolves the dimensionality issue, it does not fully address the issue of data sparsity. The technique also continues to be hampered by the presence of a large number of associated items.To address one of the problems caused by the spelling variations of tagged keywords [35], Fukui et al., [8] attempted to enhance the strategy by expanding tagged keywords based on word embeddings.

There are a few drawbacks to the methods, including hand-crafted feature extraction required by the classification methodology. Word matching serves as the foundation of the language model, which leaves out semantics. Since topic extraction requires all of the questions and answers data, topic models in dynamic environments are computationally expensive. The network-based approach is unable to take into account the questions, themes, or categories that can be leveraged from the text. In conclusion, despite the fact that numerous innovative approaches to question routing have been proposed over the years, there are still a number of unresolved issues and potential areas for development, both of which this work aims to address.The table 1 provides a summary of the different question routing models.

Table 1 Overview of the question routing methods

3 Problem formulation

A static archive of a CQA website that preserves all question-and-answer sessions that have amassed through time serves as the foundation for a CQA Network. The CQA is made up of users set \(U=u_1,u_2,\ldots u_m\), where m is the total number of users on the community website, and these users include both the asker and the expert. An expert is someone who possesses the knowledge necessary to respond to a specific question in their domain. Additionally, it includes the question set \(Q=q_1, q_2,\ldots q_n\), where n represents the overall number of questions on the community website. The questionId, askerId, creation date, answerId, answererId, acceptedAnswerId, acceptedAnswererId, tag, title-body, bounty tuple is the format for each question \(q_i\in Q\). The tuple for each \(u_i \in U\) includes score, tag, questionId, answerId, and datePosted. As a result, the task of a recommendation system can be properly defined as having a user or expert predict a given query. Therefore, the structured approach of the recommendation system’s mission is the given a question/user, forecast the experts/question respectively.

4 Proposed approach

The proposed SSC-CF expert recommendation system is depicted in Fig. 1. The recommended approach tries to generate scalable and accurate expert recommendations. The approach is defined in terms of two main parts. The construction of the recommendation models occurs in the first stage. The clustering of the voting score, dimensionality reduction using SVD++, and creation of similarity matrices for the items and users are some of the important tasks carried out in this phase. Using K-means++ with the elbow method algorithm, users are initially clustered according to their voting scores for questions, expertise, and temporal knowledge, while questions are grouped according to their tags, bodies, and titles. Then, using BERT embeddings with cosine similarity, semantic similarity is determined for each cluster. While processing, each cluster is run through SVD++ to obtain the decomposition matrices. Users and questions each have unique SVD++ models. The prediction and expert recommendation tasks are made for a specific question in the second phase after the first phase trains on the models constructed. A prioritized list of experts is actually provided by the recommender system in answer to the target question. To do this, the target question is presented to one of the clusters chosen in the first round. On the basis of the prior vote score, the SVD++ algorithm is then used to determine how similar the target expert is to the other experts. Finally, we combine user and question-based predictions using a weighted method.

Fig. 1
figure 1

Block Diagram of the Proposed Work SSC-CF

The algorithm 1 provides a comprehensive set of steps to take to achieve expert prediction. This section provides a thorough overview of each step in the process.

figure a

4.1 Clustering module

Using the K-means++ with elbow approach, the user and question profiles are first clustered.The K-means++ algorithm is an improvement on the original K-means algorithm that aims to select the initial cluster centres more successfully. The standard K-means method selects the initial cluster centres at random, which might lead to less-than-ideal clustering results. By employing a more intricate initialization process that takes the distance between data points into consideration, K-means++ chooses the initial cluster centres in a way that is more likely to result in superior clustering. This initialization technique generates clustering results with faster convergence and higher accuracy when compared to the traditional K-means algorithm. By minimizing the sum across each cluster of the square of the distance between the point and its centroid, the K-Means++ with elbow technique algorithm seeks out k centroid positions \((C_1, C_2,\ldots C_k)\). The data are clustered using the Lloyd Algorithm, an iterative approximation algorithm. Both the efficient centroid initialization and the optimum number of clusters are employed in K-Means++ with the Elbow technique. The KMeans++ with Elbow Method is explained in detail in the algorithm 2. According to their voting results for questions, expertise, and temporal knowledge, users are grouped together based on similar preferences in the process of user clustering. The target user’s prediction task is carried out using the aggregated opinions in each cluster after the clusters have been created. As a result, speed is improved because the cluster that needs to be studied contains far fewer people than the total number of users. In Fig. 2, m represents the total number of users, \(a_{ij}\) represents the average vote score by user cluster center i for question j, \(R_{ij}\) represents the voting result for user i for question j, and n and c represents the total number of questions and user centers, respectively. Question clustering groups questions together according to similar tags, bodies, and titles. Following the clustering process, the target question is predicted using the aggregated answers to all other questions in any cluster. As a result, since the cluster that needs to be processed contains a lot fewer questions than all of them combined, performance is improved. In Fig. 3, m stands for the total number of users, \(a_{ij}\) is the average voting score for user i to question cluster center j, n for the total number of questions, \(R_{ij}\) for user i to question j’s voting score, and k for the total number of question centers.

figure b
Fig. 2
figure 2

User cluster in CF

Fig. 3
figure 3

Question cluster in CF

4.2 SVD++

Then, in order to deal with predicting the unknowable values related to the sparsity, SVD++ is used in each cluster.The Singular Value Decomposition (SVD) algorithm has an extension known as SVD++ that is extensively used for matrix factorization and recommendation systems. A new term in the SVD++ technique takes into account implicit user input, such as how frequently or for how long a user engages with a certain item. With this improvement, SVD++ is more competent than the original SVD technique to express user preferences and offer recommendations. SVD++ is preferred over SVD for recommendation systems that take into account implicit user feedback. The SVD++ algorithm is an enhanced version of the classic SVD algorithm. It considers the user’s voting score matrix R to be a product of two matrices, E and F. It also maps all users and all questions into a K-dimensional latent semantic space. The semantic space is made up of a collection of latent elements. The user voting score matrix is factorized as follows:

$$\begin{aligned} \begin{aligned} R_{U*Q}&= E_{U*K}* F_{K*Q}\\&= \begin{bmatrix} e_{11}, &{} \ldots &{} e_{1K}\\ e_{21}, &{}\ldots &{} e_{2K}\\ &{} \ldots &{} \\ e_{U1} &{}\ldots &{}e_{UK} \end{bmatrix}* \begin{bmatrix} f_{11} &{}\ldots &{} f_{1Q}\\ f_{21}&{} \ldots &{}f_{2Q} \\ &{}\ldots &{} \\ f_{K1} &{} \ldots &{} f_{KQ} \end{bmatrix} \end{aligned} \end{aligned}$$
(1)

The user set is represented by \(U=(u_{1},u_{2},\ldots ,u_{n})\) and the question set is represented by \(Q=(q_{1},q_{2},\ldots ,q_{m})\). \(e_{ik}\) denotes the user i’s expertise degree for the k-th latent factor in the question. The distribution of the k-th latent component is represented by \(f_{kj}\) among the questions j. As a result, each user has a user vector \(e_{u} \in R\), which is a row of the matrix E. And each question has its own question vector \(f_{q} \in R\), which is a column in the matrix F. \(e_{u}\) denotes the user’s expertise. The question’s feature space is described by \(f_{q}\). They are now in the same data space. As a result, the usual collaborative filtering technique may be implemented here: the dot product of these two vectors is used to obtain the expert \(\hat{r}_{uq}\) forecast, which is voted on by the user u for the question q. In the equation 2, the term \(\hat{r}_{uq}\) is expressed.

$$\begin{aligned} \hat{r}_{uq}=e_{u}^{T}f_{q}=f_{q}^{T}e_{u} \end{aligned}$$
(2)

Analyzing the user’s voting score matrix R reveals that some users consistently give high or low voting scores compared to others. This suggests that the vote results are biased. However, equation 2 does not take bias into account. As a result, numerous bias factors need to be considered to obtain a more objective voting score. The Eqs. (3), (4), (5) is the voting score that has been modified.

$$\begin{aligned} \hat{r}_{uq}&=b_{uq}+f_{q}^{T}e_{u}\end{aligned}$$
(3)
$$\begin{aligned} b_{uq}&=\mu +b_{u}+b_{q}\end{aligned}$$
(4)
$$\begin{aligned} \hat{r}_{uq}&=\mu +b_{u}+b_{q}+f_{q}^{T}e_{u} \end{aligned}$$
(5)

\(b_{uq}\) indicates the user u’s total bias information for the question q, whereas \(\mu\) is the voting score mean. The bias information of the question q is represented by \(b_{q}\), which is an item offset to the voting score mean. The bias information of the user u, which is a user offset to the voting score mean, is represented by \(b_{u}\). Apart from the biased information, many implicit parameters are added to SVD++ to better reflect the user’s latent competence of the question. In general, users’ voting scores are referred to as explicit information, whereas user behavior is referred to as implicit information. Finally, a user’s preference perspective is obtained by merging the previously mentioned explicit information, bias information, and implicit information. The final voting score is given in the equation 6.

$$\begin{aligned} \hat{r}_{uq}=\mu +b_{u}+b_{q}+f_{q}^{T}\bigg (e_{u}+|N_{u}|^{\frac{-1}{2}}\sum _{j\in N_{u}}^{}y_{j}\bigg ) \end{aligned}$$
(6)

\(N_{u}\) represents the user u’s behavior data. This indicates that the user u voted on the corresponding question. The size of the behavior data is given by \(|N|\). \(-(\frac{1}{2})\) is an empirical value for the contraction factor. \(y_{j}\) represents the implicit parameters used in the recommendation to describe the implicit information, indicating that the user u voted for the question q. Finally, in addition to explicit information, the SVD++ algorithm considers both bias information and implicit information. It first examines users’ levels of expertise for the question’s latent semantic factors. The distribution of the latent semantic factors among all questions is then obtained. Finally, it considers both the bias information and the implicit information mentioned above. Based on the preceding analysis, the equation 7 represents the cost function J of the SVD++ method.

$$\begin{aligned} \begin{aligned} J&= \min _{ {b}_q, {b}_u, {f}_q, {e}_u} \sum _{(u, e) \in K}\left( r_{u q}-\mu - {b}_u- {b}_q- {f}_q^T\right. \\&\left. \quad \left( {e}_u+|N_u|^{-1 / 2} \sum _{j \in N_u} {y}_j\right) \right) ^2 \\&\quad +\lambda \left\{ \sum _u\left( {b}_u^2+\left\| {e}_u\right\| ^2\right) \right. \\&\left. \quad +\sum _q\left( {b}_q^2+\left\| {f}_q\right\| ^2+\left\| {y}_q\right\| ^2\right) \right\} \end{aligned} \end{aligned}$$
(7)

The first component of J is the loss calculated using the least square method. The regularisation term is the second part of J. The stochastic gradient descent method is used to optimize (equations (8), (9), (10), (11), (12), (13) ) the proposed SVD++ algorithm:

$$\begin{aligned}&{b}_u \longleftarrow {b}_u+\gamma \left( g_{u q}-\lambda {b}_u\right) \end{aligned}$$
(8)
$$\begin{aligned}&{b}_q \longleftarrow {b}_q+\gamma \left( g_{u q}-\lambda {b}_q\right) \end{aligned}$$
(9)
$$\begin{aligned}&{e}_u \longleftarrow {e}_u+\gamma \left( g_{u q} {f}_q-\lambda {e}_u\right) \end{aligned}$$
(10)
$$\begin{aligned}&{f}_q \longleftarrow {f}_q+\gamma \left( g_{u q}\left( {e}_u+|N_u|^{-1 / 2} \sum _{j \in N_u} {y}_j\right) -\lambda {f}_q\right) \end{aligned}$$
(11)
$$\begin{aligned}&{y}_j \longleftarrow {y}_j+\gamma \left( g_{u q}|N_u|^{-1 / 2} {f}_q-\lambda {f}_q\right) \end{aligned}$$
(12)
$$\begin{aligned}&g_{u q}=r_{u q}-\widehat{r}_{u q} \end{aligned}$$
(13)

The three variables are \(g_{uq}\) (prediction error), \(\gamma\) (learning rate), and \(\lambda\) (regularisation parameter).

4.3 Semantic similarity module

In order to further increase the predictive accuracy, semantic similarity between the questions is determined, and the final predictions are made using a hybrid of both SVD++ and semantic similarity in each cluster of the question profile. The proposed work SSC-CF uses BERT [36] with cosine distance to calculate the semantic similarity between the questions. Over time, word vectors have changed from a one-hot environment in which every word was orthogonal to every other word to one in which word vectors can change to fit the context. BERT has the capacity to incorporate word meaning into densely packed vectors. Each value contained in the dense vector has a value and a purpose for existing in that value. Each encode layer outputs a set of dense vectors thanks to BERT’s prowess in producing them. The ability to obtain word vectors that morph based on context is made possible by language modeling models. The static embedding layer, the first layer of BERT’s total of 13 layers, was selected for initial training. After obtaining the embeddings, the inputs to which the semantic similarity calculation must be applied are turned into a vector. The cosine similarity between the vectors is then used to calculate the semantic similarity. By computing the cosine of the angle created by two vectors projected in three dimensions, cosine similarity is computed. The cosine similarity between two vectors with the same orientation is 1, but the similarity between two vectors oriented at 90 degrees is 0. Equation 14 contains the mathematical formula for the cosine similarity of vectors A and B.

$$\begin{aligned} \cos (\textbf{A},\textbf{B})= {\textbf{A} \textbf{B} \over \Vert \textbf{A}\Vert \Vert \textbf{B}\Vert } = \frac{ \sum _{i=1}^{n}{\textbf{A}_i\textbf{B}_i} }{ \sqrt{\sum _{i=1}^{n}{(\textbf{A}_i)^2}} \sqrt{\sum _{i=1}^{n}{(\textbf{B}_i)^2}} } \end{aligned}$$
(14)

Finally, based on the user and hybrid-question CF, the expert in the CQA is predicted.

5 Experimental results

The analysis examines the suggested recommendation system, SSC-CF, using data from stack exchange sites. The dataset used for the experiment is described in Sect. 5.1. The performance of the recommended system SSC-CF is evaluated in the Sect. 5.2

5.1 Dataset description

Data from the three exchange sites are used in this study. The table 2 contains some statistics on the data from stack exchange sites. The data is extracted from the stack exchange sites’ dump.Footnote 4 Fig. 4 displays the entity relationship diagram for the dataset being evaluated from stack exchange platforms.

Table 2 Statistics of Stack Exchange Sites
Fig. 4
figure 4

Stack exchange sites database entity relationship

5.2 Evaluation

The SSC-CF method combines the SVD++ and semantic similarity measures with question- and user-based clustering. Comparing the proposed technique SSC-CF to the question-based+(SVD++)+clustering+semantic similarity and question-user-based+(SVD++)+clustering recommendation systems. SSC-CF is examined for throughput, which is referred to as the number of suggestions per second, on the three stack exchange sites to demonstrate the efficacy of the suggested solution in addressing the scalability issue. The performance outcomes of all methods experiments are shown in the Fig. 5. The methods throughput is graphed as a function of cluster size. The clustering process in the Proposed Approach SSC-CF employs the K-means++ with elbow method. For the suggested method SSC-CF, various clustering sizes are taken into account. Plots show that methods that use clustering, dimensionality reduction, and semantic similarity techniques have significantly higher throughput than other methods. Additionally, it can be shown from the Fig. 5 that the throughput of the approaches increases with an increase in clustering size. The Mean Absolute Error (MAE) between the projected and actual vote score is assessed using statistical measures. In contrast, decision-support metrics, for example, by measuring the overlap, compare the suggested items with the pertinent ones. Equation 15 presents MAE.

$$\begin{aligned} \text{ MAE } \text{(pred, } \text{ act } )=\sum _{i=1}^N \left|\frac{ \text{ pred } _{u, i}- \text{ act } _{u, i}}{N}\right| \end{aligned}$$
(15)

where N is how many questions a person has taken and received a score on. The MAE is used to assess the predictive accuracy of the SSC-CF approach. For various neighboring densities, it has been evaluated. The prediction accuracy for various neighborhood sizes on datasets is shown in Fig. 6. For the multi-criteria recommender assessments in terms of accuracy measures, decision-support metrics, in particular, will be crucial. The following discussion will cover the measures for this use that are well-known in the field of information retrieval. The precision Eq. 16 counts the number of relevant items in the received result.

$$\begin{aligned} \text{ Precision } =\frac{T R}{T R+F R} \end{aligned}$$
(16)

where FR stands for false relevant forecasts, TR stands for true relevant predictions, and FN stands for false non-related predictions. The precision is calculated on several Top-N numbers in order to evaluate the suggested method using decision-support accuracy metrics. In this study, N = 10, 20, 30, 40, and 50 are taken into account, meaning that we analyze the approach when recommending the top 10, 20, 30, 40, and 50 movies using the suggested recommender system. Table 3 displays the precision figures for various Top-N. The table shows that our novel method produced rather high-precision results. The metrics show that the proposed method performs better than the alternative methods. Likewise, the precision of the approach suggested by Sun et al., [37] is outperformed by the proposed SSC-CF method and the Fig. 7 giveses us the comparison between the model SSC-CF, ColdRoute [37] and Seq(Pairwise) [38]. These findings are enough to back up our contention that our strategy is relatively scalable and improves accuracy. Two major problems with recommender system design are scalability and sparsity. In order to improve the functionality of recommender systems, efforts have been made to address these problems in this research. With the use of clustering and dimensionality reduction techniques, the method created in this work takes advantage of semantic similarity in the SSC-CF. Three datasets from stack exchange sites were utilized to evaluate the approach. According to the findings presented by MAE and Precision, the performance of the CF recommender systems was enhanced by the application of semantic similarity in conjunction with clustering and dimensionality reduction approaches. The analysis’s findings showed that the scalability and sparsity problems in recommender systems can be resolved using the hybrid recommendation method.

Fig. 5
figure 5

Throughput

Fig. 6
figure 6

MAE

Table 3 Precision for Different number of Top-N
Fig. 7
figure 7

Comparison of the proposed model SSC-CF with other approaches

6 Conclusion

The proliferation of community forums has recently increased the importance of tasks associated with them. It can be challenging to identify whom to look for answers to because of the constant influx of new questions on these forums. The issue of questions with no answers is attempted to be resolved by recommending inquiries to experts. The study focused on the collaborative filtering approach’s cold start issue and data sparsity issue for the expert recommendation system. The SSC-CF is a proposed approach that employs dimensionality reduction to deal with scalability or the cold-start problem and semantic similarity to boost accuracy. The experiments make use of the three most well-known three stack exchange platforms dataset such as gaming, physics, scifi. When the metrics were used to evaluate the SSC-CF, the findings showed that the proposed technique, SSC-CF, performed better due to the help of semantic similarity in addition to clustering and dimensionality reduction. Despite the fact that the study suggests a strategy for dealing with scalability and data sparsity issues, future work will focus on increasing the accuracy of the results obtained.