Keywords

1 Introduction

With the rapid development of event-based social network services, it becomes increasing popular to participate local events through the online services such as Meetup (www.meetup.com) and Douban Events (beijing.douban.com/events). Event-based social networks (EBSNs) not only contain online social networks by joining the same groups where users can organize, participate, comment, share and advertise events, but also include offline social networks where users make friends through attending the same face-to-face activities. The core goal of EBSNs is to make easy for neighbors (users located in the same city) together to do what they are commonly interested in [20]. To better organize events, these services allow users to join online groups, in which a user can publish and announce events to other group members [14]. For example, the online social networks on Meetup are the groups that organizers create and other numbers join, while the offline social networks on Meetup are captured in offline activities. Consequently, the users in EBSNs have two kinds of heterogeneous friend relationships that one is the online friend relationships and another is offline friend relationships.

Along with dramatic rise of the users and events in EBSNs, how to choose the interesting events for users is become very important and difficult. Meetup, the world’s largest network of local groups, currently has 25.72 million users with more than 240,000 Meetup groups and 580,000 monthly eventsFootnote 1. Therefore, it becomes essential to recommend from so many dazzling events to solve information overload problem. In contrast to traditional social network services (SNS), user behavior in EBSNs is predominantly driven by offline activities and highly influenced by a set of unique factors, such as spatio-temporal constraints and special social relationships [3]. As a result, event recommendation in EBSNs becomes very difficult. In addition, the event recommendation problem is arguably more challenging than classic recommendation scenarios (e.g. movies, books), since events have time limited efficacy, which means that event recommender systems have to recommend the event after it created and before it terminate.

Generally, users can only participate local events because of the limitation of distance. Different from existing recommendation problems, the characteristic of users’ friend relationships plays very important role in event recommendation. Particularly worth mentioning is that online and offline friend relationship are not identical in EBSNs. On Meetup, the online friends mean the users who are in the same online group, and the offline friends mean the users who participate the same offline activities.

In this paper, we introduce online and offline friend relationships to recommend events to users and also introduce the tie strength of online and offline friend relationships. Based on these factors and the multi-factor model, we propose a novel method, named Heterogeneous Social Poisson Factorization (HSPF). In summary, the main contributions of this work lie in the following three aspects:

  1. 1.

    We propose the Heterogeneous Social Poisson Factorization(HSPF) model which combines online and offline social networks into one framework.

  2. 2.

    We integrate the tie strength of online and offline friend relationships to proposed HSPF model and use the coordinate ascent alogrithm for inference of the HSPF model parameters.

  3. 3.

    Our experiments on the Meetup dataset show that for the task of event recommendation, our HSPF model outperforms other models.

2 Related Works

2.1 Social Recommendation

Recent studies have proposed various methods to include social information in matrix factorization process [19]. For instance, [8, 11, 12] include trust in recommendation process; zhou et al. [21] exploits users interactions to improve recommendation qualities; duan et al. [4] uses locations to build a personalized recommendation system; and chaney et al. [2] proposes SPF (Social Poisson Factorization) model using trusted friend relationships. All the above methods for recommender systems only include a single factor. In our work, we not only divide trusted friends into online and offline friends, but also introduce the tie strength of trusted friends.

2.2 Event Recommendation

Currently, there are a small amount research works studying on event recommendation in EBSNs. EBSNs are first analyzed in data mining field in [10]. Du et al. [3] explores the modeling of EBSNs users by utilizing content preference, spatio-temporal context, and social influence (the event organizer) features. Pham et al. [14] transforms the recommendation problems into node proximity calculation problem and proposes a general graph-based model. Macedo et al. [13] takes several context-aware recommenders as input features such as content-based, social, locational and temporal signals. Zhang et al. [20] proposes CBPF (collective Bayesian Poisson factorization) model utilizing users’ relationship information and events organizer, location, and textual content information to recommend local events. However, all the above methods do not introduce the heterogenous online+offline social relationships or cannot reflect the difference between online and offline social relationships. Qiao et al. [15] presents a Bayesian latent factor mode that unify the data, i.e., the geographical features, heterogenous online+offline social relationships and user implicit rating, for event recommendation. But it ignores the tie strength of the social relationships.

3 Preliminaries

Bayesian Poisson Factorization (BPF), proposed by Gopalan [5], is a probabilistic model of users and items for recommendation. BPF assumes that an observed rating matrix \(y_{ui}\) comes from a Poisson distribution:

$$\begin{aligned} y_{ui} \sim Poisson(\theta ^{T}_{u}\beta _{i}) \end{aligned}$$

where \(\theta _{u}\) is a non-negative K-vector of user preference and \(\beta _{i}\) is a non-negative K-vector of item attributes. \(\theta _{u}\) and \(\beta _{i}\) are hidden variables with Gamma priors:

$$\begin{aligned} \theta _{u,k}\sim & {} Gamma(\lambda _{ua},\lambda _{ub}) \\ \beta _{i,k}\sim & {} Gamma(\lambda _{ia},\lambda _{ib}) \end{aligned}$$

where \(\lambda _{ua}\) and \(\lambda _{ia}\) are the shape parameters of the Gamma distribution and \(\lambda _{ub}\) and \(\lambda _{ib}\) are the rate parameters of the Gamma distribution.

BPF can handle sparse data well and is more robust to the issue of overfitting [20]. We build on BPF to develop a model of data where users attend face-to-face offline events and the same users are organized in a online network.

4 Proposed HSPF Model

In this section, we describe the Heterogeneous Social Poisson Factorization(HSPF). We are given data about users, events and groups, where each user who belongs to some groups has attend some events. The groups mean the online social networks while the events represent the offline social networks.

HSPF is a latent variable model of user-event interaction and user-group interaction. HSPF uses Poisson factorization to model both of the interactions that are typically sparse. The user’s responses to events are determined by his/her instinct and behaviours of friends. That is the users would attend the events that are not consistent with his/her preference, just because that his/her friends attend it. The intuition is that the closer the relationship between of them, the greater the impact of the choices of RSVPs(Reply, if you please). EBSNs link the online and offline social worlds [10]. For different users, the impact of online and offline social worlds is different. As the graphical model of HSPF shown in Fig. 1, HSPF captures this intuition because there are three parts in the model: user’s preference \(\theta _{u}\), online social network influence \(\delta _{uv}\), and offline social network influence \(\tau _{uf}\). Then, we provide modeling details for the Heterogeneous Social Poisson Factorization(HSPF).

Fig. 1.
figure 1

Graphical model of the Heterogeneous Social Poisson Factorization(HSPF)

Heterogeneous Social Poisson Factorization(HSPF). We assume that the rating \(y_{ui}\) of a (user u, event i) pair equals one if the choice of RSVP is yes, i.e., the user u attends event i, and is zero otherwise. The relation \(r_{uv}\) of a (user u, online friend v) pair is regarded as the degree of online closeness between them, which can be computed by the number of groups that both of them belong to. The relation \(r_{uf}\) of a (user u, offline friend f) pair is considered as the degree of offline closeness between them, which is in proportion to the number of events that both of them attend. Each user u is represented by a vector of K latent preference \(\theta _{u}\) and each event i by a vector of K latent attributes \(\beta _{i}\). \(N^{on}(u)\) and \(N^{off}(u)\) are the set of online and offline social friends of user u respectively. \(\delta _{uv}\) and \(\tau _{uf}\) are the influences of online friend v and offline friend f respectively.

The distribution of the observation \(y_{ui}\) is denoted by

$$\begin{aligned} y_{ui}|y_{\lnot ui} \sim Poisson(\theta ^{T}_{u}\beta _{i} + \sum _{f \in N^{off}(u)}\tau _{uf}r_{uf}y_{fi} + \sum _{v \in N^{on}(u)}\delta _{uv}r_{uv}y_{vi}), \end{aligned}$$
(1)

where \(y_{\lnot ui}\) denotes the responses of other users. To complete the specification of the variables, we place Gamma priors with shape and rate parameters on the user’s preference \(\theta _{uk}\), item’s attribute \(\beta _{ik}\), online social network influence \(\delta _{uv}\), and offline social network influence \(\tau _{uf}\). This is because that Gamma distribution is the conjugate prior for Poisson distribution. It is very facilitate to Bayesian learning of model parameters. Furthermore, the Gamma prior encourages sparse representations of users, events and influences. More specifically, by setting the shape parameters, i.e., \(\lambda _{ua}\),\(\lambda _{ia}\),\(\lambda _{\tau a}\) and \(\lambda _{\delta a}\), to be small (e.g., 0.3), most of the generated values will be close to zero. Note that \(\lambda _{ub}\),\(\lambda _{ib}\),\(\lambda _{\tau b}\) and \(\lambda _{\delta b}\) are rate parameters.

Based on the above description, the generative process of the Heterogeneous Social Poisson Factorization(HSPF) is as follows:

  1. (1)

    For each user \(u = 1,...,U\) and each component \(k=1,...,K\), draw latent factor

    $$\begin{aligned} \theta _{uk} \sim Gamma(\lambda _{ua},\lambda _{ub}). \end{aligned}$$
  2. (2)

    For each event \(i = 1,...,I\) and each component \(k=1,...,K\), draw latent factor

    $$\begin{aligned} \beta _{ik} \sim Gamma(\lambda _{ia},\lambda _{ib}). \end{aligned}$$
  3. (3)

    For each influence \(\delta _{uv}\) of (user u, online-friend v) pair, draw latent factor

    $$\begin{aligned} \delta _{uv} \sim Gamma(\lambda _{\delta a},\lambda _{\delta b}). \end{aligned}$$
  4. (4)

    For each influence \(\tau _{uf}\) of (user u, offline-friend f) pair, draw latent factor

    $$\begin{aligned} \tau _{uf} \sim Gamma(\lambda _{\tau a},\lambda _{\tau b}). \end{aligned}$$
  5. (5)

    For each (user u, event i) pair, draw response \(y_{ui}\) through Eq. (1).

5 Inference Algorithm

The key inferential problem that we need to solve in order to use HSPF is posterior inference. We denote all the Gamma priors of latent factors with \(\lambda \). Given \(\varTheta =\{\theta _{uk}, \beta _{ik}, \delta _{uv}, \tau _{uf}\}\) , \(\mathrm {Y} = \{y_{ui},r_{uv},r_{uf}\}\), then the posterior distribution,

$$\begin{aligned} p(\varTheta |\mathrm {Y},\lambda ) = \frac{p(\varTheta ,\mathrm {Y}|\lambda )}{p(\mathrm {Y}|\lambda )} = \frac{p(\varTheta ,\mathrm {Y}|\lambda )}{\int p(\varTheta ,\mathrm {Y}|\lambda ) d\varTheta } \end{aligned}$$
(2)

which is intractable for exact inference due to the coupling between integration variables of the normalization term shown in Eq. (2).

We adopt the variational inference to approximate the posterior distribution because that the variational inference tends to scale better than alternative sampling based algorithm such as Markov chain Monte Carlo. Variational inference algorithms approximate the posterior by defining a parameterized family of distributions over the hidden variables, i.e., \(q(\varTheta |\gamma )\), and then fitting the parameters of \(q(\varTheta |\gamma )\) that is close to the posterior, i.e., \(p(\mathrm {Y}|\lambda )\).

The basic idea of convexity-based variational inference is to take advantage of Jensen’s inequality to obtain an evidence lower bound on the log likelihood [9].

$$\begin{aligned} \begin{array}{l} \displaystyle log p(\mathrm {Y}|\lambda ) \\ \displaystyle = log \int p(\varTheta ,\mathrm {Y}|\lambda ) d\varTheta \\ \displaystyle = log \int \frac{q(\varTheta |\gamma )}{q(\varTheta |\gamma )}p(\varTheta ,\mathrm {Y}|\lambda ) d\varTheta \\ \displaystyle \ge \int q(\varTheta |\gamma )log p(\varTheta ,\mathrm {Y}|\lambda ) d\varTheta - \int q(\varTheta |\gamma )log q(\varTheta |\gamma ) d\varTheta \\ \displaystyle = E_{q}[log p(\varTheta ,\mathrm {Y}|\lambda )] - E_{q}[log q(\varTheta |\gamma )] \\ \displaystyle = \mathcal {L}(\gamma ;\lambda ) \end{array} \end{aligned}$$
(3)

The difference between \(log p(\mathrm {Y}|\lambda )\) and \(\mathcal {L}(\gamma ;\lambda )\) of Eq. (3) is the Kullback-Leibler(KL) divergence between \(q(\varTheta |\gamma )\) and \(p(\mathrm {Y}|\lambda )\), that is

$$\begin{aligned} log p(\mathrm {Y}|\lambda ) = \mathcal {L}(\gamma ;\lambda ) + D(q(\varTheta |\gamma )||p(\mathrm {Y}|\lambda )). \end{aligned}$$
(4)

This shows that maximizing the evidence lower bound \(\mathcal {L}(\gamma ;\lambda )\) equivalent to minimizing the KL divergence, thus the problem of posterior inference becomes an optimization problem.

5.1 Auxiliary Variables

Our inference algorithm for HSPF makes use of general results about the class of conditionally conjugate models [5, 7]. We first give an alternative formulation of HSPF in which we add some auxiliary variables to facilitate derivation and description of the algorithm. Without the auxiliary variables, HSPF is not conditionally conjugate model.

Note that a sum of independent Poisson random variables is itself a Poisson with rate equal to the sum of the rates. We introduce the auxiliary latent variables \(z^{M}_{uik}\), \(z^{off}_{uif}\) and \(z^{on}_{uiv}\) for each (user u, event i) pair, each (user u, offline friend f) pair and each (user u, online friend v) pair respectively such that

$$\begin{aligned} y_{ui}|y_{\lnot ui} = \sum _{k=1}^{K}z^{M}_{uik}+\sum _{f=1}^{F}z^{off}_{uif}+\sum _{v=1}^{V}z^{on}_{uiv}, \end{aligned}$$
(5)

where

$$\begin{aligned} z^{M}_{uik}\sim & {} Poisson(\theta _{uk}\beta _{ik}),\\ z^{off}_{uif}\sim & {} Poisson(\tau _{uf}r_{uf}y_{fi}),\\ z^{on}_{uiv}\sim & {} Poisson(\delta _{uv}r_{uv}y_{vi}), \end{aligned}$$

and \(F=|N^{off}(u)|\), \(V = |N^{on}(u)|\).

After adding the auxiliary variables, the variational distribution, i.e., \(q(\varTheta |\gamma )\), turns to \(q(\varTheta ,Z|\gamma ,\phi )\) where Z and \(\phi \) denote all the auxiliary variables and added parameters respectively.

5.2 Mean-Field Variational Family

We resort to the mean-field variational family, where each latent variable is independent and governed by its own variational parameter. Omitting the parameters \(\gamma \) and \(\phi \) for simplicity, the mean-field variational family is

$$\begin{aligned} q(\varTheta ,Z) = \prod _{u,k}q(\theta _{uk}) \prod _{i,k}q(\beta _{ik}) \prod _{u,f}q(\tau _{uf}) \prod _{u,v}q(\delta _{uv}) \prod _{u,i,k}q(z_{uik}), \end{aligned}$$

Each factor in the mean-field family usually is set to the same type of distribution as its complete conditional [2, 5, 6, 20]. A complete conditional which is the conditional distribution of a latent variable given the observations and other latent variables in the model.

Complete Conditional. Firstly, we compute the complete conditionals of all the latent variables in the model. For the user preferences \(\theta _{uk}\), the complete conditional is a Gamma shown as follows,

$$\begin{aligned}&p(\theta _{uk}|\lambda ,\beta ,\tau ,\delta ,z,y) \nonumber \\&\propto p(\theta _{uk}|\lambda _{ua},\lambda _{ub})\prod _{i} p(z^{M}_{uik}|\theta _{uk},\beta _{ik}) \nonumber \\&= Gamma(\theta _{uk};\lambda _{ua},\lambda _{ub})\prod _{i} Poisson(z^{M}_{uik};\theta _{uk},\beta _{ik}) \nonumber \\&= Gamma(\lambda _{ua}+\sum _{i}z^{M}_{uik},\lambda _{ub}+\sum _{i}\beta _{ik}). \end{aligned}$$
(6)

We can similarly derive the complete conditionals for \(\beta _{ik}\), \(\tau _{uf}\), and \(\delta _{uv}\).

$$\begin{aligned} \beta _{ik}|\lambda ,\theta ,\tau ,\delta ,z,y\sim & {} Gamma(\lambda _{ia}+\sum _{u}z^{M}_{uik},\lambda _{ib}+\sum _{u}\theta _{uk}) , \end{aligned}$$
(7)
$$\begin{aligned} \tau _{uf}|\lambda ,\theta ,\beta ,\delta ,z,y\sim & {} Gamma(\lambda _{fa}+\sum _{i}z^{off}_{uif},\lambda _{fb}+\sum _{i}r_{uf}y_{fi}) , \end{aligned}$$
(8)
$$\begin{aligned} \delta _{uv}|\lambda ,\theta ,\beta ,\tau ,z,y\sim & {} Gamma(\lambda _{va}+\sum _{i}z^{on}_{uiv},\lambda _{vb}+\sum _{i}r_{uv}y_{vi}) \end{aligned}$$
(9)

The conditional distribution of a set of Poisson variables, given their sum, is a multinomial for which the parameter is their normalized set of rates [1]. So the complete conditionals for \(z_{ui} = (z_{ui}^{M},z_{ui}^{off},z_{ui}^{on})\) is multinomial, i.e., \(z_{ui} \sim Mult(y_{ui},\psi _{ui})\), where \(\psi _{ui} = (\psi _{ui}^{M},\psi _{ui}^{off},\psi _{ui}^{on})\) is a point in the (\(K+F+V\))-simplex, i.e.,

$$\begin{aligned} \psi _{ui}^{M}\propto & {} \langle \theta _{u1}\beta _{i1},...,\theta _{uK}\beta _{iK} \rangle , \\ \psi _{ui}^{off}\propto & {} \langle \tau _{u1}r_{u1}y_{1i},...,\tau _{u_{F}}r_{u_{F}}y_{_{F}i} \rangle , \\ \psi _{ui}^{on}\propto & {} \langle \delta _{u1}r_{u1}y_{1i},...,\delta _{u_{V}}r_{u_{V}}y_{_{V}i} \rangle , \\ \end{aligned}$$

Note that the parameters of \(\psi _{ui}^{M},\psi _{ui}^{off}\) and \(\psi _{ui}^{on}\) should be normalized together to ensure their sum to be one.

Variational Parameter. Then, we set each factor in the mean-field family to be the same type of distribution as its complete conditional.

The complete conditionals of \(\theta _{uk}, \beta _{ik}, \delta _{uv}\), and \( \tau _{uf}\) are Gamma distributions, so their variational parameters are Gamma parameters, i.e., variational distributions for \(\theta _{uk}, \beta _{ik}, \tau _{uf}\), and \(\delta _{uv}\) are \(Gamma(\gamma ^{shp}_{uk},\gamma ^{rte}_{uk})\), \(Gamma(\gamma ^{shp}_{ik},\gamma ^{rte}_{ik})\), \(Gamma(\gamma ^{shp}_{uf},\gamma ^{rte}_{uf})\), and \(Gamma(\gamma ^{shp}_{uv},\gamma ^{rte}_{uv})\) respectively. We can similarly deduce the variational distribution for \(z_{ui} = (z_{ui}^{M},z_{ui}^{off},z_{ui}^{on})\) is \(Mult(y_{ui},\phi _{ui})\), where \(\phi _{ui} = (\phi _{ui}^{M},\phi _{ui}^{off},\phi _{ui}^{on})\).

We set each variational parameter equal to the expected parameter (under q) of the complete conditional because of the conditionally conjugate models [7].

For variational Gamma distributions, we take \(\theta _{uk}\) as an example to derive the close-form update solution of parameters, i.e., shape parameter \(\gamma ^{shp}_{uk}\) and rate parameter \(\gamma ^{rte}_{uk}\).

$$\begin{aligned} \gamma ^{shp}_{uk}&= E_{q}[\lambda _{ua}+\sum \nolimits _{i}z^{M}_{uik}] \;= \lambda _{ua}+\sum _{i}y_{ui}\phi _{uik}^{M} , \end{aligned}$$
(10)
$$\begin{aligned} \gamma ^{rte}_{uk}&= E_{q}[\lambda _{ub}+\sum \nolimits _{i}\beta _{ik}] \;= \lambda _{ub}+\sum _{i}\frac{\gamma ^{shp}_{ik}}{\gamma ^{rte}_{ik}} , \end{aligned}$$
(11)

The update solutions for the parameters of \(\beta _{ik}\), \(\tau _{uf}\), and \(\delta _{uv}\) are similarly derived, we omit the details and provide the final results.

$$\begin{aligned} \gamma ^{shp}_{ik}= & {} \lambda _{ia}+\sum _{u}y_{ui}\phi _{uik}^{M} , \end{aligned}$$
(12)
$$\begin{aligned} \gamma ^{rte}_{ik}= & {} \lambda _{ib}+\sum _{u}\frac{\gamma ^{shp}_{uk}}{\gamma ^{rte}_{uk}} , \end{aligned}$$
(13)
$$\begin{aligned} \gamma ^{shp}_{uf}= & {} \lambda _{fa}+\sum _{i}y_{uf}\phi _{uif}^{off} , \end{aligned}$$
(14)
$$\begin{aligned} \gamma ^{rte}_{uf}= & {} \lambda _{fb}+\sum _{i}r_{uf}y_{fi} , \end{aligned}$$
(15)
$$\begin{aligned} \gamma ^{shp}_{uv}= & {} \lambda _{va}+\sum _{i}y_{uv}\phi _{uiv}^{on} , \end{aligned}$$
(16)
$$\begin{aligned} \gamma ^{rte}_{uv}= & {} \lambda _{vb}+\sum _{i}r_{uv}y_{vi} . \end{aligned}$$
(17)

For variational multinomial distribution, we take \(z_{ui}^{M}\) as an example to derive update solution of parameter, i.e., \(\phi _{ui}^{M}\).

$$\begin{aligned} \phi _{uik}^{M}\propto & {} G_{q}[\theta _{uk}\beta _{ik}] \nonumber \\= & {} exp\{E_{q}[log(\theta _{uk}) + log(\beta _{ik})]\} \nonumber \\= & {} exp\{E_{q}[log(\theta _{uk})] + E_{q}[log(\beta _{ik})]\} \nonumber \\= & {} exp\{\varPsi (\gamma ^{shp}_{uk})-log\gamma ^{rte}_{uk} + \varPsi (\gamma ^{shp}_{ik})-log\gamma ^{rte}_{ik}) \}, \end{aligned}$$
(18)

where \(k=1,...,K\); \(\varPsi (\cdot )\) is the digamma function and \(G_{q}[\cdot ]=exp(E_{q}[log(\cdot )])\) denotes the geometric expectation [17]. Similarly,

$$\begin{aligned} \phi _{uif}^{off}\propto & {} exp\{\varPsi (\gamma ^{shp}_{uf})-log\gamma ^{rte}_{uf} \} + r_{uf}y_{fi}, \end{aligned}$$
(19)
$$\begin{aligned} \phi _{uiv}^{on}\propto & {} exp\{\varPsi (\gamma ^{shp}_{uv})-log\gamma ^{rte}_{uv} \} + r_{uv}y_{vi}, \end{aligned}$$
(20)

where \(f=1,...,F\) and \(v=1,...,V\). Note that the parameters of \(\phi _{ui}^{M},\phi _{ui}^{off}\) and \(\phi _{ui}^{on}\) should be normalized together to ensure their sum to be one.

5.3 Coordinate Ascent Algorithm

The coordinate ascent algorithm is illustrated in Fig. 2, which iteratively optimize each variational parameter while holding the others fixed.

Fig. 2.
figure 2

The coordinate ascent algorithm for HSPF

5.4 Event Recommendation Using HSPF

Once the posterior is fit, we use the HSPF to recommend events, which are unattend and ongoing events, to users who will like to attend. Firstly, we compute the predicting scores \(\hat{y}_{ui}\) (i.e., Eq. 21) for each (user u, unattend and ongoing event i) pair by their posterior expected Poisson parameters. Then, we rank the events for each user using the scores. Lastly, Top-N events will be recommended to each user.

$$\begin{aligned} \hat{y}_{ui}= & {} E_{q}[\theta ^{T}_{u}\beta _{i} + \sum _{f \in N^{off}(u)}\tau _{uf}r_{uf}y_{fi} + \sum _{v \in N^{on}(u)}\delta _{uv}r_{uv}y_{vi}] \nonumber \\= & {} \sum _{k=1}^{K}E_{q}[\theta _{uk}\beta _{ik}] + \sum _{f \in N^{off}(u)}E_{q}[\tau _{uf}]r_{uf}y_{fi} + \sum _{v \in N^{on}(u)}E_{q}[\delta _{uv}]r_{uv}y_{vi} \nonumber \\= & {} \sum _{k=1}^{K}\frac{\gamma ^{shp}_{uk}}{\gamma ^{rte}_{uk}}\frac{\gamma ^{shp}_{ik}}{\gamma ^{rte}_{ik}} + \sum _{f \in N^{off}(u)}\frac{\gamma ^{shp}_{uf}}{\gamma ^{rte}_{uf}}r_{uf}y_{fi} + \sum _{v \in N^{on}(u)}\frac{\gamma ^{shp}_{uv}}{\gamma ^{rte}_{uv}}r_{uv}y_{vi} \end{aligned}$$
(21)

6 Experiments

6.1 Dataset

We conduct experiments on the Meetup dataset [13]. The dataset contains all public activities on Meetup from January, 2010 to April, 2014, which we use to evaluate the HSPF model proposed in this work. We select two cities located in the USA, namely Chicago and Phoenix Table 1.

Table 1. Statistics of the Meetup dataset.

6.2 Setup and Metrics

We select data during each of last 10 months as the test set for each city. The data during preceding 6 months of each test set is regarded as the corresponding training set. We firstly compute the scores for each (user u, event i) pair in the test set, then sort the triplets by these scores according to user, lastly recommend Top-N events to each user. We report the average value of the predicted performance over the 10 test set. We choose NDCG@N and Precision@N as the metrics.

6.3 Comparison Methods

To demonstrate the effectiveness of our HSPF model, we compare it to the following state-of-the-art recommenders:

  • CMF [18] solves the recommendation task by reconstructing multiple relation matrices which are associated with some shared elements. In this work, the users are the shared elements.

  • FM [16] a general predictor and is easily applicable to a wide variety of contexts, which include user, event, online and offline social networks in Meetup dataset, by specifying only the input data.

  • BPF, proposed by Gopalan [5], is a probabilistic model of users and items for recommendation. BPF can handle sparse data well and models the long-tail of users and items.

  • SPF-on and SPF-off stem from SPF [2] which aims to bridge the gap between preference- and social-based recommendations. However, SPF only incorporates one kind of social networks information into a Poisson factorization method and can’t capture the tie strength of friend relationships. SPF-on and SPF-off means only incorporates online and offline social networks respectively.

  • HSPF-on and HSPF-off are two variants of our HSPF model and can be considered as two special cases of our model. The difference of them is that they only consider the influence of online and offline social networks respectively.

6.4 Performance Comparison

For HSPF, the latent dimensionality, i.e., K, is set to 50 for both Chicago and Phoenix dataset. The Gamma priors parameters, i.e., shape and rate parameters, of all the latent factors are fixed to be 0.3.

Fig. 3.
figure 3

Comparisons of different methods on NDCG@N and Precision@N

Effectiveness Comparison. The detailed comparison results are shown in Fig. 3. From the Fig. 3, we can observe that:

  • The performance of HSPF,CMF and FM are better than SPF-off and SPF-on. The reason is that either SPF-off or SPF-on can only integrate one kind of social networks. The performance of event recommendation can be improved by integrating the heterogeneous social networks.

  • The proposed HSPF model achieves the best performance comparing to the other models. It can be explained that the user responses to events are implicit feedback [5] while HSPF is more suitable for modeling implicit user feedback.

Factor Contribution. From the detailed comparison results which are shown in Fig. 4, we can draw the conclusions that:

  • The performance of HSPF,HSPF-off and HSPF-on outperforms that of BPF which can be considered as a special case of HSPF without social network information. It shows that the performance of recommender systems can be improved by integrating the social network information into the model.

  • HSPF-off achieves better results than HSPF-on, which means that the influence of offline social networks is greater than the online social network. This explains that face to face communication helps to become a true friend.

Fig. 4.
figure 4

Factor contribution of HSPF model

7 Conclusion

In this work, we propose a Heterogeneous Social Poisson Factorization(HSPF) model which combines online and offline social networks into one framework, and integrates the tie strength of online and offline friend relationships to the model. We test HSPF on Meetup dataset. Experimental results demonstrate that HSPF outperforms state-of-the-art recommendation methods. In the future work, we plan to exploit the influence of time, location, and content.