Abstract
Social networks are becoming important dissemination platforms, and a large body of works have been performed on viral marketing, but most are to maximize the benefits associated with the number of active nodes. In this paper, we study the benefits related to interactions among activated nodes. Furthermore, since the stochasticity caused by the dynamics of influence cascade in the social network, we propose the adaptive seeding strategy where seeds are selected one by one according to influence propagation situation of seeds already selected, and define the adaptive profit maximization problem. We analyze its complexity and prove it is not adaptive submodular. We find the upper and lower bounds which are adaptive submodular and design an adaptive sandwich policy based on the sandwich strategy which could gain a data dependent approximation solution. Through real data sets, we verify the effectiveness of our proposed algorithm.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
With the advancements in information science in the last two decades, social networks find important applications in viral marketing, where a few individuals are provided with free products, hoping that the product will be recursively recommended by each individual to his friends to create a large cascade of further adoptions. Kempe et al. [14] first formulate it as the influence maximization problem on independent cascade (IC) model and linear threshold (LT) model. After that, a large body of works have been performed on viral marketing to maximize the benefits associated with the number of active nodes.
In this paper, we study the benefits related to strength of interactions among activated nodes. For example, when multiple people participate in one game, a large amount of data are transferred to synchronize the settings of game environment and operations between players [27]. Network provider generally charges based on the amount of data used by customers. In addition, interaction behaviors between game players may bring some advertising revenues to the network provider [25]. We consider the scenario of gaming on console like Nintendo switch, which provides online play mode. Players can find online opponents and play, in this case, the interaction determines the game company profit. Let’s take the game “overcooked 2” as an example, when playing online mode, 2 or 4 players can join one game, they can select a new released level to play together, that level is not free if played locally. Thus the online gaming is actually a live advertisement of the game depending on the interaction. Affected by the advertisement, players may later buy the new levels and play locally. We use profit to represent the revenues related to the strength of interactions between players and define a profit maximization problem, which selecting a seed set to maximize the profits between activated nodes.
Furthermore, one must deal with the stochasticity caused by the dynamics of influence cascade [24]. Specifically, when a seed is selected, how the influence propagates and how many nodes it will eventually affect is uncertain. In such settings, the outcome of a seed is not known in advance and will only be discovered after it is selected.
Motivated by the above observations, we aim to consider adaptive strategies to address the profit maximization problem. We pick a node, get to see which nodes it influenced and the profit among them, then adaptively pick the next node based on these observations and so on. Note that when picking the next node, we take into account the current profits of influenced nodes. To this end, we formulate adaptive profit maximization problem (APMP) in this work.
Solving such stochastic optimization problems is a notoriously difficult challenge. However, Golovin et al. [10] introduce the concept of adaptive submodularity and prove that if the problem satisfies this property, the adaptive greedy policy can achieve a \(1-1/e\) approximate solution to the optimal policy. But unfortunately, our adaptive profit maximization problem is not adaptive submodular, thus the greedy strategy cannot obtain a guaranteed approximate solution. To solve this problem, we find the upper and lower bounds of the objective function both of which are adaptive submodular, and then use the sandwich strategy [16] to design an adaptive sandwich policy, which can obtain a data-dependent approximate solution.
The contributions of this paper are summarized as follows.
-
We propose a new problem named adaptive profit maximization problem, analyze its complexity and prove it is not adaptive submodular.
-
We find the upper and lower bounds of APMP and prove both of them are adaptive submodular.
-
Based on the upper and lower bound, we design an adaptive sandwich policy for APMP using the sandwich strategy and get a data dependent approximate solution.
-
Through real data sets, we verify the effectiveness of our proposed algorithms.
To the best of our knowledge, this is the first paper applying the sandwich strategy to the stochastic optimization problem which is not adaptive submodular.
The rest of the paper is organized as follows. Section 2 is devoted to the related work. The profit maximization problem is proposed in Sect. 3. The preliminaries about stochastic optimization and adaptive profit maximization problem are provided in Sect. 4. The upper bound and lower bound are presented in Sect. 5. In Sect. 6 we propose the adaptive sandwich policy. The experiments are presented in Sect. 7.
2 Related work
Kempe et al. [14] formulate the influence maximization problem under IC and LT models and provide a greedy algorithm with an approximation ratio. Since the problem is submodular a simple greedy algorithm gain an approximation ratio of \(1-1/e\) [9, 17]. Since then, considerable work [3, 4, 16, 18, 19, 21, 22, 22, 23] has been devoted to extending existing models to study influence maximization and its variants. Recently, Wang et al. [26] study the activity maximization problem to maximize the sum of activity strengths among the influenced users.
Our work is related to [26] but we are different from it in the following aspects. First, the optimization problem we addressed in this paper is adaptive profit maximization problem which is different from the normal activity maximization. Our optimization strategy is selecting seeds one by one according to current observations while activity maximization is to choose a seed set in one step. Second, the property of our problem is totally different from [26] as we prove our problem is adaptive non-submodular while the problem in [26] is non-submodular. And we prove the upper and lower bounds are adaptive submodular respectively while they prove they are submodular. Third, to solve the adaptive non-submodular problem in our paper, we try to apply the sandwich algorithm to solve the adaptive non-submodular problem and successfully prove it could get a data dependent approximation solution. This is the first method to solve the adaptive non-submodular problem. Actually, we successfully apply the sandwich algorithm to the adaptive non-submodular problem like Golovin and Krause in [10] apply the greedy algorithm to adaptive submodular problem which is a meaningful work.
The adaptive strategy is a stochastic optimization framework. Asadpour et al. [1] present the preliminary results of stochastic submodular maximization problem with respect to a cardinality constraint. Golovin and Krause [10] define the concept of adaptive submodularity, generalizing submodular set functions to adaptive policies, and propose a greedy adaptive algorithm with an approximation ratio. After their initial works, algorithms for stochastic maximization problems with adaptive submodular functions have been studied in [7, 8, 11]. Other applications of adaptive submodularity in artificial intelligence are shown in [5, 6, 15].
Most of them are to address the adaptive submodular problem and apply the greedy framework. There is no work to solve the adaptive non-submodular problem. To the best of our knowledge, our paper has been the first work that tries to solve the adaptive non-submodular problem since the definition of adaptive submodular is proposed by Golovin and Krause in [10].
For the application of stochastic optimization in social networks, Seeman and Singer [20] consider the adaptive approach to a variant influence maximization problem, and Horel and Singer [13] consider the strategies with two seeding steps. Hatano et al. [12] study adaptive budget allocation problem. Tong et al. [24] introduce the adaptive seeding strategy based on a dynamic independent cascade model. Yuan and Tang [28] study the influence maximization discount allocation problem.
3 Non-adaptive profit maximization problem
In this section, we first propose the non-adaptive profit maximization problem in IC model.
In IC model each node has two states, active and inactive. Every edge (u, v) is assigned with a probability \(p_{uv}\) so that when u is active, v is activated by u with probability \(p_{uv}\). And the profit between nodes is represented by a nonnegative function \(c: V\times V\rightarrow \mathbb {R}_{\ge 0}\), in which \(c(u,v)=c(v,u)\) for the unordered pair \(\{u, v\}\) of node u and v. For any seed set S, denote by I(S) the set of all active nodes at end of the diffusion process. We denote the P(A) as the collection of subsets of set A whose cardinality is two. For example, if \(A=\{1,2,3\}\), then \(P(A)=\{\{1,2\},\{1,3\},\{2,3\}\}\). Actually the set P consists of all the pairs of nodes in set A. The expected profit would be defined as
Note that for each unordered pair \(\{u, v\}\), we only compute once the profit between them denoted as c(u, v) or c(v, u).
Definition 1
Non-Adaptive Profit Maximization: Given a social network \(G=(V,E)\) under the IC model, a profit function \(c:V\times V \rightarrow \mathbb {R}_{\ge 0}\), and a positive integer k, find a set S of k seeds to maximize the expected profit between activated nodes by S through influence propagation:
Theorem 1
Non-adaptive profit maximization problem is NP-hard.
Proof
We prove it by reducing from the activity maximization problem proposed by Wang et al. [26] which is NP-hard. Note that they only consider the activity between active nodes which are connected by edges. But we consider the benefits among all active nodes, regardless of whether there is an edge connecting them. To this end, our non-adaptive profit maximization problem can be viewed as a significant extension of activity maximization problem. In order to make an obvious comparison with our problem, we give an equivalent formulation about it: given a social network \(G=(V,E)\) with the IC model, each edge \((u,v)\in E\) is associated with an activity strength \(A_{u,v}\) and a propagation probability \(B_{u,v}\), the activity of a given seed set S is defined as
The activity maximization problem can be stated as follows:
\(\square \)
Remark 1
The activity maximization problem is polynomial time reducible to the non-adaptive profit maximization problem. Given an instance of activity maximization problem, we construct a corresponding graph which has the same network topology and the propagation probability. And we construct the profit function for profit maximization problem as follows. For each pair of nodes \(\{u,v\}\) between which there are edges, we set the profit value on unordered pair \(\{u, v\}\) to be the sum of activity strength \(A_{u,v}\) and \(A_{v,u}\). For the pair of nodes \(\{u,v\}\) between which there are no edges, we set the profit value on unordered pair \(\{u, v\}\) to be 0. Therefore, the activity maximization problem is equivalent to solve the non-adaptive profit maximization problem.
According to Remark1 and Theorem 1 in [26] which says activity maximization problem is NP-hard, the theorem follows immediately. \(\square \)
Theorem 2
Give a seed S, computing the exact game profit f(S) is \(\sharp P\)-hard.
Proof
According to Remark1 and Theorem 2 in [26] which says computing the exact activity \(\delta _A(S)\) under IC model given a seed set S is \(\sharp P\)-hard, the theorem follows immediately.\(\square \)
4 Adaptive profit maximization problem
In this section, we propose the adaptive seeding strategy for profit maximization problem, where seeds are selected one by one according to influence propagation situation of seeds already selected.
4.1 Preliminaries
We employ the following concepts to formulate our problem.
Definition 2
Realization and Partial realization: Each node \(v\in V\) is in a particular state from a set O of possible states. A realization is a function \( \phi : V \rightarrow O\), representing the states of nodes. And \(\phi (v)\) is the state of v. We use \(\varPhi \) to denote a random realization. A partial realization \(\psi : H \rightarrow O , H \subseteq V \), is a function from some subset H of V to their states. It is actually a set of relations, so that \(\psi \subseteq V \times O\) equals \( \{(v,o): \psi (v) = o\} \). We use the notation \(dom (\psi ) =\{ v: (v,o) \in \psi \}\) to refer to the domain of \(\psi \). A partial realization \( \psi \) is consistent with a realization \(\phi \) if they are equal everywhere in the domain of \(\psi \). In this case, we write \( \phi \sim \psi \). We say \(\psi \) is a subrealization of \(\psi '\) if and only if \(\psi \subseteq \psi ' \).
In social networks, when a node is selected, its state corresponds to the diffusion of influence other than the state of itself, active or inactive. Specifically, we model \(\phi (u)\) as a function \(\phi _{u} : E \rightarrow \{0,1,?\}\), where \( \phi _{u}((u,v))=0\) means that activating u has revealed that the edge (u, v) is dead; \( \phi _{u}((u,v))=1\) means the edge (u, v) is live, and \( \phi _{u}((u,v))= ?\) means the status of (u, v) is unknown. We require each realization to be consistent and complete. Consistency means that no edge should be declared both live and dead by any two states of nodes. That is, for all \(u,v \in V\) and \(e \in E\), \((\phi _{u}(e),\phi _{v}(e))\notin \{(1,0),(0,1)\}\). Completeness means that the status of each edge is revealed by some activation. Let \(E(\phi )\) denote the live edges as encoded by \(\phi \).
Note that the realization \(\phi \) represent the state of each node according to its definition, and in the social network it actually revealed the state of each edge. But these two expressions are not contradictory since we model \(\phi (u)\) as a function \(\phi _{u} : E \rightarrow \{0,1,?\}\). In different contexts, we will alternate between the two representations as needed, if there is no confusion.
Definition 3
The Full-Adoption Feedback Model: After activating seed u, we get to see the status (live or dead) of all edges existing v, for all nodes v reachable from u via live edges(i.e., reachable from u in \((V,E(\phi ))\)), where \(\phi \) is the true realization.
Definition 4
Adaptive policy: We encode our adaptive strategy for picking seeds as a policy \(\pi \) which is a function from a set of partial realizations to V, specifying which node to pick next under a particular set of observations.
4.2 Adaptive profit maximization problem
For the adaptive stochastic settings of profit maximization problem in the full adoption feedback model, we pick a node, get to see which nodes it influenced and compute the profit between them which is called the current observations. And when picking the next node, we need to take into account the current profits of influenced nodes. In other words, making a decision based on these observations and so on. We define the profit function as follows
where \(\phi \) is a realization which encode the states of each node selected (i.e., actually reveals the states of all edges, live or dead), \(I(S,\phi )\) is nodes activated by seed set S under realization \(\phi \), \(P(I(S,\phi ))\) is the all pairs of nodes in \(I(S,\phi )\). \(N(\pi ,\phi )\) is the seeds selected by policy \(\pi \) under realization \(\phi \). Based on this notation, the expected profit of a policy \(\pi \) is
where the expectation is taken with respect to \(p(\phi )\).
Definition 5
(Adaptive Profit Maximization Problem, APMP) Given a social network \(G=(V,E)\) under the full feedback model, a profit function \(c:V\times V \rightarrow \mathbb {R}_{\ge 0}\), and a positive integer k, find a policy \(\pi ^*\) to maximize the expected profit between nodes activated through influence propagation:
Remark 2
Let us first consider the very special case of adaptive profit maximization where the distribution \(p(\phi )\) is deterministic. Note that, we model \(\phi (i)\) as a function \(\phi _{i} : E \rightarrow \{0,1,?\}\). Because of completeness of the realization \(\phi \), the state of each edge is revealed, live or dead. And no edge is declared to be both live and dead by any two states of nodes, since the property of consistency. Through the above analysis, a deterministic realization corresponds to a live edge graph [14]which is obtained from live edges \(E(\phi )\) under realization \(\phi \) and all nodes. In this case, the realization \(\phi \) is known to the decision maker in advance, and thus there is no benefit in adaptive selection and the policies are non-adaptive. So when the distribution is deterministic, the APMP is transferred to the following problem: given a live graph \(G(V,E(\phi ))\) (i.e. given the realization \(\phi \)), where each edge(u, v) in \(E(\phi )\) is live, find a seed set S that max the profit of influenced nodes:
Our research work is mainly based on the theory of adaptive submodularity which is the first proposed by [10], the specific concepts associated with it are as follows
Definition 6
Conditional Expected Marginal Benefit: Given a partial realization \(\psi \) and node v, the conditional expected marginal benefit of v conditioned on having observed \(\psi \), denoted \(\triangle (v \mid \psi ) \), is
Definition 7
Adaptive Monotonicity: A function \( f : 2^{V} \times O^{V} \rightarrow \mathbb {R}_{\ge 0} \) is adaptive monotone with respect to distribution \( p(\phi ) \) if the conditional expected marginal benefit of any item is nonnegative, i.e., for all \(\psi \) with \(\mathbb {P}[\varPhi \sim \psi ] > 0\) and all \(v\in V\) we have
Definition 8
Adaptive Submodularity: A function \( f : 2^{V} \times O^{V} \rightarrow \mathbb {R}_{\ge 0} \) is adaptive submodular with respect to distribution \( p(\phi ) \) if the conditional expected marginal benefit of any fixed item does not increase as more items are selected and their states are observed. Formally, f is adaptive submodular w.r.t. \(p(\phi )\) if for all \(\psi \) and \(\psi '\) such that \(\psi \) is a subrealization of \(\psi '\) (i.e., \(\psi \subseteq \psi '\)), and for all \( v\in V\setminus dom(\psi ')\), we have
Theorem 3
\(f(S,\phi )\) is not adaptive submodular under full adoption feedback model.
Proof
We prove it by a special case proposed by Remark 2. Actually, when the distribution is deterministic, the diminishing returns property of adaptive submodularity reduces to the classical characterization of submodular set functions. So we construct a live graph as shown in Fig. 1. Note that each edge in the figure is bidirectional, and the first number in the tuple attached to each edge represents the propagation probability, the second number is the profit between its two endpoints. For example in Fig. 1, the tuple (0, 1) on edge \((v_{1},v_{2})\) means propagation probability \(p(v_{1},v_{2})= p(v_{2},v_{1})=0\) and \(c(v_{1},v_{2})=1\). And for the vertices between which there is no edge we set \(c(v_{i},v_{j})=0\). Then we have \(f(\{v_{1}\})=0\), \(f(\{v_{1}, v_{4}\})=1\), \(f(\{v_{1}, v_{2}\})=1\) and \(f(\{v_{1}, v_{2},v_{4}\})=3\). Thus, \(f(\{v_{1}, v_{4}\})-f(\{v_{1}\})<f(\{v_{1}, v_{2},v_{4}\})-f(\{v_{1}, v_{2}\})\), which implies \(f(\cdot )\) is not submodular. \(\square \)
Theorem 4
Adaptive profit maximization problem is NP-hard.
Proof
We prove APMP is NP-hard by showing a special case of it is NP-hard, where \(p(\phi )\) is deterministic as proposed by Remark 2. Now we prove by reducing from the set cover problem, which is NP-complete. Given a ground set \(U = \{ u_{1}, u_{2},\dots , u_{n}\}\) and a collection of sets \(\{S_{1}, S_{2}, \dots ,S_{m}\}\) whose union equals the ground set, the set cover problem is to decide if there exist k sets in S so that the union equals U. Given an instance of the set cover problem, we construct a corresponding live graph with \(m +2n\) nodes as follows. For each set \(S_{i}\) we create one node \(p_{i}\), and for each element \(u_{j}\) we create two nodes \(q_{j}\) and \(q'_{j}\). If the \(S_{i}\) contains the element \(u_{j}\), then we create two edges \((p_{i},q_{j})\) and \((p_{i},q'_{j})\). Note that each edge is live. Now we create the profit function over pairs of nodes. For the pairs \(\{q_{j},q'_{j}\}\), the profit equals to 1, and the other pairs equal to 0. Then the set cover problem is equivalent to deciding if there is a set S of k nodes such that the profit of S equals to n. The deterministic case of Adaptive profit maximization problem is NP-hard follows immediately and so is the theorem. \(\square \)
Through above analysis, we find that the problem APMP is not adaptive submodular and hard to solve. There is no general way to optimize or approximate a non-adaptive submodular function.
5 Adaptive submodular upper bound and lower bound
In this section, we propose an upper and a lower bound of the original problem APMP, and we prove that both of them are adaptive submodular. Thus, the adaptive greedy policy can achieve a \(1-1/e\) approximate solution to the expected reward of the best policy according to [10].
5.1 Upper bound
Under realization \(\phi \), the upper bound equals to the profit between nodes both of which are active plus half of the profit between nodes which are active and nodes which are not active. The former is exactly the profit we want to compute. And it can be defined as
where \(I(S,\phi )\) denotes the set of all active nodes under the realization \(\phi \), when S is the seed set and
Thus, we can see that the upper bound is essentially a weighted version of influence spread, where the weight of node v is \(\frac{1}{2}\sum _{u\in V}c(v,u)\), which equals to half of the sum of profit between v and the other nodes in V.
Lemma 1
Under any fixed realization \(\phi \), \(U(S,\phi )\) is monotone and submodular.
Proof
Given a fixed realization \(\phi \) which means the state of all the edges are determined according to Remark 2, we just need to prove \(U(S,\phi )\) is monotone and submodular in a live graph. Since the profit function \(c: V\times V \rightarrow \mathbb {R}_{\ge 0}\) is nonnegative which means the profit of each pair of nodes is non-negative. Thus the weight of every node is non-negative and the monotonicity of \(U(S,\phi )\) follows immediately. For the submodularity, we need prove \(U(A\cup \{v\}) -U(A) \ge U(B \cup \{v\}) -U(B)\), such that \( A \subseteq B \subseteq V\) and \(v\in V\setminus B\). It is the second equivalent definitions of submodularity as shown in Theorem 1 of work [9]. The left side of inequality is the weight of nodes which can be activated by v but can not by A. The right side is the weight of nodes which can be activated by v but can not by B. We have \(I(v) -I(A) \supseteq I(v) - I(B)\), since \(A \subseteq B\). And the submodularity follows immediately. \(\square \)
Lemma 2
If we have \(U(S,\phi ) = f(I(S,\phi ))\), where \(I(S,\phi )\) denotes the set of all activated nodes when S is the seed set selected under the realization \(\phi \), then \(f: 2^V \rightarrow \mathbb {R}_{\ge 0}\) is a monotone submodular function indicating the profit of influenced nodes set .
Proof
Since \( U(S,\phi )=\sum _{v\in I(S,\phi )}w(v) \), and \(U(S,\phi ) = f(I(S,\phi ))\), we have \( f(I(S,\phi )) =\sum _{v\in I(S,\phi )}w(v)\). Since w(v) is a constant for each node v, apparently \(f: 2^V \rightarrow \mathbb {R}_{\ge 0}\) is linear function. And for each node v, its weight \(w(v) \ge 0\). So \(f: 2^V \rightarrow \mathbb {R}_{\ge 0}\) is an monotone submodular function indicating the profit for influencing a set. \(\square \)
Lemma 3
In the full-adoption feedback model, given realizations \(\phi \) and \(\phi '\), and a node \(v\in V\), if \(\phi (v) = \phi '(v)\), then \(I(v, \phi ) = I(v,\phi ')\).
Proof
It means that if the states of a seed v are the same under different realizations, then the nodes that can be activated by it are the same too. For any node \(w \in I(v, \phi )\), there is a live path from v to w under realization \(\phi \). According to the definition of realization, we know that the node v reveals a set of edges denoted as A to be dead or live under realization \(\phi \) and \(\phi '\). And for any edge \((i,j)\in A\), \(\phi _{v}(i,j) =\phi '_{v}(i,j)\) since \( \phi (v) = \phi '(v) \), which means the node v revealed edge (i, j) to the same state under realization \(\phi \) and \(\phi '\) respectively. So if there a live path from v to w under realization \(\phi \), then there a live path from v to w under realization \(\phi '\) too. It follows that \(w \in I(v, \phi ')\) and vice versa. \(\square \)
Lemma 4
In the full-adoption feedback model, given realizations \(\phi \) and partial realization \(\psi \), such that \(\phi \sim \psi \) , then \(I(dom(\psi ), \phi ) = \cup _{x\in dom(\psi )} I(x, \phi ) \)
Proof
Since the realization is complete according to the definition of it, the state of every edge in social network is deterministic which means we get a live graph under the realization \( \phi \). Apparently, in a live graph the influence spread of nodes set \(dom(\psi )\) equals to the union of influenced spread of node in \(dom(\psi )\). The lemma follows immediately. \(\square \)
Lemma 5
In the full-adoption feedback model, given realizations \(\phi \) and \(\phi '\), and partial realization \(\psi \) and \(\psi '\), such that \( \phi \sim \psi \), \(\phi ' \sim \psi ' \), and \(\psi \) is subrealization of \(\psi '\) , then \(I(dom(\psi ), \phi ) = I(dom(\psi ),\phi ')\).
Proof
Since \(\psi \) is subrealization of \(\psi '\), we know that \( \psi \subseteq \psi '\) which means that \(\psi (x) =\psi '(x)\) for all \(x \in dom(\psi )\). Then we get \( \phi (x) = \phi '(x) \), for \(x \in dom(\psi )\), since the consistency of realization and the given fact \( \phi \sim \psi \), \(\phi ' \sim \psi ' \). And based on Lemma 3, we have \(I(x, \phi ) = I(x,\phi ')\), for \(x \in dom(\psi )\). And we have \(I(dom(\psi ), \phi ) = \cup _{x\in dom(\psi )} I(x, \phi ) \), \(I(dom(\psi ), \phi ' ) = \cup _{x\in dom(\psi )} I(x, \phi ') \) according to Lemma 4. Then the lemma follows immediately. \(\square \)
Theorem 5
\(U(S,\phi )\) is adaptive monotone and submodular.
Proof
Since \( U(\cdot , \phi )\) is monotonic for each fixed \(\phi \) according to Lemma 1, adaptive monotonicity is readily proved.
Moving on to adaptive submodularity, we first define a coupled distribution \(\theta (\phi , \phi ')\) over pairs of realizations \(\phi \thicksim \psi \) and \(\phi ' \thicksim \phi '\) using the technique proposed by [10]. We denote the state of edge (u, v) as \(X_{uw} \), for \((u,v)\in E\). Then a realization can be represented by a random variable \( \mathbf{X} = \{X_{uw} : (u,w) \in E\}\). Then \(\theta (\phi ,\phi ')\) can be defined implicitly in terms of a joint distribution \(\hat{\theta }\) on \( \mathbf{X} \times \mathbf{X} '\), where \(\theta = \theta (\mathbf{X} )\) and \(\theta ' = \theta '(\mathbf{X} ')\) are the realizations induced by the two sets of edge statuses, respectively. Hence \( \theta (\phi , \phi ') = \hat{\theta }(\mathbf{X} ,\mathbf{X} ')\). There are two constraints we need to consider in constructing \(\hat{\theta }\). First, the status of all edges which are revealed by \(\psi \) are the same since \(\psi \subseteq \psi '\). Second, the status of all edges which are unobserved by both \(\psi \) and \(\psi '\) are the same in \(\mathbf{X} \) and \(\mathbf{X} '\), meaning \( X_{uw} = X'_{uw}\) for all such edges (u, w). The bove constraints leave us with the following degrees of freedom: we may select \( X_{uw}\) for all \((u,w) \in E\) which are unobserved by \(\psi \).
Hence for all \((\mathbf{X} ,\mathbf{X} ') \) satisfying the above two constraints, we have
and otherwise \(\hat{\theta }(\mathbf{X} ,\mathbf{X} ')= 0\). Note that \(p(\phi \mid \psi )=\varSigma _{\phi '}\theta (\phi ,\phi ')\) and \(p(\phi '\mid \psi ') = \varSigma _{\phi } \theta (\phi ,\phi ').\) And all \((\mathbf{X} ,\mathbf{X} ') \in \text{ support } (\hat{\theta })\) satisfy the two constraints. We next claim that for all \((\phi ,\phi ')\in \text{ support }(\theta )\),
Recall \(U(S,\phi ) =f(I(S,\phi ))\) proposed in Lemma 2, where \(I (S,\phi )\) is the set of all activated nodes when S is the seed set of activated nodes and \(\phi \) is the realization. Let \(B := I(dom(\psi ),\phi )\) and \( C := I(dom(\psi )\cup \{v\}, \phi )\) denote the active nodes before and after selecting v after \(dom(\psi )\) under realization \(\phi \), and similarly define \(B' := I(dom(\psi '),\phi ')\) and \( C' := I(dom(\psi ')\cup \{v\}, \phi ' )\) with respect to \(\psi '\) and \(\psi '\). Let \(D := C\backslash B\), \( D':= C'\backslash B'\). Then Eq. 17 is equivalent to
According to Lemma 2, we know that \(f: 2^V \rightarrow \mathbb {R}_{\ge 0}\) is a monotone submodular function. So it suffices to show that \(B\subseteq B'\) and \(D' \subseteq D\) to prove the above inequality, which we will now do.
We first prove \(B \subseteq B'\). Since \((\phi ,\phi ') \in \text{ support } (\theta )\), this implies that \(\psi \subseteq \psi '\) and \(dom(\psi ) \subseteq dom(\psi ')\). We denote the set \( H = dom(\psi ') -dom(\psi )\). Then \(B' = I(dom(\psi '),\phi ') = I(dom(\psi )\cup H,\phi ') = I(dom(\psi ),\phi ') \cup I(H,\phi ') \) according Lemma 4. And \(B = I(dom(\psi ),\phi ) = I(dom(\psi ),\phi ' ) \) according Lemma 5, so we have \(B \subseteq B'\).
Next we show \(D' \subseteq D\). According to Lemma 4, we have \( D = C - B = I(dom(\psi )\cup \{v\}, \phi ) - I(dom(\psi ),\phi ) = I(v, \phi ) - I(dom(\psi ),\phi )\), and \( D' = C' - B' = I(dom(\psi ')\cup \{v\}, \phi ' ) - I(dom(\psi ' ),\phi ' ) = I(v, \phi ' ) - I(dom(\psi ' ),\phi ' )\). We have \(\phi (v) = \phi '(v)\) since \((\phi ,\phi ')\in \text{ support }(\theta )\), and then \( I(v, \phi ) = I(v, \phi ' )\) according to Lemma 3. And we know that \( I(dom(\psi ),\phi ) \subseteq I(dom(\psi ' ),\phi ' )\) which has been proved above(i.e., \(B \subseteq B'\)), then we have \(D' \subseteq D\).
Having proved Eq. 17, we now proceed to use it to show \(\triangle (v\mid \psi ') \le \triangle (v\mid \psi )\). Since
we have \(\triangle (v\mid \psi ')\le \triangle (v\mid \psi ) \) which completes the proof. \(\square \)
5.2 Lower bound
The major idea is that we only consider the profit between nodes which are activated by the same seed node. Accordingly, the lower bound can be defined as
where \(I(\{x\},\phi )\) are nodes activated by node x under realization \(\phi \). It is easy to see that \(L(S,\phi )\le f(S,\phi )\) for any \(S \subseteq V\) since we ignore the profit of pairs of nodes which are activated by different seeds.
Lemma 6
Under any fixed realization \(\phi \), \(L(S,\phi )\) is monotone and submodular.
Proof
Given a fixed realization \(\phi \) which means the state of all the edges in the social network is determined according to Remark 2, we just need to prove \(L(S,\phi )\) is monotone and submodular in a live graph. Since the profit function \(c: V\times V \rightarrow \mathbb {R}_{\ge 0}\) is nonnegative which means the profit of each pair of nodes is non-negative, the monotonicity of \(L(S,\phi )\) follows immediately. For the submodularity, we need prove \(L(A\cup \{e\}) -L(A) \ge L(B \cup \{e\}) -L(B)\), such that \( A \subseteq B \subseteq V\) and \(e\in V\setminus B\). We have \(I(e) -\cup _{x\in A}I(x) \supseteq I(e) - \cup _{x\in B}I(x)\), since \(A \subseteq B\). Thus \(P(I(e) -\cup _{x\in A}I(x)) \supseteq P(I(e) - \cup _{x\in B}I(x))\), and the submodularity follows immediately. \(\square \)
To prove the lower bound is adaptive submodular, we define a new function
which quantifies the profit \(L(S,\varPhi )\) when seeds S are in arbitrary states. Note \(\hat{f}\) is a set function taking a set of (seeds, state) pairs as input. The profit \(L(S,\phi )\) of choosing seeds S under realization \(\phi \) is then
We make the value of \(\hat{f} \) equals 0, when the status of the seeds are not consistent with each other. For example \(\hat{f}({(v_1,\phi (v_1)),(v_2,\phi '(v_2)) }) =0\), if there is a edge which is declared to be live and dead by \(\phi (v_1)\) and \(\phi '(v_2)\) respectively.
Lemma 7
\(\hat{f} : 2 ^{V\times O}\rightarrow \mathbb {R}_{\ge 0} \) is a monotone submodular function.
Proof
Since \(\hat{f}((x,\phi (x))) = L(x, \phi ) \) is nonnegative, for any \((x,\phi )\). The monotonicity of \(\hat{f}\) follows immediately.
For the submodularity, we need prove \(\hat{f}(A\cup \{x\}) -\hat{f}(A) \ge \hat{f}(B \cup \{x\}) -\hat{f}(B)\), such that \( A \subseteq B \subseteq 2 ^{V\times O}\) and \(x\in 2 ^{V\times O}\setminus B\). Note that x is actually a tuple \((e, \phi (e))\) or \((e , \phi '(e))\) which means that the node e is in state \(\phi (e)\). We have \(A:=\psi \) and \(B:=\psi '\), and \(\phi (e) = \phi '(e)\). We address the problem in two situations. One is called effective domain which means given realizations \(\phi \) and \(\phi '\), and partial realization \(\psi \) and \(\psi '\), such that \( \phi \sim \psi \), \(\phi ' \sim \psi ' \), and \(\psi \) is subrealization of \(\psi '\). The other condition is called non-effective domain where the above condition is not satisfied. Actually it means that there exist some nodes whose states are not consistent with each other.
For the effective domain,
in the same way
According Lemma 3, we have \(I(e,\phi ) = I(e,\phi ')\) since \(\phi (e) = \phi '(e)\). So we get \( P( I(e,\phi ')) = P( I(e,\phi ))\). And since \(\psi \) is subrealization of \(\psi '\), we have \( dom(\psi )\subset dom(\psi ') \), and \(I(x,\phi ) =I(x,\phi ')\) for \(x\in dom(\psi )\). So we know \(\cup _{x\in dom(\psi )}P(I(\{x\},\phi ) \subseteq \cup _{x\in dom(\psi ')}P(I(\{x\},\phi ' ))\). Thus \( P( I(e,\phi '))-\cup _{x\in dom(\psi ')}P(I(\{x\},\phi ' )) \subseteq P( I(e,\phi ))-\cup _{x\in dom(\psi ')}P(I(\{x\},\phi ))\}\). Then we have
For the non-effective domain, we make the value of \(\hat{f} \) equals 0 when the status of some seeds are not consistent with each other according to its definition: if the state of \(\phi '(e)\) is not consistent with \(\psi '\) which means \(\phi ' \sim \psi ' \) is not satisfied and at the same time the state of \(\phi (e)\) is not consistent with \(\psi \) which means \( \phi \sim \psi \) is not satisfied, then \( \hat{f} (\psi '\cup (e,\phi '(e))) = 0\), and \(\hat{f} (\psi \cup (e,\phi (e))) =0\). We have \(\hat{f}(\psi ) \le \hat{f}(\psi ') \), since \( \psi \subseteq \psi '\) and the monotonicity of \(\hat{f} \). Thus
If the state of \(\phi '(e)\) is not consistent with \(\psi '\) which means \(\phi ' \sim \psi ' \) is not satisfied, but the state of \(\phi (e)\) is consistent with \(\psi \), then \( \hat{f} (\psi '\cup (e,\phi '(e))) = 0\) and \(\hat{f} (\psi \cup (e,\phi (e))) \ge 0\). We have \(\hat{f}(\psi ) \le \hat{f}(\psi ') \), since \( \psi \subseteq \psi '\) and the monotonicity of \(\hat{f} \). Thus
which complete the proof. \(\square \)
Theorem 6
\(L(S,\phi )\) is adaptive monotone and submodular.
Proof
Adaptive monotonicity is readily proved after observing that \( L(\cdot , \phi )\) is monotonic for each \(\phi \) according to Lemma 6. Moving on to adaptive submodularity, we define a coupled distribution \(\theta (\phi , \phi ')\) over pairs of realizations \(\phi \thicksim \psi \) and \(\phi ' \thicksim \phi '\) which is the same as Theorem 5. So we have
and otherwise \(\hat{\theta }(\mathbf{X} ,\mathbf{X} ')= 0\). Note that \(p(\phi \mid \psi )=\varSigma _{\phi '}\theta (\phi ,\phi ')\) and \(p(\phi '\mid \psi ') = \varSigma _{\phi } \theta (\phi ,\phi ').\) We next claim that for all \((\phi ,\phi ')\in \text{ support }(\theta )\),
from the submodularity of \(\hat{f}\) according to Lemma 7. We now proceed to use it to show \(\triangle (v\mid \psi ') \le \triangle (v\mid \psi )\).
we have \(\triangle (v\mid \psi ') \le \triangle (v\mid \psi )\) according to (22), which completes the proof. \(\square \)
6 Algorithm
In this section, we propose a greedy policy with lazy evaluation based on submodularity of the upper and lower bound, and design an adaptive sandwich strategy for APMP which can get a data dependent approximation solution. To the best of our knowledge, this is the first paper trying to address the problem that is not adaptive submodular.
6.1 Lazy evaluation
As we know, when adaptive greedy algorithm selects a seed in the round i, it computes \(\varDelta (v|\psi _{i}) \) for all \(v \in V\) and \( 0\le i \le k\), unless \(v \in dom(\psi _{i})\), where \(\psi _{i}\) is the partial realization after round i and k is constraint on number of seeds. The key insight is that \(i \mapsto \varDelta (v|\psi _{i})\) is nonincreasing for all \( v\in V\), because of the adaptive submodularity of the objective. Specifically speaking, in the round i of selecting a seed, if we know \(\varDelta (v'|\psi _{j}) \le \varDelta (v|\psi _{i})\) for some items \(v'\) and v and \(j \le i\), then \(\varDelta (v'|\psi _{i}) \le \varDelta (v|\psi _{i})\) follows, since \(\varDelta (v'|\psi _{j}) \ge \varDelta (v'|\psi _{i})\) based on the adaptive submodularity. So there is no need to compute \(\varDelta (v'|\psi _{i})\) in the round i.
6.2 Accelerated adaptive greedy algorithm
Pseudocode of the adaptive greedy algorithm with lazy evaluation is given in Algorithm 1. Its core idea is to choose the seed with the most expected marginal benefit at each step. S is used to store the seeds selected, T is the profit gained by the seeds in S. The expected marginal gain of each node is stored in the list Q. According to the strategy of lazy evaluation, the while loop computes conditional expected marginal benefit \(\varDelta (v|\psi )\) for node v in decreasing order of upper bounds knows on them, until it finds an item whose value is at least as great as the upper bounds of all other nodes. Note the Q.max( ) return max value in the marginal gain list Q, and Max(Q) return the node which has the maximum marginal gain. When a seed with the greatest marginal benefit is found, we observe its realization denoted by \(\psi '\) as shown in line 15. After observing the realization, we need to compute the actual marginal benefit of node \(v_{max}\) denoted by \( \delta _{v_{max}}^{\psi '}\) under partial realization \(\psi '\) as shown in line 16. We will delete node \(v_{max}\) and the nodes activated by it as shown in line 20 since their marginal gain will be 0 in the next round. The greedy policy will return the seeds and its profit.
6.3 Adaptive sandwich approximation policy
By adopting the sandwich strategy which is first proposed by [16], and using the proving technique introduced by [26], we design an adaptive greedy sandwich approximation policy for adaptive profit maximization problem, which is given in Algorithm 2. The basic idea of the sandwich strategy is that we find an adaptive greedy strategy based on the upper bound, the lower bound, and original function respectively. Then we choose the best greedy strategy for the original problem as the solution. Specifically, the adaptive sandwich approximation strategy works as follows. First, according to the upper bound we find a greedy strategy by calling Algorithm 1. When selecting the seed, we need to calculate the expected marginal gain \(\triangle (v|\psi )\) according to the upper bound function as shown in line 9, i.e., \(\triangle (v|\psi ) = \mathbb {E}[U(dom(\psi ) \cup {v}, \varPhi ) - U(dom(\psi ), \varPhi ) |\varPhi \sim \psi ] \). Note that after a node with the largest expected return is found, we need to calculate the actual marginal benefit of the node based on the original function instead of upper bound, i.e., \( \delta _{v_{max}}^{\psi '}=f(v_{max}\cup S|\psi ')-f(S|\psi )\) as shown in line 16. This is a very important point. Because our purpose is choosing the best strategy for the original function, we need to evaluate the strategy with benefits of the original function. In the same way, for the lower bound we calculate \(\triangle (v|\psi )\) according to the lower bound function, i.e., \(\triangle (v|\psi ) = \mathbb {E}[L(dom(\psi ) \cup {v}, \varPhi ) - L(dom(\psi ), \varPhi ) |\varPhi \sim \psi ] \). And calculate the actual marginal benefit of the node based on the original function instead of lower bound, i.e., \( \delta _{v_{max}}^{\psi '}=f(v_{max}\cup S|\psi ')-f(S|\psi )\).
6.4 Approximation ratio
Theorem 7
Let \(\hat{\pi }_{max}\) be the policy return by Algorithm 2, then we have
where \(\pi _{f}^{*}\), \(\pi _{U}^{*}\), \(\pi _{L}^{*}\) is the optimal policy for \(f(\cdot )\), \(U(\cdot )\), \(L(\cdot )\) respectively, and \(\alpha = \beta = 1-1/e.\)
Proof
We have
Let \(\pi _{max}= \mathrm{argmax}_{\pi \in \{ \pi _{f},\pi _{U},\pi _{L} \}}f_{avg}(\pi ) \), then
Since \(\forall \pi \in \{ \pi _{f},\pi _{U},\pi _{L} \} \), \(|\hat{f}_{avg}(\pi )-f_{avg}(\pi ) | \le \gamma f_{avg}(\pi ) \), we have \((1+\gamma )f_{avg}(\hat{\pi }_{max})\ge (1-\gamma )f_{avg}(\pi _{max})\). It follows that
\(\square \)
6.5 Implementation
To implement the proposed adaptive greedy algorithm, the key step is to calculate the conditional expected marginal gain \(\triangle (v|\psi )\) for the upper bound, lower bound, and original function. Unfortunately, all of them are \(\sharp \)P-hard.
Theorem 8
Given a seed v and partial realization \(\psi \), computing conditional expected marginal gain \(\triangle (v|\psi )\) for upper bound is \(\sharp \) P-hard.
Proof
We prove the theorem through a special case in which the partial realization is \(\emptyset \). Since \(\psi \) is empty set, we have \(\triangle (v |\psi ) = \mathbb {E}[U({v}, \varPhi )]\), where the expectation is taken with respect to \(p(\phi )\). Apparently, it equals to compute the upper bound profit of node v in non-adaptive settings under the IC model. This is actually weighted influence spread computation problem which was proved \(\sharp P\)-hard [2], the theorem follows immediately. \(\square \)
Theorem 9
Given a seed v and partial realization \(\psi \), computing conditional expected marginal benefit \(\triangle (v | \psi ) \) for lower bound is \(\sharp \) P-hard.
Proof
We prove the theorem through a special case in which the partial realization is \(\emptyset \). Since \(\psi \) is empty set , we have \(\triangle (v|\psi ) = \mathbb {E}[L({v}, \varPhi )]\), where the expectation is taken with respect to \(p(\phi )\). Apparently, it equals to the problem about computing the lower bound profit of node v in non-adaptive settings. According to Remark 1 we can reduce it from AMP problem, and computing the lower bound of AMP is \(\sharp P\)-hard which is proved in [26], the theorem follows immediately. \(\square \)
Theorem 10
Given a seed v and partial realization \(\psi \), computing conditional expected marginal benefit \(\triangle (v | \psi ) \) for original problem is \(\sharp \) P-hard.
Proof
We prove the theorem through a special case in which the partial realization is \(\emptyset \). Since \(\psi \) is empty set, we have \(\triangle (v \mid \psi ) := \mathbb {E}[f({v}, \varPhi )]\), where the expectation is taken with respect to \(p(\phi )\). Apparently, it equals to the game profit computing problem in non-adaptive settings which is \(\sharp \)P-hard according to Theorem 2, the theorem follows immediately. \(\square \)
Although they are difficult to calculate, we can employ the Monte Carlo simulation to obtain an accurate estimation. By the Hoeffding’s Inequality, the error of the estimation can be infinitely small when a sufficient number of simulations are performed.
7 Experiment
7.1 Settings
We use four social networks in our experiments. All datasets are publicly available. Email, DBLP can be obtained from SNAP website (http://snap.stanford.edu), while Facebook and Douban can be obtained from KONECT website (http://konect.uni-koblenz.de). The details of datasets are shown in Table 1.
The propagation probability for IC model is set to \(\frac{1}{degree(v)}\) as widely used in other literatures [2, 22], and the profit between nodes is proportional to propagation probability on corresponding edges. For the nodes that are not directly adjacent, we set the profit between them as 0. For comparison, we propose two other algorithms as baselines. One is non-adaptive greedy algorithm in which we select k seeds in advance regardless of their realization. We denote it as Non in Figs. 2, 3 and 4. The other one is a random algorithm which selects the next node in a random way based on the current realization. We mark it as Random in Figs. 2, 3 and 4.
For non-adaptive greedy algorithm, we actually select k seeds in one step without observing the realization of any seed. So the non-adaptive greedy algorithm is similar to the greedy algorithm propose by Kempe [14] except that our objective function is the profit between activated node while their objective function is the number of the activated nodes. For random algorithm, it selects the next node v in a random way based on the current realization.We can employ the Monte Carlo simulation to obtain an estimation. As shown in [14] 10k iterations could get a acceptable quality of approximation. We simulate the process 10k times for each targeted node v and get the average value for conditional expected marginal gain\(\triangle (v|\psi )\). For the adaptive policy value we use the same idea simulating the realization 1k times and get the average value. We implement all the algorithms in Python and the experiments run on a workstation with an Intel Xeon 4.0 GHz CPU and 64 GB memory.
7.2 Effectiveness and analysis
The results of profit computed by our proposed algorithms on four data sets are shown in Fig.2, 3 and 4 respectively. As the number of selected seeds increases, the performance of sandwich algorithm donated as Adapt in the figure is always superior to the baseline algorithms. This is because non-adaptive greedy algorithm ignores the stochastics of influence propagation and always chooses the same seed set under any realization. And our adaptive strategy must make a decision based on the actual situation of current network propagation when selecting a seed. Moreover, his submodular upper bound and lower bound guarantee its performance. An interesting question in adaptive optimization is how much better adaptive policies can perform when compared to non-adaptive policies. This is quantified by the adaptivity gap, and the gap as shown in Fig.2, 3 and 4 between the two algorithms is a good illustration of the effectiveness of adaptive strategy in social network impact issues. Not surprisingly, the performance of a random algorithm is the worst because it does not consider any validity or propensity when making a choice. On both small-scale and large-scale networks, the adaptive sandwich algorithm perform quite well and demonstrate good scalability.
8 Conclusion
In this paper, we propose the adaptive profit maximization problem which is not adaptive submodular, and analyze its complexity which is NP-hard. To address the stochastic optimization problem, we find the upper and lower bounds of it and prove both of them are adaptive submodular, so an adaptive greedy algorithm could gain a guaranteed approximate solution for them. Based on the upper and lower bound, we design an adaptive sandwich policy for APMP to get a data dependent approximate solution. To the best of our knowledge, this is the first paper applying the sandwich strategy to stochastic optimization problem which is not adaptive submodular.
References
Asadpour, A., Nazerzadeh, H., Saberi, A.: Stochastic submodular maximization. In: Proceedings of the 4th International Workshop on Internet and Network Economics, WINE ’08, pp. 477–489. Springer-Verlag, Berlin, Heidelberg (2008). https://doi.org/10.1007/978-3-540-92185-1_53
Chen, W., Wang, C., Wang, Y.: Scalable influence maximization for prevalent viral marketing in large-scale social networks. In: International Conference on Knowledge Discovery and Data Mining, KDD ’10, pp. 1029–1038. ACM (2010). https://doi.org/10.1145/1835804.1835934
Chen, W., Yuan, Y., Zhang, L.: Scalable influence maximization in social networks under the linear threshold model. In: 2010 IEEE International Conference on Data Mining, pp. 88–97 (2010). https://doi.org/10.1109/ICDM.2010.118
Chen, W., Lin, T., Tan, Z., Zhao, M., Zhou, X.: Robust influence maximization. In: International Conference on Knowledge Discovery and Data Mining, KDD ’16, pp. 795–804. ACM, New York, NY, USA (2016). https://doi.org/10.1145/2939672.2939745
Chen, Y., Shioi, H., Montesinos, C.F., Koh, L.P., Wich, S., Krause, A.: Active detection via adaptive submodularity. In: ICML, pp. 55–63 (2014)
Chen, Y., Javdani, S., Karbasi, A., Bagnell, J.A., Srinivasa, S.S., Krause, A.: Submodular surrogates for value of information. In: AAAI, pp. 3511–3518 (2015)
Gabillon, V., Kveton, B., Wen, Z., Eriksson, B., Muthukrishnan, S.: Adaptive submodular maximization in bandit setting. In: International Conference on Neural Information Processing Systems, NIPS’13, pp. 2697–2705. Curran Associates Inc., USA (2013). http://dl.acm.org/citation.cfm?id=2999792.2999913
Gabillon, V., Kveton, B., Wen, Z., Eriksson, B., Muthukrishnan, S.: Large-scale optimistic adaptive submodularity. In: AAAI, pp. 1816–1823 (2014)
Goldengorin, B.: Maximization of submodular functions: theory and enumeration algorithms. Eur. J. Oper. Res. 198, 102–112 (2009)
Golovin, D., Krause, A.: Adaptive submodularity: theory and applications in active learning and stochastic optimization. J. Artif. Intell. Res. 42, 427–486 (2011)
Gotovos, A., Karbasi, A., Krause, A.: Non-monotone adaptive submodular maximization. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (2015)
Hatano, D., Fukunaga, T., Kawarabayashi, K.I.: Adaptive budget allocation for maximizing influence of advertisements. In: IJCAI, pp. 3600–3608 (2016)
Horel, T., Singer, Y.: Scalable methods for adaptively seeding a social network. In: International Conference on World Wide Web, WWW ’15, pp. 441–451 (2015). https://doi.org/10.1145/2736277.2741127
Kempe, D., Kleinberg, J., Tardos, E.: Maximizing the spread of influence through a social network. In: International Conference on Knowledge Discovery and Data Mining, KDD ’03, pp. 137–146. ACM (2003). https://doi.org/10.1145/956750.956769
Krause, A., Golovin, D., Converse, S.: Sequential decision making in computational sustainability via adaptive submodularity. Ai Mag. 35(2), 8–18 (2014)
Lu, W., Chen, W., Lakshmanan, L.V.S.: From competition to complementarity: comparative influence diffusion and maximization. Proc. VLDB Endow. 9(2), 60–71 (2015). https://doi.org/10.14778/2850578.2850581
Nemhauser, G.L., Wolsey, L.A.: Best algorithms for approximating the maximum of a submodular set function. Math. Oper. Res. 3(3), 177–188 (1978). https://doi.org/10.1287/moor.3.3.177
Nguyen, H.T., Dinh, T.N., Thai, M.T.: Cost-aware targeted viral marketing in billion-scale networks. In: The 35th Annual IEEE International Conference on Computer Communications, INFOCOM 2016, pp. 1–9 (2016). https://doi.org/10.1109/INFOCOM.2016.7524377
Nguyen, H.T., Thai, M.T., Dinh, T.N.: Stop-and-stare: Optimal sampling algorithms for viral marketing in billion-scale networks. In: International Conference on Management of Data, SIGMOD ’16, pp. 695–710. ACM (2016). https://doi.org/10.1145/2882903.2915207
Seeman, L., Singer, Y.: Adaptive seeding in social networks. In: 2013 IEEE 54th Annual Symposium on Foundations of Computer Science, pp. 459–468 (2013). https://doi.org/10.1109/FOCS.2013.56
Tang, J., Tang, X., Yuan, J.: Towards profit maximization for online social network providers. CoRR (2017). arXiv:1712.08963
Tang, Y., Xiao, X., Shi, Y.: Influence maximization: near-optimal time complexity meets practical efficiency. In: International Conference on Management of Data, SIGMOD ’14, pp. 75–86. ACM (2014). https://doi.org/10.1145/2588555.2593670
Tang, Y., Shi, Y., Xiao, X.: Influence maximization in near-linear time: a martingale approach. In: Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, SIGMOD ’15, pp. 1539–1554. ACM (2015). https://doi.org/10.1145/2723372.2723734
Tong, G., Wu, W., Tang, S., Du, D.Z.: Adaptive influence maximization in dynamic social networks. IEEE/ACM Trans. Netw. 25(1), 112–125 (2017)
Wang, A., Wu, W., Cui, L.: On Bharathi–Kempe–Salek conjecture for influence maximization on arborescence. J. Comb. Optim. 31(4), 1678–1684 (2016)
Wang, Z., Yang, Y., Pei, J., Chu, L., Chen, E.: Activity maximization by effective information diffusion in social networks. IEEE Trans. Knowl. Data Eng. 29(11), 2374–2387 (2017). https://doi.org/10.1109/TKDE.2017.2740284
Xue, Z., Wu, D., He, J., Hei, X., Liu, Y.: Playing high-end video games in the cloud: a measurement study. IEEE Trans. Circuits Syst. Video Technol. 25(12), 2013–2025 (2015). https://doi.org/10.1109/TCSVT.2014.2364419
Yuan, J., Tang, S.J.: Adaptive discount allocation in social networks. In: Proceedings of the 18th ACM International Symposium on Mobile Ad Hoc Networking and Computing, p. 22. ACM (2017)
Funding
Funding was provided by National Natural Science Foundation of China (Grant Nos. 61832012, 61771289 and 61672321).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Gao, C., Gu, S., Yu, J. et al. Adaptive seeding for profit maximization in social networks. J Glob Optim 82, 413–432 (2022). https://doi.org/10.1007/s10898-021-01076-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10898-021-01076-1