Abstract
Maximizing a submodular function has a wide range of applications in machine learning and data mining. One such application is data summarization whose goal is to select a small set of representative and diverse data items from a large dataset. However, data items might have sensitive attributes such as race or gender, in this setting, it is important to design fairness-aware algorithms to mitigate potential algorithmic bias that may cause over- or under- representation of particular groups. Motivated by that, we propose and study the classic non-monotone submodular maximization problem subject to novel group fairness constraints. Our goal is to select a set of items that maximizes a non-monotone submodular function, while ensuring that the number of selected items from each group is proportionate to its size, to the extent specified by the decision maker. We develop the first constant-factor approximation algorithms for this problem. We also extend the basic model to incorporate an additional global size constraint on the total number of selected items.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Submodular function refers to a broad class of functions which satisfy the natural diminishing returns property: adding an additional item to a larger existing subset is less beneficial. A wide range of machine learning and AI problems, including exemplar-based clustering (Dueck and Frey 2007), feature selection (Das and Kempe 2008), active learning (Golovin and Krause 2011), influence maximization in social networks (Tang and Yuan 2020), recommender system (El-Arini and Guestrin 2011), and diverse data summarization (Sipos et al. 2012), can be formulated as a submodular maximization problem. This problem, whose goal is to select a set of items to maximize a submodular function, and its variants (Gu et al. 2022; Shi et al. 2021) have been extensively studied in the literature subject to various constraints, including cardinality, matroid, or knapsack-type restrictions.
We notice that in practise, items or individuals are often associated with different groups based on various attributes, such as gender, race, age, religion, or other factors. Existing algorithms might exhibit bias if left unchecked, for example, some of the groups might be over- or under-represented in the final selected subset. Therefore, it becomes increasingly important to design fairness-aware algorithms to mitigate such issues. Towards this end, we propose and study the classic non-monotone submodular maximization problem subject to novel group fairness constraints. Our goal is to select a balanced set of items that maximizes a non-monotone submodular function, such that the ratio of selected items from each group to its size is within a desired range, as determined by the decision maker. Non-monotone submodular maximization has multiple compelling applications, such as feature selection (Das and Kempe 2008), profit maximization (Tang and Yuan 2021), maximum cut (Gotovos et al. 2015) and data summarization (Mirzasoleiman et al. 2016). Formally, we consider a set V of items (e.g., datapoints) which are partitioned into m groups: \(V_1, V_2, \cdots , V_m\) such that items from the same group share same attributes (e.g., gender). We say that a set \(S\subseteq V\) of items is \((\alpha ,\beta )\)-fair if for all groups \(i \in [m]\), it holds that \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i| \le \lfloor \beta |V_i|\rfloor \). Using our model, it allows for the decision maker to specify the desired level of fairness by setting appropriate values of \(\alpha \) and \(\beta \). Specifically, setting \(\alpha =\beta \) leads to the highest level of fairness in that the number of selected items is strictly proportional to its group size; if we set \(\alpha =0\) and \(\beta =1\), there are no fairness constraints. Our goal is to find such a \((\alpha ,\beta )\)-fair subset of items that maximizes a submodular objective function. Our definition of fairness, which balances solutions with respect to sensitive attributes, has gained widespread acceptance in the academic community, as demonstrated by its frequent use in previous studies (Celis et al. 2018b; El Halabi et al. 2020; Chierichetti et al. 2019). There are several other notations of fairness that can be captured by our formulation such as the \(80\%\)-rule (Biddle 2017), statistical parity (Dwork et al. 2012) and proportional representation (Monroe 1995).
1.1 Our contributions
-
Our study breaks new ground by examining the classic (non-monotone) submodular maximization problem under \((\alpha ,\beta )\)-fairness constraints. Our model offers flexibility in capturing varying degrees of fairness as desired by the decision maker, by adjusting the values of \(\alpha \) and \(\beta \).
-
We develop the first constant-factor approximation algorithm for this problem. We observe that the parameter \(\alpha \) is closely linked to the complexity of solving the \((\alpha ,\beta )\)-fair non-monotone submodular maximization problem. In particular, when \(\alpha \le 1/2\), we design a \(\frac{\gamma }{2}\)-approximation algorithm and when \(\alpha > 1/2\), we develop a \(\frac{\gamma }{3}\)-approximation algorithm, where \(\gamma \) is the approximation ratio of the current best algorithm for matroid-constrained submodular maximization. We also extend the basic model to incorporate an additional global size constraint on the total number of selected items. We provide approximation algorithms that have a constant-factor approximation ratio for this extended model.
1.2 Additional related works
In recent years, there has been a growing awareness of the importance of fair and unbiased decision-making systems. This has led to an increased interest in the development of fair algorithms in a wide range of applications, including influence maximization (Tsang et al. 2019), classification (Zafar et al. 2017), voting (Celis et al. 2018b), bandit learning (Joseph et al. 2016), and data summarization (Celis et al. 2018a). Depending on the specific context and the type of bias that one is trying to mitigate, existing studies adopt different metrics of fairness. This can lead to different optimization problems and different fair algorithms that are tailored to the specific requirements of the application. Our notation of fairness is general enough to capture many existing notations such as the \(80\%\)-rule (Biddle 2017), statistical parity (Dwork et al. 2012) and proportional representation (Monroe 1995). Unlike most of existing studies on fair submodular maximization (Celis et al. 2018b) whose objective is to maximize a monotone submodular function, (El Halabi et al. 2020) develop fair algorithms in the context of streaming non-monotone submodular maximization. Their proposed notation of fairness is more general than ours, leading to a more challenging optimization problem which does not admit any constant-factor approximation algorithms. Tang et al. (2023), Tang and Yuan (2023) aim to develop randomized algorithms that satisfy average fairness constraints. Very recently, Tang and Yuan (2022) extend the studies of fair algorithms to a more complicated adaptive setting and they propose a new metric of fairness called group equality.
2 Preliminaries and problem statement
We consider a set V of n items. There is a non-negative submodular utility function \(f: 2^V\rightarrow {\mathbb {R}}_+\). Denote by \(f(e\mid S)\) the marginal utility of \(e\in V\) on top of \(S\subseteq V\), i.e., \(f(e\mid S)=f(\{e\}\cup S)-f(S)\). We say a function \(f: 2^V\rightarrow {\mathbb {R}}_+\) is submodular if for any two sets \(X, Y\subseteq V\) such that \(X\subseteq Y\) and any item \(e \in V\setminus Y\),
Assume V is partitioned into m disjoint groups: \(V_1, V_2,\cdots , V_m\). We assume that there is a given lower and upper bound on the fraction of items of each group that must be contained in a feasible solution. These two bounds, namely \(\alpha \) and \(\beta \), represent group fairness constraints. The problem of \((\alpha , \beta )\)-fair submodular maximization problem (labelled as \({\textbf {P.0}}\)) can be written as follows.
One can adjust the degree of group fairness in a feasible solution through choosing appropriate values of \(\alpha \) and \(\beta \). I.e., strict group fairness is achieved at \(\alpha =\beta \) in which case every feasible solution must contain the same \(\alpha \) fraction of items from each group; if we set \(\alpha =0\) and \(\beta =1\), then there is no group fairness constraints. We next present the hardness result of this problem.
Lemma 1
Problem \({\textbf {P.0}}\) is NP-hard.
Proof
We prove this through reduction to the classic cardinality constrained submodular maximization problem which we define below.
Definition 1
The input of cardinality constrained submodular maximization problem is a group of items U, a submodular function \(h: 2^U\rightarrow {\mathbb {R}}_+\), and a cardinality constraint b; we aim to select a group of items \(S\subseteq U\) such that h(S) is maximized and \(|S|\le b\).
We next show a reduction from cardinality constrained submodular maximization problem to \({\textbf {P.0}}\). Consider any given instance of cardinality constrained submodular maximization problem, we construct a corresponding instance of \({\textbf {P.0}}\) as follows: Let \(V=U\), \(f=h\), assume there is only one group, i.e., \(V=V_1\), and let \(\alpha =0\), \(\beta =b/|U|\). It is easy to verify that these two instances are equivalent. This finishes the proof of the reduction. \(\Box \)
3 Non-monotone submodular maximization with group fairness
Warm-up: Monotone Utility Function If f is monotone and submodular, we can easily confirm that \({\textbf {P.0}}\) can be simplified to \({\textbf {P.1}}\) by removing the lower bound constraints. This is because in this case, increasing the size of a solution by adding more items will not decrease its utility. As a result, the lower bound constraints in \({\textbf {P.0}}\), which state that \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i|\) for all \(i \in [m]\), can always be met by adding sufficient items to the solution.
Since f is a monotone submodular function, \({\textbf {P.1}}\) is a well-known problem of maximizing a monotone submodular function subject to matroid constraintsFootnote 1. This problem has a \((1-1/e)\)-approximation algorithm.
We then proceed to develop approximation algorithms for non-monotone functions. We will examine two scenarios, specifically when \(\alpha \le 1/2\) and when \(\alpha > 1/2\).
3.1 The case when \(\alpha \le 1/2\)
In the scenario where \(\alpha \le 1/2\), we use the solution of \({\textbf {P.1}}\) as a building block to construct our algorithm. First, it is easy to verify that \({\textbf {P.1}}\) is a relaxed version of \({\textbf {P.0}}\) with lower bound constraints \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i|\) in \({\textbf {P.0}}\) being removed. Because f is a submodular function, \({\textbf {P.1}}\) is a classic problem of maximizing a (non-monotone) submodular function subject to matroid constraints. There exist effective solutions for \({\textbf {P.1}}\). Now we are ready to present the design of our algorithm as below.
-
1.
Apply the state-of-the-art algorithm \({\mathcal {A}}\) for matroid constrained submodular maximization to solve \({\textbf {P.1}}\) and obtain a solution \(A^{P.1}\).
-
2.
Note that \(A^{P.1}\) is not necessarily a feasible solution to \({\textbf {P.0}}\) because it might violate the lower bound constraints \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i|\) for some groups. To make it feasible, we add additional items to \(A^{P.1}\). Specifically, for each group \(i\in [m]\) such that \(|A^{P.1} \cap V_i|< \lfloor \alpha |V_i|\rfloor \), our algorithm selects a backup set \(B_i\) of size \(\lfloor \alpha |V_i|\rfloor -|A^{P.1} \cap V_i|\), by randomly sampling \(\lfloor \alpha |V_i|\rfloor -|A^{P.1} \cap V_i|\) items from \(V_i\setminus A^{P.1}\). Define \(B_i=\emptyset \) if \(|A^{P.1} \cap V_i|\ge \lfloor \alpha |V_i|\rfloor \).
-
3.
At the end, add \(\cup _{i\in [m]}B_i\) to \(A^{P.1}\) to build the final solution \(A^{approx}\), i.e., \(A^{approx}= A^{P.1}\cup (\cup _{i\in [m]}B_i)\).
The pseudocode of this approximation algorithm is given as Algorithm 1. Observe that \(A^{P.1}\) is a feasible solution to \({\textbf {P.1}}\), hence, \(A^{P.1}\) satisfies upper bound constraints of \({\textbf {P.1}}\) and hence \({\textbf {P.0}}\), i.e., \(|S \cap V_i| \le \lfloor \beta |V_i|\rfloor , \forall i \in [m]\). According to the construction of \(B_i\), it is easy to verify that adding \(\cup _{i\in [m]}B_i\) to \(A^{P.1}\) does not violate the upper bound constraints because \(\cup _{i\in [m]}B_i\) are only supplemented to those groups which do not satisfy the lower bound constraints of \({\textbf {P.0}}\), i.e., \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i|\). Moreover, adding \(\cup _{i\in [m]}B_i\) to \(A^{P.1}\) makes it satisfy lower bound constraints of \({\textbf {P.0}}\). Hence, \(A^{approx}\) is a feasible solution to \({\textbf {P.0}}\).
Lemma 2
\(A^{approx}\) is a feasible solution to \({\textbf {P.0}}\).
3.1.1 Performance analysis
We next analyze the performance of Algorithm 1. We first introduce a useful lemma from Buchbinder et al. (2014).
Lemma 3
If f is submodular and S is a random subset of V, such that each item in V is contained in S with probability at most p, then \({\mathbb {E}}_S[f(S)] \ge (1-p)f(\emptyset )\).
The next lemma states that if \(A^{P.1}\) is a \(\gamma \)-approximate solution of \({\textbf {P.1}}\), then \(f(A^{P.1})\) is at least \(\gamma \) fraction of the optimal solution of \({\textbf {P.0}}\).
Lemma 4
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint. Let OPT denote the optimal solution of \({\textbf {P.0}}\), we have \(f(A^{P.1})\ge \gamma f(OPT)\).
Proof
Because \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint, we have \(f(A^{P.1})\ge \gamma f(O^{P.1})\) where \(O^{P.1}\) denotes the optimal solution of \({\textbf {P.1}}\). Moreover, because \({\textbf {P.1}}\) is a relaxed version of \({\textbf {P.0}}\), we have \( f(O^{P.1})\ge f(OPT)\). Hence, \(f(A^{P.1})\ge \gamma f(OPT)\). \(\Box \)
We next show that augmenting \(A^{P.1}\) with items from the random set \(\cup _{i\in [m]}B_i\) reduces its utility by a factor of at most 1/2 in expectation. Here the expectation is taken over the distribution of \(\cup _{i\in [m]}B_i\).
Lemma 5
Suppose \(\alpha \le 1/2\), we have \({\mathbb {E}}_{A^{approx}}[f(A^{approx})]\ge \frac{1}{2}f(A^{P.1})\) where \(A^{approx}= A^{P.1}\cup (\cup _{i\in [m]}B_i)\).
Proof
Recall that \(B_i= \emptyset \) for all \(i\in [m]\) such that \(|A^{P.1} \cap V_i|\ge \lfloor \alpha |V_i|\rfloor \), hence, adding those \(B_i\) to \(A^{P.1}\) does not affect its utility. In the rest of the proof we focus on those \(B_i\) with
Recall that for every \(i\in [m]\) such that \(|A^{P.1} \cap V_i|< \lfloor \alpha |V_i|\rfloor \), \(B_i\) is a random set of size \(\lfloor \alpha |V_i|\rfloor -|A^{P.1} \cap V_i|\) that is sampled from \(V_i\setminus A^{P.1}\). It follows that each item in \(V_i\setminus A^{P.1}\) is contained in \(B_i\) with probability at most
We next give an upper bound of (2). First,
where the second inequality is by the assumption that \(\alpha \le 1/2\). Moreover,
where the inequality is by the assumption that \(\alpha \le 1/2\).
Hence,
where the first inequality is by (5); the second inequality is by (3) and the assumption that \(\lfloor \alpha |V_i|\rfloor -|A^{P.1} \cap V_i|>0\) (listed in (1)). That is, the probability that each item in \(V_i\setminus A^{P.1}\) is contained in \(B_i\) is at most 1/2. It follows that the probability that each item in \(V \setminus A^{P.1}\) is contained in \(\cup _{i\in [m]}B_i\) is at most 1/2. Moreover, Lemma 3 states that if f is submodular and S is a random subset of V, such that each item in V appears in S with probability at most p, then \({\mathbb {E}}_A[f(A)] \ge (1-p)f(\emptyset )\). With the above discussion and the observation that \(f(A^{P.1}\cup \cdot )\) is submodular, it holds that \({\mathbb {E}}_{A^{approx}}[f(A^{approx})]={\mathbb {E}}_{\cup _{i\in [m]}B_i}[f(A^{P.1}\cup (\cup _{i\in [m]}B_i))]\ge (1-\frac{1}{2})f(A^{P.1}\cup \emptyset )=\frac{1}{2}f(A^{P.1})\). \(\Box \)
Our main theorem as below follows from Lemma 4 and Lemma 5.
Theorem 1
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint and \(\alpha \le 1/2\), we have \({\mathbb {E}}_{A^{approx}}[f(A^{approx})]\ge \frac{\gamma }{2}f(OPT)\).
One option of \({\mathcal {A}}\) is the continuous double greedy algorithm proposed in Feldman et al. (2011) which gives a \(1/e-o(1)\)-approximation solution, that is, \(\gamma \ge 1/e-o(1)\). This, together with Theorem 1, implies that \({\mathbb {E}}_{A^{approx}}[f(A^{approx})]\ge \frac{1/e-o(1)}{2}f(OPT)\).
3.2 The case when \(\alpha > 1/2\)
We next consider the case when \(\alpha > 1/2\). We first introduce a new utility function \(g: 2^V\rightarrow {\mathbb {R}}_+\) as below:
We first present a well-known result, which states that submodular functions maintain their submodularity property when taking their complement.
Lemma 6
If f is submodular, then g must be submodular.
With utility function g, we present a new optimization problem \({\textbf {P.2}}\) as below:
\({\textbf {P.2}}\) is a flipped version of the original problem \({\textbf {P.0}}\) in the sense that if there is a \(\gamma \)-approximate solution \(A^{P.2}\) to \({\textbf {P.2}}\), it can be easily verified that \(V\setminus A^{P.2}\) is a \(\gamma \)-approximate solution to \({\textbf {P.0}}\). As a result, we will focus on solving \({\textbf {P.2}}\) for the rest of this section.
To solve \({\textbf {P.2}}\), we introduce another problem (labeled as \({\textbf {P.3}}\)) as follows:
\({\textbf {P.3}}\) is relaxed version of \({\textbf {P.2}}\) with lower bound constraints \(|V_i|-\lfloor \beta |V_i|\rfloor \le |S \cap V_i|\) in \({\textbf {P.2}}\) being removed. Because g is a submodular function, \({\textbf {P.3}}\) is a classic problem of maximizing a submodular function subject to matroid constraints. Now we are ready to present the design of our algorithm.
-
1.
Apply the state-of-the-art algorithm \({\mathcal {A}}\) for matroid constrained submodular maximization to solve \({\textbf {P.3}}\) and obtain a solution \(A^{P.3}\).
-
2.
Note that \(A^{P.3}\) is not necessarily a feasible solution to \({\textbf {P.2}}\) because it might violate the lower bound constraints \(|V_i|-\lfloor \beta |V_i|\rfloor \le |S \cap V_i|\) for some groups. We add additional items to \(A^{P.3}\) to make it feasible. Specifically, for each group \(i\in [m]\) such that \(|A^{P.3} \cap V_i|< |V_i|-\lfloor \beta |V_i|\rfloor \), our algorithm selects a backup set \(B_i\) of size \(|V_i|-\lfloor \beta |V_i|\rfloor -|A^{P.3} \cap V_i|\), by randomly sampling \(|V_i|-\lfloor \beta |V_i|\rfloor -|A^{P.3} \cap V_i|\) items from \(V_i\setminus A^{P.3}\). Define \(B_i=\emptyset \) if \(|A^{P.1} \cap V_i|\ge |V_i|-\lfloor \beta |V_i|\rfloor \).
-
3.
Add \(\cup _{i\in [m]}B_i\) to \(A^{P.3}\) to build \(A^{approx}\), i.e., \(A^{approx}= A^{P.3}\cup (\cup _{i\in [m]}B_i)\). Return \(V\setminus A^{approx}\) as the final solution.
The pseudocode of this approximation algorithm is given as Algorithm 2. Observe that \(A^{P.3}\) satisfies upper bound constraints of \({\textbf {P.3}}\) and hence \({\textbf {P.2}}\) because \(A^{P.3}\) is a feasible solution to \({\textbf {P.3}}\). According to the construction of \(B_i\), adding \(\cup _{i\in [m]}B_i\) to \(A^{P.1}\) does not violate the upper bound constraints because \(\cup _{i\in [m]}B_i\) are added to meet the lower bound constraints of \({\textbf {P.2}}\) if necessary. Moreover, adding \(\cup _{i\in [m]}B_i\) to \(A^{P.3}\) makes it satisfy lower bound constraints of \({\textbf {P.2}}\). Hence, \(A^{approx}\) is a feasible solution to \({\textbf {P.2}}\).
Lemma 7
\(A^{approx}\) is a feasible solution to \({\textbf {P.2}}\).
3.2.1 Performance analysis
We first introduce a technical lemma which states that if \(A^{P.3}\) is a \(\gamma \)-approximate solution of \({\textbf {P.3}}\), then \(f(A^{P.3})\) is at least \(\gamma \) fraction of the optimal solution of \({\textbf {P.2}}\). This lemma follows from the observation that \({\textbf {P.3}}\) is a relaxation of \({\textbf {P.2}}\) .
Lemma 8
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint. Let \(O^{P.2}\) denote the optimal solution of \({\textbf {P.2}}\), it holds that \(g(A^{P.3})\ge \gamma g(O^{P.2})\).
We next show that augmenting \(A^{P.3}\) with items from \(\cup _{i\in [m]}B_i\) reduces its utility by a factor of at most 2/3 in expectation.
Lemma 9
Suppose \(\alpha > 1/2\), \({\mathbb {E}}_{A^{approx}}[g(A^{approx})]\ge \frac{1}{3}g(A^{P.3})\) where \(A^{approx}= A^{P.3}\cup (\cup _{i\in [m]}B_i)\).
Proof
Recall that \(B_i= \emptyset \) for all \(i\in [m]\) such that \(|A^{P.3} \cap V_i|\ge |V_i|-\lfloor \beta |V_i|\rfloor \), hence, adding those \(B_i\) to \(A^{P.3}\) does not affect its utility. Therefore, we focus on those groups \(i\in [m]\) with \(|A^{P.3} \cap V_i|< |V_i|-\lfloor \beta |V_i|\rfloor \) in the rest of the proof. Let \(M=\{i\mid |A^{P.3} \cap V_i|< |V_i|-\lfloor \beta |V_i|\rfloor \}\) denote the set containing the indexes of all such groups and we assume \(M\ne \emptyset \) to avoid trivial cases. We next show that it is safe to assume \(\min _{i\in M}|V_i|>1\) without loss of generality, i.e., the smallest group in M contains at least two items. To prove this, we consider two cases, depending on the value of \(\beta \). If \(\beta =1\), then \(|A^{P.3} \cap V_i|< |V_i|-\lfloor \beta |V_i|\rfloor \) does not hold for any group i such that \(|V_i|=1\), that is, \(\min _{i\in M}|V_i|>1\). If \(\beta <1\), then according to the group fairness constraints listed in \({\textbf {P.0}}\), we are not allowed to select any items from those groups with \(|V_i|=1\). Hence, removing all groups with size one from consideration does not affect the quality of the optimal solution.
With the assumption that \(\min _{i\in M}|V_i|>1\), we are now in position to prove this lemma. Recall that for every \(i\in M\), \(B_i\) is a random set of size \(|V_i|-\lfloor \beta |V_i|\rfloor -|A^{P.3} \cap V_i|\) that is sampled from \(V_i\setminus A^{P.3}\). It follows that each item in \(V_i\setminus A^{P.3}\) appears in \(B_i\) with probability at most
We next give an upper bound of (8). Because we assume \(\alpha > 1/2\), we have \(\beta \ge \alpha >1/2\). This, together with the assumption that \(\min _{i\in M}|V_i|>1\), implies that for all \(i\in M\),
Moreover,
where the inequality is by the observation that \(\beta >1/2\). It follows that
where the first inequality is by (13) and the second inequality is by (9) and the assumption that \(|V_i|-\lfloor \beta |V_i|\rfloor -|A^{P.3} \cap V_i|>0\). That is, each item in \(V_i\setminus A^{P.3}\) appears in \(B_i\) with probability at most 2/3. Lemma 3 and the observation that \(g(A^{P.3}\cup \cdot )\) is submodular imply that \({\mathbb {E}}_{A^{approx}}[g(A^{approx})]={\mathbb {E}}_{\cup _{i\in [m]}B_i}[g(A^{P.3}\cup (\cup _{i\in [m]}B_i))]\ge (1-\frac{2}{3})g(A^{P.3}\cup \emptyset )= \frac{1}{3}g(A^{P.3})\). \(\Box \)
Lemmas 8 and 9 together imply that
By the definition of function g, we have
where the last equality is by the observation that \({\textbf {P.2}}\) and \({\textbf {P.0}}\) share the same value of the optimal solution. Hence, the following main theorem holds.
Theorem 2
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint and \(\alpha > 1/2\), we have \({\mathbb {E}}_{A^{approx}}[f(V\setminus A^{approx})]\ge \frac{\gamma }{3}f(OPT)\).
If we adopt the continuous double greedy algorithm (Feldman et al. 2011) as \({\mathcal {A}}\) to compute \(A^{P.3}\), it gives a \(1/e-o(1)\)-approximation solution, that is, \(\gamma \ge 1/e-o(1)\). This, together with Theorem 2, implies that \({\mathbb {E}}_{A^{approx}}[f(V\setminus A^{approx})]\ge \frac{1/e-o(1)}{3}f(OPT)\).
4 Extension: incorporating global cardinality constraint
In this section, we extend \({\textbf {P.0}}\) to incorporate a global cardinality constraint. A formal definition of this problem is listed in \({\textbf {P.A}}\). Our objective is to find a best S subject to a group fairness constraint \((\alpha ,\beta )\) and an additional cardinality constraint c.
4.1 The case when \(\alpha \le 1/2\)
We first consider the case when \(\alpha \le 1/2\). We introduce a new optimization problem \({\textbf {P.B}}\) as follows:
It is easy to verify that \({\textbf {P.B}}\) is a relaxation of \({\textbf {P.A}}\) in the sense that every feasible solution to \({\textbf {P.A}}\) is also a feasible solution to \({\textbf {P.B}}\). Hence, we have the following lemma.
Lemma 10
Let OPT denote the optimal solution of \({\textbf {P.A}}\) and \(O^{P.B}\) denote the optimal solution of \({\textbf {P.B}}\), we have \(f(O^{P.B})\ge f(OPT)\).
It has been shown that the constraints in \({\textbf {P.B}}\) gives rise to a matroid (El Halabi et al. 2020). This, together with the assumption that f is a submodular function, implies that \({\textbf {P.B}}\) is a classic problem of maximizing a submodular function subject to matroid constraints. Now we are ready to present the design of our algorithm.
-
1.
Apply the state-of-the-art algorithm \({\mathcal {A}}\) for matroid constrained submodular maximization to solve \({\textbf {P.B}}\) and obtain a solution \(A^{P.B}\).
-
2.
Note that \(A^{P.B}\) is not necessarily a feasible solution to \({\textbf {P.A}}\) because it might violate the lower bound constraints \(\lfloor \alpha |V_i|\rfloor \le |S \cap V_i|\) for some groups. To make it feasible, we add additional items to \(A^{P.B}\). Specifically, for each group \(i\in [m]\) such that \(|A^{P.B} \cap V_i|< \lfloor \alpha |V_i|\rfloor \), our algorithm selects a backup set \(B_i\) of size \(\lfloor \alpha |V_i|\rfloor -|A^{P.B} \cap V_i|\), by randomly sampling \(\lfloor \alpha |V_i|\rfloor -|A^{P.B} \cap V_i|\) items from \(V_i\setminus A^{P.B}\). Define \(B_i=\emptyset \) if \(|A^{P.1} \cap V_i|\ge \lfloor \alpha |V_i|\rfloor \).
-
3.
At the end, add \(\cup _{i\in [m]}B_i\) to \(A^{P.B}\) to build the final solution \(A^{approx}\), i.e., \(A^{approx}= A^{P.B}\cup (\cup _{i\in [m]}B_i)\).
The pseudocode of this approximation algorithm is given as Algorithm 3. Observe that \(A^{P.B}\) satisfies the group-wise upper bound constraints of \({\textbf {P.A}}\) because \(A^{P.B}\) meets the first set of constraints in \({\textbf {P.B}}\). According to the construction of \(B_i\), adding \(\cup _{i\in [m]}B_i\) to \(A^{P.B}\) does not violate the group-wise upper bound constraints of \({\textbf {P.A}}\) because \(\cup _{i\in [m]}B_i\) are added to meet the lower bound constraints of \({\textbf {P.A}}\) if necessary. Moreover, adding \(\cup _{i\in [m]}B_i\) to \(A^{P.B}\) does not violate the global cardinality constraint of \({\textbf {P.A}}\) because \(A^{P.B}\) meets the second set of constraints in \({\textbf {P.B}}\). At last, it is easy to verify that adding \(\cup _{i\in [m]}B_i\) to \(A^{P.B}\) makes it satisfy the lower bound constraints of \({\textbf {P.A}}\). Hence, \(A^{approx}\) is a feasible solution to \({\textbf {P.A}}\).
Lemma 11
\(A^{approx}\) is a feasible solution to \({\textbf {P.A}}\).
Following the same proof of Theorem 1, we have the following theorem.
Theorem 3
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint and \(\alpha \le 1/2\), we have \({\mathbb {E}}_{A^{approx}}[f(A^{approx})]\ge \frac{\gamma }{2}f(OPT)\).
4.2 The case when \(\alpha > 1/2\)
We next consider the case when \(\alpha > 1/2\). Recall that \(g(\cdot ) = f(V\setminus \cdot )\). We first present a flipped formation of \({\textbf {P.A}}\) as below:
Suppose there is a \(\gamma \)-approximate solution \(A^{P.C}\) to \({\textbf {P.C}}\), it is easy to verify that \(V\setminus A^{P.C}\) is a \(\gamma \)-approximate solution to \({\textbf {P.A}}\). We focus on solving \({\textbf {P.C}}\) in the rest of this section. We first introduce a new optimization problem (labeled as \({\textbf {P.D}}\)) as follows:
\({\textbf {P.D}}\) is relaxed version of \({\textbf {P.C}}\) with both group-wise lower bound constraints \(|V_i|-\lfloor \beta |V_i|\rfloor \le |S \cap V_i|\) and global lower bound constraints \(|S|\ge n-c\) in \({\textbf {P.C}}\) being removed. Hence, we have the following lemma.
Lemma 12
Let \(O^{P.C}\) denote the optimal solution of \({\textbf {P.C}}\) and \(O^{P.D}\) denote the optimal solution of \({\textbf {P.D}}\), we have \(g(O^{P.D})\ge g(O^{P.C})\).
Recall that if f is submodular, g must be submodular (by Lemma 6). Hence, \({\textbf {P.D}}\) is a classic problem of maximizing a submodular function subject to matroid constraints. We next present the design of our algorithm.
-
1.
Apply the state-of-the-art algorithm \({\mathcal {A}}\) for matroid constrained submodular maximization to solve \({\textbf {P.D}}\) and obtain a solution \(A^{P.D}\).
-
2.
Note that \(A^{P.D}\) is not necessarily a feasible solution to \({\textbf {P.C}}\) because it might violate the group-wise or the global lower bound constraints of \({\textbf {P.C}}\). We add additional items to \(A^{P.D}\) to make it feasible. Specifically, for each group \(i\in [m]\), our algorithm selects a backup set \(B_i\) of size \(|V_i|-\lfloor \alpha |V_i|\rfloor -|A^{P.D} \cap V_i|\), by randomly sampling \(|V_i|-\lfloor \alpha |V_i|\rfloor -|A^{P.D} \cap V_i|\) items from \(V_i\setminus A^{P.D}\). Define \(B_i=\emptyset \) if \(|V_i|-\lfloor \alpha |V_i|\rfloor -|A^{P.D} \cap V_i|=0\).
-
3.
Add \(\cup _{i\in [m]}B_i\) to \(A^{P.D}\) to build \(A^{approx}\), i.e., \(A^{approx}= A^{P.D}\cup (\cup _{i\in [m]}B_i)\). Return \(V\setminus A^{approx}\) as the final solution.
The pseudocode of this approximation algorithm is given as Algorithm 4. Observe that adding \(\cup _{i\in [m]}B_i\) to \(A^{P.D}\) ensures that each group contributes exactly \(|V_i|-\lfloor \alpha |V_i|\rfloor \) number of items to the solution. Because \(n-c\le \sum _{i\in [m]}(|V_i|-\lfloor \alpha |V_i|\rfloor )\) (otherwise \({\textbf {P.C}}\) does not have a feasible solution), \(A^{P.D}\cup (\cup _{i\in [m]}B_i)\) must satisfy all constraints in \({\textbf {P.C}}\). Hence, we have the following lemma.
Lemma 13
\(A^{approx}\) is a feasible solution to \({\textbf {P.C}}\).
We next analyze the performance of \(A^{approx}\). The following lemma states that adding \(\cup _{i\in [m]}B_i\) to \(A^{P.D}\) reduces its utility by a factor of at most 2/3 in expectation.
Lemma 14
Suppose \(\alpha > 1/2\), we have \({\mathbb {E}}_{A^{approx}}[g(A^{approx})]\ge \frac{1}{3}g(A^{P.D})\).
Proof
Observe that \(B_i= \emptyset \) for all \(i\in [m]\) such that \(|A^{P.D} \cap V_i|= |V_i|-\lfloor \alpha |V_i|\rfloor \), hence, adding those \(B_i\) to \(A^{P.D}\) does not affect its utility. Therefore, we focus on those groups \(i\in [m]\) with \(|A^{P.D} \cap V_i|< |V_i|-\lfloor \alpha |V_i|\rfloor \) in the rest of the proof. Let \(Z=\{i\in [m]\mid |A^{P.D} \cap V_i|< |V_i|-\lfloor \alpha |V_i|\rfloor \}\) denote the set containing the indexes all such groups. We assume \(Z\ne \emptyset \) to avoid trivial cases. We next show that it is safe to assume \(\min _{i\in Z}|V_i|>1\) without loss of generality, i.e., the smallest group in Z contains at least two items. To prove this, we consider two cases, depending on the value of \(\alpha \). If \(\alpha =1\), then \(|A^{P.D} \cap V_i|< |V_i|-\lfloor \alpha |V_i|\rfloor \) does not hold for any group i such that \(|V_i|=1\). Hence, \(\min _{i\in Z}|V_i|>1\). If \(\alpha <1\), then according to the group fairness constraints listed in \({\textbf {P.A}}\), we are not allowed to select any items from those groups with \(|V_i|=1\). Hence, removing all groups with size one from consideration does not affect the quality of the optimal solution.
With the assumption that \(\min _{i\in Z}|V_i|>1\), we are now ready to prove this lemma. Recall that for every \(i\in Z\), \(B_i\) is a random set of size \(|V_i|-\lfloor \alpha |V_i|\rfloor -|A^{P.D} \cap V_i|\) that is sampled from \(V_i\setminus A^{P.D}\). It follows each item in \(V_i\setminus A^{P.D}\) appears in \(B_i\) with probability at most
We next give an upper bound of (15). Because we assume \(\alpha > 1/2\) and \(\min _{i\in Z}|V_i|>1\), it holds that for all \(i\in M\),
Moreover,
where the inequality is by the assumptions that \(\alpha >1/2\) and \(|V_i|>1\). It follows that
where the first inequality is by (20) and the second inequality is by (16) and the assumption that \(|V_i|-\lfloor \alpha |V_i|\rfloor -|A^{P.D} \cap V_i|>0\). That is, each item in \(V_i\setminus A^{P.D}\) appears in \(B_i\) with probability at most 2/3. Lemma 3 and the observation that \(g(A^{P.D}\cup \cdot )\) is submodular imply that \({\mathbb {E}}_{A^{approx}}[g(A^{approx})]={\mathbb {E}}_{\cup _{i\in [m]}B_i}[g(A^{P.D}\cup (\cup _{i\in [m]}B_i))]\ge (1-\frac{2}{3})g(A^{P.D}\cup \emptyset )=\frac{1}{3}g(A^{P.D})\). \(\Box \)
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint, we have
where the first inequality is by Lemma 14. This, together with \(g(O^{P.D})\ge g(O^{P.C})\) (as proved in Lemma 12), implies that \({\mathbb {E}}_{A^{approx}}[g(A^{approx})]\ge \frac{\gamma }{3}g(O^{P.C})\). By the definition of function g, we have
where the last equality is by the observation that \({\textbf {P.A}}\) and \({\textbf {P.C}}\) share the same value of the optimal solution. Hence, the following main theorem holds.
Theorem 4
Suppose \({\mathcal {A}}\) is a \(\gamma \)-approximate algorithm for non-monotone submodular maximization subject to a matroid constraint and \(\alpha > 1/2\), we have \({\mathbb {E}}_{A^{approx}}[f(V\setminus A^{approx})]\ge \frac{\gamma }{3}f(OPT)\).
5 Conclusion
This paper presents a comprehensive investigation of the non-monotone submodular maximization problem under group fairness constraints. Our main contribution is the development of several constant-factor approximation algorithms for this problem. In the future, we plan to expand our research to explore alternative fairness metrics.
Data Availability
Enquiries about data availability should be directed to the authors.
Notes
A matroid is a pair \({\mathcal {M}} = (V, {\mathcal {I}})\) where \({\mathcal {I}} \subseteq 2^V\) and 1. \(\forall Y \in {\mathcal {I}}, X \subseteq Y \rightarrow X \in {\mathcal {I}}\). 2. \(\forall X, Y \in {\mathcal {I}}; |X| < |Y| \rightarrow \exists e\in Y \setminus X; X \cup \{ e\} \in {\mathcal {I}}\).
References
Biddle D (2017) Adverse impact and test validation: a practitioner’s guide to valid and defensible employment testing. Routledge, Oxfordshire
Buchbinder N, Feldman M, Naor J, Schwartz R (2014) Submodular maximization with cardinality constraints. In: Proceedings of the twenty-fifth annual ACM-SIAM symposium on Discrete algorithms, pp 1433–1452. SIAM
Celis E, Keswani V, Straszak D, Deshpande A, Kathuria T, Vishnoi N (2018) Fair and diverse dpp-based data summarization. In: International conference on machine learning, pp 716–725. PMLR
Celis LE, Huang L, Vishnoi NK (2018) Multiwinner voting with fairness constraints. In: Proceedings of the 27th international joint conference on artificial intelligence, pp 144–151
Chierichetti F, Kumar R, Lattanzi S, Vassilvtiskii S (2019) Matroids, matchings, and fairness. In: The 22nd international conference on artificial intelligence and statistics, pp 2212–2220. PMLR
Das A, Kempe D (2008) Algorithms for subset selection in linear regression. In: Proceedings of the fortieth annual ACM symposium on Theory of computing, pp 45–54
Dueck D, Frey BJ (2007) Non-metric affinity propagation for unsupervised image categorization. In: 2007 IEEE 11th international conference on computer vision, pp 1–8. IEEE
Dwork C, Hardt M, Pitassi T, Reingold O, Zemel R (2012) Fairness through awareness. In: Proceedings of the 3rd innovations in theoretical computer science conference
El-Arini K, Guestrin C (2011) Beyond keyword search: discovering relevant scientific literature. In: Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining, pp 439–447
El Halabi M, Mitrović S, Norouzi-Fard A, Tardos J, Tarnawski JM (2020) Fairness in streaming submodular maximization: algorithms and hardness. Adv Neural Inf Process Syst 33:13609–13622
Feldman M, Naor J, Schwartz R (2011) A unified continuous greedy algorithm for submodular maximization. In: 2011 IEEE 52nd annual symposium on foundations of computer science, pp 570–579. IEEE
Golovin D, Krause A (2011) Adaptive submodularity: theory and applications in active learning and stochastic optimization. J Artif Intell Res 42:427–486
Gotovos A, Karbasi A, Krause A (2015) Non-monotone adaptive submodular maximization. In: Twenty-fourth international joint conference on artificial intelligence
Gu S, Gao C, Wu W (2022) A binary search double greedy algorithm for non-monotone DR-submodular maximization. In: Algorithmic aspects in information and management: 16th international conference, AAIM 2022, Guangzhou, China, 13–14 Aug, 2022, proceedings. Springer, pp. 3–14
Joseph M, Kearns M, Morgenstern JH, Roth A (2016) Fairness in learning: classic and contextual bandits. Adv. Neural Inf. Process. Syst. 29
Mirzasoleiman B, Badanidiyuru A, Karbasi A (2016) Fast constrained submodular maximization: personalized data summarization. In: ICML, pp 1358–1367
Monroe BL (1995) Fully proportional representation. Am Polit Sci Rev 89(4):925–940
Shi G, Gu S, Wu W (2021) k-submodular maximization with two kinds of constraints. Discr Math Algorithms Appl 13(04):2150036
Sipos R, Swaminathan A, Shivaswamy P, Joachims T (2012) Temporal corpus summarization using submodular word coverage. In: Proceedings of the 21st ACM international conference on Information and knowledge management, pp 754–763
Tang S, Yuan J (2020) Influence maximization with partial feedback. Oper Res Lett 48(1):24–28
Tang S, Yuan J (2021) Adaptive regularized submodular maximization. In: 32nd international symposium on algorithms and computation (ISAAC 2021). Schloss Dagstuhl-Leibniz-Zentrum für Informatik
Tang S, Yuan J (2022) Group equility in adaptive submodular maximization. arXiv preprint arXiv:2207.03364
Tang S, Yuan J (2023) Beyond Submodularity: A unified framework of randomized set selection with group fairness constraints (Under Review)
Tang S, Yuan J, Mensah-Boateng T (2023) Achieving long-term fairness in submodular maximization through randomization (Under Review)
Tsang A, Wilder B, Rice E, Tambe M, Zick Y (2019) Group-fairness in influence maximization. arXiv preprint arXiv:1903.00967
Zafar MB, Valera I, Rogriguez MG, Gummadi KP (2017) Fairness constraints: mechanisms for fair classification. In: Artificial intelligence and statistics, pp 962–970. PMLR
Funding
The authors have not disclosed any funding.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have not disclosed any competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Yuan, J., Tang, S. Group fairness in non-monotone submodular maximization. J Comb Optim 45, 88 (2023). https://doi.org/10.1007/s10878-023-01019-4
Accepted:
Published:
DOI: https://doi.org/10.1007/s10878-023-01019-4