1 Introduction

Hamiltonian cycles are important objects of study from both theoretical and application perspectives and for deterministic graphs, there are many known sufficient conditions for their existence. For example, Dirac (1952) proved that if the minimum degree of a graph \(H\) on \(n\) vertices is at least \(\frac{n}{2},\) then \(H\) is Hamiltonian and for further results along this direction, we refer to Chapter \(7\) of West (2000).

In the case of random graphs, one of the earliest studies was initiated by Pósa (1976), who used rotation techniques to obtain edge probability threshold for the occurrence of Hamiltonian paths with high probability, i.e. with probability converging to one as \(n \rightarrow \infty .\) For a detailed survey of further techniques, we refer to the book (Bollobás 2001). In Lee and Sudakov (2012) a Dirac-type theorem is obtained for random subgraphs of a graph that is not necessarily complete. The main result there is that if the minimum degree of the “parent” graph is at least \(\frac{n}{2},\) then the Hamiltonicity threshold is again of the order of \(\frac{\log {n}}{n}\) and they also demonstrated the tightness of this bound. We also remark that recently (Condon et al. 2021) has obtained Dirac-type theorems random regular graphs.

In this paper, we consider inhomogenous random graphs with independently open edges, whose edge probabilities are not all the same. We obtain sufficient “two-sided” conditions on the edge probabilities, for the existence of a Hamiltonian path with high probability. We then show that under weaker one-sided conditions, there is a perfect matching with high probability.

The paper is organized as follows. In Section 2 we state and prove our results regarding existence of Hamiltonian paths and perfect matchings in inhomogenous random graphs and in Section 3, we briefly describe an application of our results in the context of channel assignment in wireless networks.

2 Inhomogenous Random Graphs

Let \(K_n\) be the complete graph on \(n\) vertices and let \(\{Z(f)\}_{f \in K_n}\) be independent random variables with distribution

$$\begin{aligned} \mathbb {P}(Z(f) = 1) = p(f) = 1-\mathbb {P}(Z(f) = 0) \end{aligned}$$
(2.1)

where \(0 \le p(f) \le 1\) and let \(G\) be the random graph formed by the union of all edges \(f\) satisfying \(Z(f) = 1.\) A path in \(G\) is a sequence of vertices \(P = (u_1,\ldots ,u_t)\) such that \(u_i\) is adjacent to \(u_{i+1}\) in \(G,\) for each \(1 \le i \le t-1.\) The length of \(P\) is the number of edges in \(P\) which in this case is \(t-1.\) If \(P\) has the maximum possible length of \(n-1,\) then we say that \(P\) is a Hamiltonian path.

Our first result obtains Hamiltonian paths in \(G\) for the following class of “well-behaved” probability assignments. For constants \(0< \alpha < 1\) and \(c_1,c_2 > 0\) and \(0< p = p(n) < 1,\) we say that the probability assignment \(\{p(u,v)\}_{1 \le u < v \le n}\) in (2.1) is \((\alpha ,c_1,c_2,p)-\)good if for any vertex \(u\) and any set \(S\) containing \(r \ge \alpha n\) vertices,

$$\begin{aligned} c_1 rp \le \sum \limits _{v \in S} p(u,v) \le c_2 rp. \end{aligned}$$
(2.2)

Here and throughout, constants do not depend on \(n.\) Letting \(E_{HAM}\) be the event that \(G\) contains a Hamiltonian path, we have the following result.

Theorem 1

Suppose \(\{p(u,v)\}\) is \((\alpha ,c_1,c_2,p)-\)good for \(p = Cn^{-\frac{k}{k+1}}\) where

\(C > 0\) is a constant and \(k \ge 1\) is an integer constant. For every

$$\begin{aligned} C \le \frac{1}{10c_2} \text { and }\alpha \le \min \left( \frac{7}{8},\left( \frac{Cc_1}{8}\right) ^{k+1}\right) , \end{aligned}$$
(2.3)

there exists a constant \(\theta > 0\) such that

$$\begin{aligned} \mathbb {P}(E_{HAM}) \ge 1- e^{-\theta np}. \end{aligned}$$
(2.4)

In other words, the resulting random graph \(G\) contains a Hamiltonian path with high probability, i.e., with probability converging to one as \(n \rightarrow \infty .\)

For example, suppose we consider the particular case when \(G\) is homogenous with common edge probability \(p_{hom} \ge \frac{1}{n^{\beta }}\) for some \(0< \beta < 1.\) If \(k\) is a large enough integer so that \(0< \beta< \frac{k}{k+1} < 1,\) then by direct coupling, it suffices to demonstrate Hamiltonicity for the homogenous random subgraph \(G_{low}\) with edge probability \(p = Cn^{-\frac{k}{k+1}}\) where \(C\) is chosen sufficiently small so that the first condition in (2.3) holds. In this case condition (2.2) holds trivially with \(c_1 = c_2=1\) and all constant \(\alpha >0.\) Therefore the second condition in (2.3) is also satisfied and so from (2.4), we then get that \(G_{low}\) (and therefore \(G\)) contains a Hamiltonian path with high probability.

Our proof of Theorem 1 involves a combination of the Pósa rotation technique (Pósa 1976) and the “single vertex exclusion” method used in Ganesan (2020) for obtaining Hamiltonicity in inhomogenous random graphs when \(k = 1.\) Throughout, we use the following standard deviation estimate. Let \(Z_i, 1 \le i \le t\) be independent Bernoulli random variables satisfying \(\mathbb {P}(Z_i = 1) = p_i = 1-\mathbb {P}(Z_i = 0).\) If \(W_t = \sum _{i=1}^{t} Z_i\) and \(\mu _t = \mathbb {E}W_t,\) then for any \(0< \eta < \frac{1}{2}\) we have that

$$\begin{aligned} \mathbb {P}\left( \left| W_t-\mu _t\right| \ge \eta \mu _t\right) \le 2\exp \left( -\frac{\eta ^2}{4}\mu _t\right) . \end{aligned}$$
(2.5)

For a proof of (2.5), we refer to Corollary \(A.1.14,\) pp. \(312,\) Alon and Spencer (2008).

Proof of Theorem 1: We obtain the Hamiltonian path of \(G\) in three steps with the first two steps are preliminary calculations. In the first step, we define an event \(E_{good}\) regarding the neighbourhood size of arbitrary sets and obtain probability estimates for \(E_{good}.\) Next, in the second step, we define the concept of pivots and use the occurrence of \(E_{good}\) to estimate the number of pivots in the maximum length path of \(G\) obtained after excluding a single vertex. Finally, in the third step, we combine the two estimates to show that \(\pi \) is Hamiltonian with high probability.

Step 1: Let \(S \subset \{1,2,\ldots ,n\} \) be any set of size \(s := \#S\) and let \(\mathcal{N}_{out}(S)\) be the set of all vertices of the complement set \(S^c,\) adjacent to at least one vertex of \(S.\) Letting \(N_{out}(S) = \#\mathcal{N}_{out}(S)\) denote the size of \(\mathcal{N}_{out}(S),\) we begin by showing that for any set \(S\) of size \(1 \le s \le \frac{1}{10c_2p},\) we have

$$\begin{aligned} \frac{3c_1nps}{4} \le \mathbb {E} N_{out}(S) \le c_2nps. \end{aligned}$$
(2.6)

Indeed, by definition \(s \le \frac{1}{10c_2p}=o(n)\) and so \(n-s \ge \frac{7n}{8}\) for all \(n\) large. Therefore using (2.2), we get that the expected number of vertices adjacent to \(u\) in \(G\) is at least \(c_1np\) and at most \(c_2np.\) This proves (2.6) for the case \(s=1\) and also the upper bound in (2.6) for general \(s.\)

We prove the lower bound in (2.6) by induction on \(s.\) Pick a set \(S\) of size \(2 \le s \le \frac{1}{10c_2p}\) and for \(u \in S,\) define \(S_u := S \setminus \{u\}.\) The set \(S_u\) has size \(s-1\) and so by the induction assumption

$$\begin{aligned} \frac{3c_1np(s-1)}{4} \le \mathbb {E} N_{out}(S_u) \le c_2np(s-1). \end{aligned}$$
(2.7)

Using the concentration estimate (2.5) with \(\eta >0\) small, we then get that

$$\begin{aligned} \mathbb {P}\left( N_{out}(S_u) \le (1+\eta )c_2np(s-1)\right) \ge 1-e^{-C_0np(s-1)} \ge 1-e^{-C_0np} \end{aligned}$$
(2.8)

for some constant \(C_0 > 0,\) not depending on \(s.\) If the event defined in the left hand side of (2.8) occurs, then there are at least \(n-(1+\eta )c_2np(s-1)\) vertices in \(S_u^c\) that are not adjacent to any vertex of \(S_u\) and at least \(n-1-(1+\eta )c_2np(s-1)\) among these are adjacent to \(u\) in \(K_n.\)

Recalling that \(s \le \frac{1}{10c_2p},\) we then choose \(\eta > 0\) small so that 

$$\begin{aligned} n-1-(1+\eta )c_2np(s-1) \ge n-1-\frac{(1+\eta )}{10} \ge \frac{7n}{8} \end{aligned}$$

for all \(n \ge N\) large not depending on \(s.\) With this choice of \(\eta \) we get from (2.2) that the expected number of vertices of \(S^c\) adjacent only to \(u\) and no other vertex of \(S,\) is at least \(\frac{7c_1np}{8}(1-e^{-C_0np}).\) Thus

$$\begin{aligned} \mathbb {E}N_{out}(S) \ge \frac{3c_1np(s-1)}{4} + \frac{7c_1np}{8}(1-e^{-C_0np}) \ge \frac{3c_1 nps}{4} \end{aligned}$$

and this proves the induction step for the lower bound in (2.6).

Defining \(E_{good}(S) := \left\{ \frac{c_1nps}{2} \le N_{out}(S) \le 4c_2nps\right\} \) and using the deviation estimate (2.5) with \(\eta = \frac{1}{4},\) we obtain from the bounds in (2.6) that

$$\begin{aligned} \mathbb {P}(E_{good}(S)) \ge 1-\exp \left( -2C_1 nps\right) \end{aligned}$$

for some constant \(C_1 > 0.\) There are at most \({n \atopwithdelims ()s} \le n^{s}\) choices for \(S\) and so setting \(E_{good} := \bigcap _{S} E_{good}(S),\) where the intersection is over all permissible \(S,\) we get from the union bound that

$$\begin{aligned} \mathbb {P}(E_{good}) \ge 1- \sum \limits _{s \ge 1} n^{s}\exp \left( -2C_1 nps \right) . \end{aligned}$$

Since \(p = Cn^{-\frac{k}{k+1}}\) is much larger than \(\frac{\log {n}}{n},\) we have that

$$\begin{aligned} n^{s}\exp \left( -2C_1 nps \right) = \exp \left( s\left( \log {n} - 2C_1 np \right) \right) \le \exp \left( -C_1 nps \right) , \end{aligned}$$

and so

$$\begin{aligned} \mathbb {P}(E_{good}) \ge 1-\sum \limits _{s \ge 1} e^{-C_1 nps} \ge 1-2e^{-C_1np} \end{aligned}$$
(2.9)

for all \(n\) large.

Fig. 1
figure 1

The path \(\pi \) with endvertices \(u\) and \(v\) is shown at the top, together with a neighbour \(y\) of \(u.\) We perform a Pósa rotation to obtain a new path with the pivot \(x\) as the endvertex, as illustrated in the bottom figure

The event \(E_{good}\) as defined above, allows us to count the number of pivots in the maximum length paths of \(G\) as demonstrated in Step \(2\) below.

Step 2: Let \(\pi = (\pi (1),\pi (2),\ldots ,\pi (t))\) be the longest path in \(G\) with endvertices \(\pi (1)\) and \(\pi (t).\) If \(\mathcal{S}_1\) is the set of neighbours of \(\pi (1)\) in \(G,\) then all vertices in \(\mathcal{S}_1\) must also be present in \(\pi \) and so \(\mathcal{S}_1 = \{\pi (a_1),\ldots ,\pi (a_r)\}\) for some integers \(2 = a_1< a_2< \ldots < a_r.\) For each \(a_j, j \ge 1\) we can do a Pósa rotation as shown in Fig. 1 and obtain a maximum length path with endvertex \(\pi (a_{j}-1).\) We refer to \(\pi (a_j-1), 2 \le j \le r\) as first generation pivots or pivots associated to \(\pi (1).\)

Consider now the pivot \(\pi (a_j-1)\) and let \(\gamma _j = (\gamma _j(1),\ldots ,\gamma _j(t))\) be a maximum length path in \(G\) with endvertex \(\gamma _j(1) = \pi (a_j-1),\) obtained via the Pósa rotation as described in the previous paragraph. Again all neighbours of \(\pi (a_j-1)\) in the graph \(G\) must be present in \(\gamma _j\) as well and if \(\mathcal{S}_j\) is the set of neighbours of \(\pi (a_j-1)\) in \(G,\) then we obtain \(\#\mathcal{S}_j\) pivots associated to \(\pi (a_j-1),\) which we call as second generation pivots. Continuing this way we define \(i^{th}-\)generation pivots for arbitrary \(i \ge 1.\) Summarizing, the longest path \(\pi \) in \(G\) satisfies the following properties:

\((p1)\) All neighbours of the two endvertices of \(\pi \) are present in \(\pi \) itself.

\((p2)\) If \(v\) is a \(i^{th}-\)generation pivot of \(\pi \) and \(\gamma _v\) is a maximum length path obtained from successive Pósa rotations of \(\pi \) and containing \(v\) as an endvertex, then all neighbours of \(v\) belong to \(\gamma _v\) (and therefore to \(\pi \)) as well.

We now assume that the event \(E_{good}\) described in Step \(1\) occurs and estimate the number of pivots in the maximum length path obtained by excluding a single vertex. Specifically, we perform an iterative pivot counting procedure consisting of \(k\) steps at the end of which we demonstrate that the overall number of pivots grows linearly in \(n.\) Details follow.

For \(1 \le j \le n\) we let \(\pi _j\) be the maximum length path in the graph \(G_j\) obtained by removing the vertex \(j\) from \(G\) and show by induction that for each \(1 \le l \le k,\) there are at least \(\left( \frac{c_1np}{8}\right) ^{l}\) and at most \((8c_2np)^{l}\) pivots belonging to the \(l^{th}\) generation (here we recall that the integer \(k\) is defined via \(p = Cn^{-\frac{k}{k+1}}\)); i.e., our goal is to obtain the bounds

$$\begin{aligned} \left( \frac{c_1 np}{8}\right) ^{l} \le \#\mathcal{P}_l \le (8c_2np)^{l} \end{aligned}$$
(2.10)

for each \(1 \le l \le k.\)

We begin with the base case \(l=1.\) Indeed if \(\pi _j = (\pi _j(1),\ldots ,\pi _j(t))\) then since \(E_{good}\) occurs, the vertex \(\pi _j(1)\) contains at least \(\frac{c_1np}{2}\) neighbours and at most \(4c_2np\) neighbours in \(G.\) Therefore there are at least \(\frac{c_1np}{2}-1 \ge \frac{c_1np}{8}\) vertices adjacent to \(\pi _j(1)\) in \(G_j\) and so there are at least \(\frac{c_1np}{8}\) pivots associated with \(\pi _j(1).\) Each vertex adjacent to \(\pi _j(1)\) in \(G_j\) gives rise to at most two pivots and so there are at least \(\frac{c_1np}{8}\) and at most \(2(4c_2np) = 8c_2np,\) first generation pivots. This completes the proof of (2.10) for the case \(l=1.\)

To argue for general \(l,\) we first see that the \(E_{good}\) probability estimate is valid as long as \((8c_2np)^{l} \le \frac{1}{10c_2p}\) (see step \(1\)). Since \(p = \frac{C}{n^{k/k+1}},\) we have that 

$$\begin{aligned} p(8c_2np)^{k} = C^{k+1}(8c_2)^{k} \le \frac{1}{10c_2} \end{aligned}$$

if \(C > 0\) satisfies the first condition in (2.3). Fixing such a \(C\) henceforth, we now proceed to the induction step. For \(l \ge 1,\) let \(\mathcal{P}_l\) be the set of all \(l^{th}-\)generation pivots of \(\pi _j\) and let \(\bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v)\) be the set of all vertices adjacent to at least one vertex of \(\mathcal{P}_l\) in \(G_j,\) where \(\mathcal{N}_j(v)\) denotes the set of all neighbours of \(v\) in \(G_j.\) By induction assumption

$$\begin{aligned} \left( \frac{c_1 np}{8}\right) ^{l} \le \#\mathcal{P}_l \le (8c_2np)^{l} \end{aligned}$$
(2.11)

and due to the occurrence of the event \(E_{good},\) we have that

$$\begin{aligned} \frac{c_1np}{2} \cdot \left( \frac{c_1np}{8}\right) ^{l}-1 \le \#\left( \bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v)\right) \le (4c_2np) (8c_2np)^{l}. \end{aligned}$$
(2.12)

Thus the number of \((l+1)^{th}-\)generation pivots is at most 

$$\begin{aligned} 2\cdot (4c_2np) \cdot (8c_2np)^{l} = (8c_2np)^{l+1}. \end{aligned}$$

For a lower bound on the number of \((l+1)^{th}-\)generation pivots, we pick a \(l^{th}-\)generation pivot \(v \in \mathcal{P}_l\) and perform \(l\) Pósa rotations to obtain a path \(\gamma _{v,j}\) containing \(v\) as an endvertex. In doing so, we remove exactly \(l\) edges from \(\pi _j\) and add \(l\) other edges, to obtain \(\gamma _{v,j}.\) We let \(\mathcal{R}_{j}(v)\) be the union of the set of all endvertices of the \(l\) removed edges, \(l\) newly added edges and the endvertices of \(\gamma _{v,j}\) so that there are at most \(4l+2\) vertices in \(\mathcal{R}_j(v).\)

We now obtain a lower bound on the number of pivots that arise out of some \(l^{th}\) generation pivot in \( \Lambda _j := \bigcup _{v \in \mathcal{P}_l} \left( \mathcal{N}_j(v) \setminus \mathcal{R}_j(v)\right) .\) Let \(w \in \Lambda _j\) be adjacent to \(v \in \mathcal{P}_l\) in \(G_j.\) From property \((p2)\) described above, all neighbours of \(v\) in \(G_j\) belong to the path \(\pi _j\) and so \(w = \pi _j(x)\) for some \(x.\) Moreover, because \(w \notin \mathcal{R}_j(v),\) both the edges 

$$\begin{aligned} (\pi _j(x),\pi _j(x+1)) \text { and }(\pi _j(x-1),\pi _j(x)) \end{aligned}$$

must belong to the path \(\gamma _{v,j}\) with \(v\) as an endvertex, obtained by performing \(l\) Pósa rotations on \(\pi _j.\) Thus at least one of the vertices in

$$\begin{aligned} \{\pi _j(x-1),\pi _j(x+1)\}, \end{aligned}$$

say for example \(\pi _j(x-1),\) must necessarily be a \((l+1)^{th}-\)generation pivot and we say that \(\pi _j(x-1)\) is an \((l+1)^{th}-\)generation pivot created by an \(l^{th}-\)generation pivot.

Conversely, any \((l+1)^{th}-\)generation pivot created from an \(l^{th}-\)generation pivot is of the form \(\pi _j(y)\) and is created from an \(l^{th}-\)generation pivot that is necessarily adjacent to either \(\pi _j(y-1) \in \Lambda _j\) or \(\pi _j(y+1) \in \Lambda _j.\) We have a couple of remarks here. It may happen that more than one \(l^{th}-\)generation pivot itself could be adjacent to either \(\pi _j(y-1)\) or \(\pi _j(y+1)\) or some \((l+1)^{th}-\)generation pivot is already a \(y^{th}-\)generation pivot for some \( y \le l.\) In any case, our count of the number of pivots in the \((l+1)^{th}\) generation above depends only on the neighbours of the \(l^{th}-\)generation pivots, i.e., vertices in \(\Lambda _j.\) Therefore from the above argument, we get

$$\begin{aligned} \#\mathcal{P}_{l+1}{} & {} \ge \frac{\#{\Lambda _j}}{2} \nonumber \\{} & {} = \frac{1}{2}\#\left( \bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v) \setminus \mathcal{R}_j(v)\right) \nonumber \\{} & {} \ge \frac{1}{2} \#\left( \left( \bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v)\right) \setminus \left( \bigcup _{v \in \mathcal{P}_l} \mathcal{R}_j(v)\right) \right) \nonumber \\{} & {} \ge \frac{1}{2}\#\left( \bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v)\right) - \sum \limits _{v \in \mathcal{P}_l} \#\mathcal{R}_j(v)\nonumber \\{} & {} \ge \frac{1}{2}\#\left( \bigcup _{v \in \mathcal{P}_l} \mathcal{N}_j(v)\right) -(4l+2)\#\mathcal{P}_l, \end{aligned}$$
(2.13)

since the number of vertices in any \(\mathcal{R}_j(v)\) is at most \(4l+2.\) Plugging the bounds from (2.11) and (2.12) into (2.13), we get that

$$\begin{aligned} \#\mathcal{P}_{l+1} \ge \frac{1}{2}\left( \frac{c_1np}{2} \cdot \left( \frac{c_1np}{8}\right) ^{l}-1\right) - (4l+2)(8c_2np)^{l} \ge \left( \frac{c_1np}{8}\right) ^{l+1} \end{aligned}$$

for all \(n\) large. This obtains (2.10) for the general case.

From (2.10), we see that the number of \((k+1)^{th}-\)generation pivots is at least 

$$\begin{aligned} \left( \frac{c_1np}{8}\right) ^{k+1} = \left( \frac{c_1}{8}\right) ^{k+1} \cdot n \cdot p(np)^{k} =: Dn, \end{aligned}$$

where \(D = \left( \frac{Cc_1}{8}\right) ^{k+1}>0\) is a constant. Summarizing, we have shown that if the event \(E_{good}\) occurs then there are at least \(Dn\) pivots in \(\pi _j\) for any \(1 \le j \le n\) and this completes the second step of the proof.

Step 3: We now combine the estimates obtained in the previous two steps to obtain the desired Hamiltonian path. We begin by noting that if the vertex \(j\) does not belong to the maximum length path \(\pi ,\) then the longest path \(\pi _j\) in the random graph \(G_j\) is also \(\pi \) and more importantly, \(j\) is not adjacent to any pivot of \(\pi _j = \pi ;\) otherwise, we would have a path containing \(j\) as an endvertex and with length strictly larger than \(\pi .\) Therefore letting \(Q_j(\pi _j)\) be the event that \(j\) is not adjacent to any pivot of \(\pi _j,\) we get that

$$\begin{aligned} \mathbb {P}\left( \{j \notin \pi \} \cap E_{good}\right) \le \sum _{\Gamma } \mathbb {P}\left( \{\pi _j = \Gamma \} \cap Q_j(\Gamma )\right) \!. \end{aligned}$$
(2.14)

where the summation is over all deterministic paths \(\Gamma \) not containing \(j,\) satisfying \((p1)-(p2)\) and containing at least \(Dn\) pivots. The event \(Q_j(\Gamma )\) is independent of the event \(\{\pi _j = \Gamma \}\) and moreover if the probability assignment is \((\alpha ,c_1,c_2,p)-\)good for \(\alpha \le D,\) then we get that \(\mathbb {P}(Q_j(\Gamma )) \le e^{-c_1Dnp}.\) Consequently,

$$\begin{aligned} \mathbb {P}\left( \{\pi _j = \Gamma \} \cap Q_j(\Gamma )\right){} & {} = \mathbb {P}(\pi _j = \Gamma ) \mathbb {P}(Q_j(\Gamma )) \\{} & {} \le e^{-c_1Dnp}\mathbb {P}(\pi _j = \Gamma ) \end{aligned}$$

Summing over \(\Gamma \) we therefore get from (2.14) that \(\mathbb {P}\left( \{j \notin \pi \} \cap E_{good}\right) \le e^{-c_1Dnp}\) and so using (2.9), we get that

$$\begin{aligned} \mathbb {P}(j \notin \pi ) \le e^{-c_1Dnp} + \mathbb {P}(E^c_{good}) \le e^{-c_1Dnp} + 2e^{-C_1np} \le e^{-C_2np} \end{aligned}$$

for some constant \(C_2 > 0.\) Therefore by the union bound, we get that \(\pi \) is Hamiltonian with probability at least \(1-ne^{-C_2 np}\) and this completes the proof of Theorem 1. \(\square \)

In Theorem 1 we have obtained sufficient conditions to be satisfied by the probability assignments, for the existence of Hamiltonian paths. Specifically, (2.2) has “two-sided” conditions on both the upper and lower bounds of the edge probability sums. A natural question therefore is whether under weaker conditions, we can say anything about the existence of Hamiltonian paths or perhaps at least perfect matchings. For context, we recall that a set of vertex disjoint edges in the random graph \(G\) is called a matching and a matching \(\mathcal{W}\) of \(G\) is said to be perfect if every vertex except at most one, is an endvertex of some edge in \(\mathcal{W}.\) By definition, if \(G\) has a Hamiltonian path \(\pi ,\) then picking every alternate edge of \(\pi \) gives us a perfect matching. Thus the existence of a Hamiltonian path implies the existence of a perfect matching but the converse, however, is not true.

We now consider probability assignments with “one-sided” conditions involving only lower bounds and obtain perfect matchings in the resulting random graph \(G.\) For constants \(0< \beta ,\gamma < 1\) and \(0< p = p(n) < 1,\) we say that \(\{p(u,v)\}\) is a \((\beta ,d_1,d_2,p)-\)nice probability assignment if the following holds for any two vertices \(u,v:\) For any two disjoint sets \(S_1,S_2\) not containing \(u\) or \(v\) and having cardinality \(r \ge \beta n\) each, we have

$$\begin{aligned} \sum \limits _{w \in S_1} p(u,w) \ge d_1 rp\;\;\text { and } \sum \limits _{w_1 \in S_1,w_2 \in S_2}p(u,w_1)p(w_2,v) \ge d_2 rp^2. \end{aligned}$$
(2.15)

Letting \(E_{PER}\) be the event that \(G\) contains a perfect matching, we have the following result. As before, constants do not depend on \(n.\)

Theorem 2

Suppose \(\{p(u,v)\}\) is \((\beta ,d_1,d_2,p)-\)nice for some \(p \ge \frac{\log {n}}{\sqrt{n}}\) and

\(\beta \le \frac{1}{4}.\) There is a constant \(D > 0\) such that

$$\begin{aligned} \mathbb {P}(E_{PER}) \ge 1- e^{-D(\log {n})^2}. \end{aligned}$$
(2.16)

In other words, the one-sided conditions in (2.15) ensure the existence of perfect matchings with high probability. In the next Section, we briefly describe an application of our result in the context of channel assignment in wireless networks.

Proof of Theorem 2: For simplicity we assume throughout that \(n = 4z\) is even and let \(G_{bip}\) be the bipartite subgraph of \(G\) with left vertex set \(X = \{1,2,\ldots ,2z\}\) and right vertex set \(Y = \{2z+1,\ldots ,n\}.\) We obtain a perfect matching in \(G_{bip}\) using a bootstrapping technique as follows. In the first step, we let \(\mathcal{M}\) be the maximum matching in \(G_{bip}\) (picked according to a deterministic rule) and estimate the probability \(p_{0}\) that \(\#\mathcal{M} \ge z.\) In the next step, we then use the estimate for \(p_{0}\) as a bootstrap to bound the probability of a perfect matching. We provide the details below.

For \(\mathcal{L} \subseteq X\) and \(\mathcal{R} \subseteq Y\) let \(E\left( \mathcal{L}, \mathcal{R}\right) \) be the event that no vertex of \(\mathcal{L}\) is adjacent to any vertex of \(\mathcal{R},\) in \(G_{bip}.\) If \(\#\mathcal{L} \ge z\) and \(\#\mathcal{R} \ge z,\) then any vertex \(v \in \mathcal{L}\) is adjacent to at least \( z = \frac{n}{4}\) vertices of \(\mathcal{R}\) in \(K_n\) and so using the first condition in (2.15), we get that the expected number of vertices adjacent to \(v\) in \(G_{bip}\) is at least \(d_1 zp.\) Therefore the total expected number of edges of \(G_{bip}\) containing one endvertex in \(\mathcal{L}\) and the other endvertex in \(\mathcal{R}\) is at least \(d_1 z^2p = \frac{d_1 n^2p}{16}\) and consequently,

$$\begin{aligned} \mathbb {P}\left( E\left( \mathcal{L}, \mathcal{R}\right) \right) \le \exp \left( -\frac{d_1 n^2p}{16}\right) . \end{aligned}$$
(2.17)

Now if the maximum matching \(\mathcal{M}\) of \(G_{bip}\) has size \(\#\mathcal{M} < z,\) then removing the edges and vertices of \(\mathcal{M}\) from \(G_{bip}\) we are left with an edgeless bipartite graph containing at least \(z\) left vertices and at least \(n-2z-z \ge z\) right vertices. Thus \(\mathbb {P}\left( \#\mathcal{M} < z\right) \le \mathbb {P}\left( \bigcup E\left( \mathcal{L}, \mathcal{R}\right) \right) \) where the union is over all sets \(\mathcal{L} \subseteq X\) and \(\mathcal{R} \subseteq Y\) satisfying \(\#\mathcal{L} \ge z\) and \(\#\mathcal{R}\ge z.\) Letting \(E_{low} := \{\#\mathcal{M} \ge z\},\) we then get from the union bound and (2.17) that

$$\begin{aligned} \mathbb {P}\left( E_{low}^c\right) \le \sum \limits _{\mathcal{L}, \mathcal{R}} \mathbb {P}\left( E\left( \mathcal{L},\mathcal{R}\right) \right) \le 4^{n}\cdot \exp \left( -\frac{d_1n^2p}{16}\right) \end{aligned}$$
(2.18)

since there are at most \(2^{n}\) choices each, for \(\mathcal{L}\) and for \(\mathcal{R}.\)

Suppose the event \(E_{low}\) occurs and for \(1 \le j \le z\) let \(E_j\) be the event that \(\mathcal{M}\) does not have any edge with an endvertex in \(\{w_j,v_j\}\) where \( w_j \in X\) and \(v_j \in Y.\) If \(E_j \cap E_{low}\) occurs, then in the random graph \(G^{(j)}\) obtained by removing the vertices \(w_j\) and \(v_j\) from \(G_{bip},\) the maximum matching \(\mathcal{M}(G^{(j)})\) in \(G^{(j)}\) is still \(\mathcal{M}.\) Consequently

$$\begin{aligned} \mathbb {P}(E_j \cap E_{low}){} & {} = \mathbb {P}\left( E_j \cap E_{low} \cap \{\mathcal{M}(G^{(j)}) = \mathcal{M}\}\right) \nonumber \\{} & {} = \sum \limits _{\mathcal{E}} \mathbb {P}\left( E_j \cap \{\mathcal{M}(G^{(j)}) = \mathcal{E}\} \cap \{\mathcal{M} = \mathcal{E}\}\right) \end{aligned}$$
(2.19)

where the summation in (2.19) is over all sets of edges \(\mathcal{E}\) of size at least \(z\) and satisfying the property that no edge of \(\mathcal{E}\) has an endvertex in the set \(\{w_j,v_j\}.\) Moreover, 

$$\begin{aligned} E_j \cap \{\mathcal{M}(G^{(j)}) = \mathcal{E}\} \cap \{\mathcal{M} = \mathcal{E}\} \subset \{\mathcal{M}(G^{(j)}) = \mathcal{E}\} \cap V\left( \mathcal{E}\right) , \end{aligned}$$

where \(V\left( \mathcal{E}\right) \) is the event that there is no edge \(e = (u,v) \in \mathcal{E}\) such that \(v_j\) is adjacent to \(u \in X\) and \(w_j\) is adjacent to \(v \in Y,\) in the graph \(G_{bip}.\) This is because if there existed such an edge \((u,v),\) then we could remove \((u,v)\) from \(\mathcal{M}\) and add the edges \((u,v_j)\) and \((w_j, v)\) to get a matching of \(G_{bip}\) whose size is strictly larger than that of \(\mathcal{E},\) a contradiction. This is illustrated in Fig. 2.

Fig. 2
figure 2

Replacing the edge \((u,v)\) with the edges \((u,v_j)\) and \((w_j,v)\) to get a bigger matching

From (2.19) and the discussion in the above paragraph, we therefore get that

$$\begin{aligned} \mathbb {P}(E_j \cap E_{low}){} & {} \le \sum \limits _{\mathcal{E}} \mathbb {P}\left( \{\mathcal{M}(G^{(j)}) = \mathcal{E}\} \bigcap V\left( \mathcal{E}\right) \right) \nonumber \\{} & {} = \sum \limits _{\mathcal{E}} \mathbb {P}\left( \mathcal{M}(G^{(j)}) = \mathcal{E}\right) \mathbb {P}\left( V\left( \mathcal{E}\right) \right) \end{aligned}$$
(2.20)

because for any \(\mathcal{E},\) the event \(V\left( \mathcal{E}\right) \) depends only on the state of edges containing an endvertex in \(\{w_j,v_j\}\) and is therefore independent of \(\{\mathcal{M}(G^{(j)}) = \mathcal{E}\},\) by definition. To estimate \(\mathbb {P}(V(\mathcal{E})),\) we let \(\mathcal{E} = \{(x_i,y_i)\},1 \le i \le t, t \ge z = \frac{n}{4},\) be the edges in \(\mathcal{E}\) and obtain that

$$\begin{aligned} \mathbb {P}(V(\mathcal{E})) \le \prod _{i=1}^{t}\left( 1-p(w_j,y_i)p(x_i,v_j)\right) \le \exp \left( -\sum \limits _{i=1}^{t}p(w_j,y_i)p(x_i,v_j)\right) . \end{aligned}$$

From the second condition in (2.15) we get that 

$$\begin{aligned} \mathbb {P}\left( V\left( \mathcal{E}\right) \right) \le e^{-d_2tp^2} \le \exp \left( -\frac{d_2p^2n}{4}\right) . \end{aligned}$$

Plugging this into (2.20) we get

$$\begin{aligned} \mathbb {P}(E_j \cap E_{low}){} & {} \le \exp \left( -\frac{d_2p^2n}{4}\right) \sum \limits _{\mathcal{E}} \mathbb {P}\left( \mathcal{M}(G^{(j)}) = \mathcal{E}\right) \\{} & {} \le \exp \left( -\frac{d_2p^2n}{4}\right) \end{aligned}$$

and so from (2.18) we therefore get that

$$\begin{aligned} \mathbb {P}(E_j) \le 4^{n}\cdot \exp \left( -\frac{d_1 n^2p}{16}\right) + \exp \left( -\frac{d_2p^2n}{4}\right) =:Q. \end{aligned}$$

Summarizing, we get that at least one of vertices \(w_j\) and \(v_j\) belong to the maximum matching of \(G_{bip}\) with probability at least \(1-Q.\) There are at most \(n^2\) edges with one endvertex in \(X\) and other endvertex in \(Y\) and so applying the union bound, we see that \(G_{bip}\) contains a perfect matching with probability at least \(1-n^2Q.\) Using the fact that \(p \ge \frac{\log {n}}{\sqrt{n}},\) we then get that \(\mathbb {P}(E_{PER}) \ge 1-e^{-D_1(\log {n})^2}\) for some constant \(D_1 > 0\) and this completes the proof of Theorem 2. \(\square \)

3 Channel Assignment Problem

In this section, we briefly describe an application of our results in the context of channel assignment problem in wireless networks. Let \(K_{n,n}\) be the complete bipartite graph with vertex sets \(X = Y=\{1,2,\ldots ,n\}\) and let \(\{Z(u,v)\}_{1 \le u,v \le n} \) be positive independent random variables with \(F_{u,v}\) denoting the distribution of \(Z_{u,v}.\)

In the context of wireless networks, the set \(X\) denotes the set of users and \(Y\) denotes the set of channels to be assigned to the users with the constraint that no two users are assigned the same channel. The random variable \(Z(u,v)\) denotes the fading gain (Goldsmith 2005) of the \(u^{th}\) user on the \(v^{th}\) channel and in many practical scenarios, the fading gains are independent but not necessarily identically distributed. It is of interest to assign the “best possible” channel to each user and one straightforward way to implement this would be to set a predetermined threshold \(\lambda \) and assign each user a distinct channel whose fading gain is at least \(\lambda .\) The natural question is whether such an assignment does in fact exist and we use perfect matchings described in the previous section to demonstrate an answer.

Indeed, let \(G_{bip}\) be the random bipartite graph obtained by retaining all edges \((u,v)\) satisfying \(Z(u,v) > \lambda \) and we set \(p(u,v) := \mathbb {P}(Z_{u,v} > \lambda ).\) If the condition (2.15) in Theorem 2 holds, then we are guaranteed a perfect matching in \(G_{bip}\) with high probability, i.e., with probability converging to one as \(n \rightarrow \infty .\) Any perfect matching in \(G_{bip}\) provides a valid channel assignment as described before and so the condition (2.15) could be interpreted as a sufficient condition for assigning each user a distinct channel, with fading gain at least \(\lambda .\)

Finally, we remark that the iterative analysis described in our paper could also be potentially extended to preferential attachment models (Piva et al. 2021) and we plan to analyze applicability to these models in the future.