1 Introduction

With the explosive growth of smart phones and tablets, mobile data traffic has been approximately doubling each year in the early 2010s. According to the last Ericsson report [1], the mobile data traffic grew around 54% between the first quarter of 2017 and the first quarter of 2018. Such a traffic growth raises big challenges to cellular networks. In dense areas, it is quite necessary to deploy heterogeneous networks [2,3,4] or to combine cellular and WiFi technologies [7, 8] to cope with traffic growth.

A lot of applications can tolerate a delay (e.g., non-urgent file download, pull services). The delay tolerance feature has already been exploited in [5,6,7,8,9,10]. In [5], the trade-off between delaying a service and minimizing the energy consumption is studied. In [6], the authors propose to combine the use of neighbor terminals (crowd computing approach) and mobility prediction to limit the delay before getting the service. In [7], a model is proposed to predict WiFi connectivity and offload cellular networks by steering delay-tolerant data onto WiFi. In [8], the authors propose an integrated architecture to migrate data traffic from cellular networks to WiFi networks and quantify the number of WiFi access points (APs) required for a city-wide WiFi offloading. The performance of offloading in [7, 8] is clearly closely related to the WiFi availability, which is better in urban than in rural environments. In [9], the authors study how to select a number of key locations in cellular networks to upgrade capacity, and shift delay-tolerant traffic to them. In [10], the authors investigate solutions for network-controlled WiFi offloading in Long Term Evolution (LTE) cellular networks when performance needs exceed the capacity of the LTE.

Both the capital and the operational expenses that are required for the deployment of WiFi or LTE small cells are significant. In rural environments, such additional expenses can be prohibitive, especially in developing countries where the monthly subscription fee should be kept as low as possible. Also in some cases, deploying micro-cells is technically difficult because of the lack of energy sources.

Data traffic generally exhibits some degrees of heterogeneity in both the time and space domains. It is well known that the dimensioning of network resource is done to cope with the traffic conditions for peak hours in the day or even for peak periods in one given day of the week. The average usage of network resource is thus generally very low.

Our objective is to analyze the capacity increase of cellular networks by exploiting user delay tolerance without deploying new base stations or access points and without adding any resource. The capacity is defined in this paper as the maximum traffic arrival rate for which the blocking probability of the system is below a target [11]. In [12], we proposed an architecture based on a specific server in the network and a mobile application that queues non-urgent download requests and determines the best time to trigger a download that is queued. We also presented a proof of concept. In [13], we analyze the impact of user delay tolerance on cellular networks but from the energy efficiency perspective. In this paper, we focus on the capacity increase that is possible by delaying the service of some users in the network.

In [14], the authors consider different priority schemes and the model they propose could be used for delay-tolerant systems. However, they assume the service rate is the same for all users in a cell. In contrast, our model considers two zones (inner and outer) with different rates, through a simple but appropriate model to take into account the effect of the radio conditions. In [15], the authors propose a call admission scheme that uses mobility prediction. The proposal shows good performance but it applies to voice calls as it considers a given number of channels and a loss system (a call is rejected if all channels are busy). Here, we consider data traffic with different scheduling policies and a queuing system. We propose an analytical model, that is of course a simplified view of reality, but is based on the widely accepted assumption of a regular hexagonal network, for which we build a Markov chain. Unlike simulation models, the way to compute all variables is explained, which ensures reproducibility.

We consider that the load of cells is mainly due to the data traffic, and focus on interactive services (for example, web browsing). Each user alternatively downloads some content and reads this content. We refer to the download request as a packet call. We assume that a call admission control is activated by the operator to guarantee a minimum bit rate. Hence, a packet call can be blocked in case of overload. Of course, enough frequency bandwidth should be allocated to the cell to ensure a low blocking probability (typically 0.1% in normal conditions).

In this paper, we refer to a user with delay-tolerant data traffic as a Delay Tolerant User (DTU). Users whose calls cannot tolerate delays are referred to as non-DTUs. Note that a user can be a DTU for a service and a non-DTU for another service. We propose to queue the call of a DTU, upon its arrival, if the user has bad radio conditions and the current load of the system is above a given threshold. The call gets served when either the radio conditions improve or the load goes below the threshold. We analyze the impact of DTUs on the capacity of cellular networks. Of course, delaying the service is not possible in all environments and all circumstances. However, due to its simplicity, it is worth studying such a solution in constrained networks (e.g., cost constraints, lack of energy, etc.).

The remainder of the paper is organized as follows. Section 2 presents the system model considered in this paper, including the mobility model, the traffic model, the scheduling strategies and the admission control policy. In Sect. 3, a Markov chain is defined for that system, and the blocking probability, mean service time and mean waiting time of DTUs are derived. Numerical results are given in Sect. 4. Finally, Sect. 5 concludes the paper.

2 System model

We consider a regular hexagonal cellular network with the Okumura–Hata propagation model [16]: the received power at the terminal is proportional to \(1/x^{\eta }\) where x is the distance to the base station and \(\eta \) is an environment-dependent parameter (typically \(\eta =3.3\)). The same frequency carrier is assumed to be used in all cells (the reuse cluster size is 1). The SINR (Signal-to-Interference-and-Noise Ratio) thus depends only on x (see Sect. 2.2.1).

Each cell is divided into two zones: the inner zone in which users have a SINR higher than a threshold and can have a high transmission rate, the outer zone in which the SINR is low and hence the rate is reduced. In a simple hexagonal cellular network (see Fig. 1), the inner zone is a disk of radius \(r_{g}\) (we use subscript g for “good” instead of i for inner to avoid any confusion with index i, which will be used in the Markov chain) and the outer zone is the complementary of the disk in the hexagon of radius r. In order to have a system simple enough to make the analysis, we consider that each mobile in the inner (resp. outer) zone gets the same rate as the one at distance \(r_g\) (resp. r).

Fig. 1
figure 1

Inner and outer zones of a cell

The inner zone is a disk whose area is

$$\begin{aligned} S_g = \pi r_g^2. \end{aligned}$$
(1)

The outer zone is defined by the part of the hexagonal cell that is not in the inner zone. Let \(S_o\) be the area of the outer zone. Thus,

$$\begin{aligned} S_o = \frac{3 \sqrt{3}}{2} \; r^2 - \pi r_g^2. \end{aligned}$$
(2)

2.1 Mobility model

We assume the residence time of a user in any type of zone (inner or outer) is exponentially distributed and that mobility is a memory-less process [17]. When a user is in the outer zone, he/she can go either in the inner zone or in a neighbor cell. In the latter case, he/she is necessarily in the outer zone of the neighbour cell. The user can go an unlimited number of times in the inner zone and then go back to the outer zone. The network is assumed to be regular and we thus consider only two cells; the current cell and a neighbor cell (which becomes the current cell as soon as the user enters it, from the user point of view). The transition rate from the outer zone to the inner zone (resp. to the neighbor cell) is denoted by \(\alpha \) (resp. \(\delta \)). The transition rate from the inner zone to the outer zone is denoted by \(\beta \). The mobility of each user is modeled by a continuous-time Markov chain as shown in Fig. 2.

Fig. 2
figure 2

Continuous-time Markov chain for the user mobility model

Let \(q_g\) and \(q_o\) be the steady state probability to be in the inner and the outer zone, respectively. We have

$$\begin{aligned} \left\{ \begin{array}{ll} q_g &{} = \frac{\alpha }{\alpha + \beta }\\ q_o &{} = \frac{\beta }{\alpha + \beta }\\ \end{array} \right. \end{aligned}$$
(3)

We consider an underlying random walk mobility model, where the distribution of the speed is uniform in all the directions. In that case [19, 20], it is possible to compute the average number of terminals \(\frac{dN}{dT}\) that cross (outwards) the perimeter L of an area S per time unit:

$$\begin{aligned} \frac{dN}{dT} = \frac{v \rho _T L}{\pi } \end{aligned}$$
(4)

where \(\rho _T\) is the density of terminals and v is the average speed. But with a Markovian model, that average number of outgoing crossings should equal the individual outgoing rate \(\omega \) of the area multiplied by the number \(\rho _T S\) of terminals in the area: \(\frac{dN}{dT}= \omega \rho _T S\). Combining with (4) and denoting by T the average dwelling time in the area, we get:

$$\begin{aligned} \omega = \frac{1}{T} = \frac{v L}{\pi S}. \end{aligned}$$
(5)

Note that (5) is very general and can be applied to any type of shape. In Sect. 2.1.1, we use it both for the inner zone (circle) and the global cell (hexagon).

Finally, since all users have the same mobility pattern, \(q_g\) is also the proportion of users in the inner zone. If the repartition of users over space is uniform, the steady state probability is clearly proportional to the area of each zone. This implies that

$$\begin{aligned} \frac{\alpha }{\beta } = \frac{S_g}{S_o} \text{. } \end{aligned}$$
(6)

2.1.1 Computation of the transition rates of the mobility Markov chain

The inner zone is a circle and thus, \(L=2 \pi r_g\). By using (5) for \(S=S_g\) and \(\beta =\omega \), we get:

$$\begin{aligned} \beta = \frac{2}{\pi } \; \frac{v}{r_g} \text{. } \end{aligned}$$
(7)

Consider a terminal that enters a cell. It is necessarily in the outer zone, where it stays on average \(1/(\alpha +\delta )\) seconds. Then it leaves the cell with probability \(\delta /(\alpha +\delta )\) or goes in the inner zone with probability \(\alpha /(\alpha +\delta )\). In the latter case it can go again in the inner zone then in the outer zone several times before leaving the cell. The dwell time in the inner zone is \(1/\beta \). Let \(T_c\) be the average cell dwell time. Using the renewal theory we can write

$$\begin{aligned} T_c = \frac{1}{\alpha +\delta } + \frac{\alpha }{\alpha +\delta } \left( \frac{1}{\beta } + T_c \right) , \end{aligned}$$
(8)

yielding

$$\begin{aligned} T_c = \frac{1}{\delta }\left( \frac{\alpha }{\beta } + 1 \right) \text{. } \end{aligned}$$
(9)

The cell crossing rate is given by \(1/T_c\) and can be computed with (5). As for an hexagon \(S=3\sqrt{3} r^2/2\) and \(L=6r\), we have:

$$\begin{aligned} \frac{1}{T_c} = \frac{4}{\pi \sqrt{3}} \frac{v}{r} \text{. } \end{aligned}$$
(10)

Given \(r_g\), r, v, it is possible to compute \(\alpha \), \(\beta \) and \(\delta \) very simply by solving a system of equations:

$$\begin{aligned} \left\{ \begin{array}{l} \beta = \frac{2}{\pi } \; \frac{v}{r_g}\\ \frac{\alpha }{\beta } = \frac{S_g}{S_o}\\ \delta \frac{1}{1+ \frac{\alpha }{\beta }} = \frac{4}{\pi \sqrt{3}} \frac{v}{r}. \end{array} \right. \end{aligned}$$
(11)

After a few elementary steps, we get

$$\begin{aligned} \left\{ \begin{array}{l} \alpha = \frac{ \frac{2}{\pi } \; \frac{v}{r_g} }{\frac{3 \sqrt{3}}{2\pi } \; \frac{r^2}{r_g^2} - 1}\\ \beta = \frac{2}{\pi } \; \frac{v}{r_g}\\ \delta = \frac{\frac{6}{\pi ^2} \; \frac{v \; r}{r_g^2}}{\frac{3 \sqrt{3}}{2\pi } \; \left( \frac{r}{r_g}\right) ^2 - 1}. \end{array}\right. \end{aligned}$$
(12)

2.2 Traffic model

2.2.1 Rate for each zone

In this subsection we compute the minimum SINR for each zone and then deduce the achievable rate. We assume that noise is negligible compared to interference; the SINR is thus equal to the Signal to Interference Ratio (SIR). In Kelif et al. [18] propose to compute the interference by considering a continuum of interference base stations. With this approach, they got a closed formula of the SIR f(x) as a function of the distance x between the terminal and the base station in an hexagonal network. The formula can be written as:

$$\begin{aligned} f(x) = 4 \pi \frac{\sqrt{3}}{3} \frac{1}{\eta -2} \, \frac{\left( \frac{x}{\sqrt{3}r} \right) ^{\eta }}{\left( 1 - \frac{x}{\sqrt{3}r} \right) ^{\eta -2}} \end{aligned}$$
(13)

where \(\eta \) is the propagation exponent and r is the cell radius.

The rate g(x) achievable at distance x is computed according to the Shannon formula:

$$\begin{aligned} g(x) = B \log _2 \left( 1+f(x)\right) \end{aligned}$$
(14)

where B is the bandwidth used for the transmission. In the inner zone, the SIR is equal to or higher than \(f(r_g)\). We consider that the rate in the whole zone is given by \(g(r_g)\). The maximum rate R is given when there is only one terminal in the inner zone (and none in the outer zone). In that case, B is the system bandwidth. We thus have:

$$\begin{aligned} R = g\left( r_g \right) = B \log _2 \left( 1+ \frac{4 \pi \sqrt{3}}{3(\eta -2)} \, \frac{\left( \frac{r_g}{\sqrt{3}r} \right) ^{\eta }}{\left( 1 - \frac{r_g}{\sqrt{3}r} \right) ^{\eta -2}}\right) \text{. }\nonumber \\ \end{aligned}$$
(15)

Similarly, the rate for one terminal in the outer zone (and no terminal in the inner zone) is given by g(r). Let \(\chi \) be the ratio between the rates in the outer zone and in the inner zone (we have \(0<\chi <1\)):

$$\begin{aligned} \chi = \frac{g(r)}{g\left( r_g\right) } = \frac{\log _2 \left( 1+ \frac{4 \pi \sqrt{3}}{3(\eta -2)} \, \frac{\left( \frac{1}{\sqrt{3}} \right) ^{\eta }}{\left( 1 - \frac{1}{\sqrt{3}}\right) ^{\eta -2}}\right) }{\log _2 \left( 1+ \frac{4 \pi \sqrt{3}}{3(\eta -2)} \, \frac{\left( \frac{r_g}{\sqrt{3}r} \right) ^{\eta }}{\left( 1 - \frac{r_g}{\sqrt{3}r} \right) ^{\eta -2}}\right) }\text{. } \end{aligned}$$
(16)

2.2.2 Practical considerations

In the model, the position of each mobile is known and the rate is a decreasing function of the distance between the terminal and the base station. The two bit-rate values are linked to geographical zones. In a real network, due to propagation phenomena (shadowing, fading), two terminals at the same distance can have different bit rates. Hence, the criterion to decide whether a DTU should be queued or not would no longer be based on the distance but on an SINR threshold. Considering these propagation phenomena is out of the scope of the paper.

2.2.3 Source model

As explained in Sect. 1, data services are modeled with transmission phases triggered by packet calls and reading periods [21]. Call arrivals are assumed to follow a Poisson process with rate \(\lambda \). Each data transmission phase is equivalent to the transfer of an average amount of data (i.e., a file) equal to F bits. That file can be a video, a web page, etc. The size of the file is modeled as an exponential random variable.

2.3 Scheduling strategies

The transmission rate needs to be shared among all users. We adopt two schedulers as proposed in [22]: equal throughput (ET) scheduler and round robin (RR) scheduler. In this paper, we always use subscript i and j to denote the number of users being served in the inner zone and in the outer zone, respectively. We denote by \(R_{g,i,j}\) and \(R_{o,i,j}\) the average rate of a user in the inner zone and in the outer zone, respectively, when i users in the inner zone and j in the outer zone are served. Note that all users in the same zone have the same bit rate.

2.3.1 Equal throughput scheduler (ET)

An ET scheduler allocates the same rate to all users. It can be seen as a round robin scheduler where one bit is transmitted to each user. The time to transmit one bit to a user in the inner zone and in the outer zone is 1/R and \(1/(\chi R)\), respectively. The total time for a round (one bit to each user) is thus \(i/R + j/(\chi R)\). The rate, which is the same for all users, is therefore:

$$\begin{aligned} R_{g,i,j} = R_{o,i,j} = \frac{R}{i+\frac{j}{\chi }}. \end{aligned}$$
(17)

Let \(\mu _{g,i,j}\) and \(\mu _{o,i,j}\) be the service rate of a user in the inner zone and in the outer zone, respectively. With ET, we have

$$\begin{aligned} \mu _{g,i,j} = \mu _{o,i,j} =\frac{R}{F(i+\frac{j}{\chi })}. \end{aligned}$$
(18)

2.3.2 Round Robin scheduler (RR)

An RR scheduler allocates the same airtime to each user (but not the same rate). Thus,

$$\begin{aligned} \left\{ \begin{array}{l} {R_{g,i,j}} = \frac{R}{ i + j}\\ {R_{o,i,j}} = \frac{\chi R}{ i + j}, \end{array}\right. \end{aligned}$$
(19)

giving the service rates

$$\begin{aligned} \left\{ \begin{array}{l} {\mu _{g,i,j}} = \frac{R}{F(i + j)}\\ {\mu _{o,i,j}} = \frac{\chi R}{F(i + j)}. \end{array}\right. \end{aligned}$$
(20)

2.3.3 Load indicator

The resource management strategy is based on the global load of the cell. In this report, as already proposed in [11], we use the harmonic mean \(R_{h,i,j}\) of the user rates as an indicator of the load. With ET, all users have the same rate and the harmonic mean is thus \(\frac{R}{i+j/\chi }\). With RR, i terminals get \(\frac{R}{ i + j}\) and j terminals get \(\frac{\chi R}{ i + j}\). By definition of the harmonic mean, we have:

$$\begin{aligned} R_{h,i,j} = \frac{i+j}{\frac{i}{R_{g,i,j}} + \frac{j}{R_{o,i,j}}} \end{aligned}$$
(21)

By combining (19) and (21), we easily get

$$\begin{aligned} R_{h,i,j} = \frac{R}{i+\frac{j}{\chi }}. \end{aligned}$$
(22)

Note that the harmonic rate is given by (22) for both ET and RR, i.e., the load indicator does not depend on the scheduling policy.

2.4 Admission control

We propose and analyze the following admission control policy:

  • in low load conditions, all packet calls are accepted and served immediately;

  • in medium load conditions, packet calls made by DTUs in the outer zone are queued (within the queue length limit) and other packet calls are served immediately;

  • in high load conditions, all calls are blocked.

More precisely, we define a maximum length \(K^{T}_{b}\) of the queue, and two thresholds \(R^{T}_{q}\) and \(R^{T}_{b}\) (with \(R^{T}_{q} > R^{T}_{b}\)) on the harmonic rate on which the admission decision will be based. Denoting by k the size of the queue at a given time, we propose that

  • when \(R_{h,i,j} > R^{T}_{q}\), all packet calls are served immediately;

  • when \(R^{T}_{b} < R_{h,i,j} \le R^{T}_{q}\), a call made by a DTU in the outer zone is queued if \(k<K^{T}_{b}\) and is blocked otherwise, while other calls are served (including calls by DTUs in the inner zone);

  • when \(R_{h,i,j} \le R^{T}_{b}\), all calls are blocked.

That policy is illustrated in Fig. 3, where we use \(1/R_{h,i,j}\) on the horizontal axis to have an indicator that increases with the load. Note that the comparison with a threshold is a linear function of i and j. In other words, \(R_{h,i,j} > R^{T}_{q}\) is equivalent to \(i+j/\chi <R/R^{T}_{q}\). Similar expressions can be found for all tests.

Fig. 3
figure 3

Illustration of the studied admission control policy

Mobility events can happen at any time. The mobility events consist of:

  • movement between the inner zone and the outer zone,

  • handover of a queued call, indicating that a user with a call in the queue of a neighbor cell is entering the current cell,

  • handover of an active call, indicating that a user in data transmission phase in a neighbor cell is entering the current cell.

In order to provide a good quality of service, the data transmission phase should not be interrupted. Hence, in the latter case, the call is accepted regardless of the load. Furthermore, a user who is served and who is switching from the inner to the outer zone (or vice-versa) is always kept in the system. If the current load is below the threshold, a queued call from a neighbor cell is immediately served. In other cases, it is put in the queue of the current cell, regardless the current length of the queue. When a queued user enters the inner zone, he/she is immediately served.

For convenience, we summarize the main notations in Table 1.

Table 1 Main notations

3 Analytical method

3.1 Markov chain

The system is modeled by a continuous-time Markov chain with states (ijk), representing the respective number of calls that are served in the inner zone, served in the outer zone, and queued. This chain has an infinite number of states. In order to allow numerical processing, we keep the Markov chain with a finite number of states. We define \(I_{\mathrm{max}}\), \(J_{\mathrm{max}}\) and \(K_{\mathrm{max}}\) as the upper bounds for i, j and k, respectively. We denote by \(\varOmega \) the set of possible values for (ijk) (\(\varOmega \subset [0,I_{\mathrm{max}}]\times [0,J_{\mathrm{max}}]\times [0,K_{\mathrm{max}}]\)). Considering a chain with a finite number of states slightly modifies the behavior of the system. For example, according to the admission control policy, handovers are always accepted and a data transmission phase is never interrupted, while for our truncated Markov chain, a handover can be blocked when the system is in state \((i, J_{\mathrm{max}}, k)\). However, \(I_{\mathrm{max}}, J_{\mathrm{max}}\) and \(K_{\mathrm{max}}\) are chosen large enough to make such blocking probabilities negligible (typically \(10^{-15}\)).

Due to the complexity of the Markov chain, it is not possible to represent it clearly in a figure. However, the transitions from state (ijk) to other states are shown in Table 2. For the sake of clarity, we consider a typical state (ijk), where \(0< i< I_{\mathrm{max}}, 0<j<J_{\mathrm{max}}, 0< k <K_{\mathrm{max}}\). For the extreme states, such as \(i=0\) or \(j = 0\) or \(k=0\), the transition rates can also be easily derived. Table 2 is valid for ET and RR: variables \(\mu _{g,i,j}\) and \(\mu _{o,i,j}\) are chosen according to the scheduler selected by using either (18) or (20).

Table 2 Transitions from state (ijk)

The infinitesimal generator \({\mathbf {Q}}\) of the Markov Chain is easily obtained from the transition rates. The stationary probability vector \(\mathbf {\theta } = (\theta _{i,j,k})\) is computed from \(\mathbf {\theta } {\mathbf {Q}} = {\mathbf {0}}\) and \(\sum _{(i,j,k)} \theta _{i,j,k} = 1 \), which can be solved by a classical method.

3.2 Arrival rates of handovers of active calls and queued calls

Let \(\lambda _H\) be the arrival rate of handovers of active calls and \(\lambda _W\) the arrival rate of handovers of queued calls from DTUs. In homogeneous networks (all cells are equivalent), the departure rate of users is equal to the arrival rate. Therefore, we have

$$\begin{aligned} \lambda _H&= \sum _{(i,j,k) \in \varOmega } j \delta \theta _{i, j,k} \end{aligned}$$
(23)
$$\begin{aligned} \lambda _W&= \sum _{(i,j,k)\in \varOmega } k \delta \theta _{i, j,k} \end{aligned}$$
(24)

Solving the Markov chain involves solving some fixed-point system (the steady-state probabilities depend on \(\lambda _H\) and \(\lambda _W\), that depend on the steady-state probabilities). We solve that problem numerically, using Algorithm 1, i.e., iterating those dependencies until variations are below a threshold. We know from simulation process that it usually requires from 4 to 9 iterations to solve the Markov chain.

figure d

In the remainder of this section, we explain how we use the obtained steady-state distribution to derive performance measures.

3.3 Blocking probability

A new call from the inner zone or from a non-DTU in the outer zone is blocked if \(R_{h,i,j} \le R^{T}_{b}\), so the blocking probability for such users is

$$\begin{aligned} P_{b,1} = \sum \nolimits _{(i,j,k) \in \varOmega _1} \theta _{i,j,k} \end{aligned}$$
(25)

where \(\varOmega _1 = \{ (i,j,k) | i + j/\chi \ge R/R^{T}_{b}, 0 \le k \le K_{\mathrm{max}}\}\).

A new call from a DTU in the outer zone is blocked if \(R_{h,i,j} \le R^{T}_{b}\) or if \(R^{T}_{b} < R_{h,i,j} \le R^{T}_{q}\) and \(k \ge K^{T}_{b}\). The blocking probability is given as

$$\begin{aligned} P_{b,2} = \sum \nolimits _{(i,j,k) \in \varOmega _2} \theta _{i,j,k} \end{aligned}$$
(26)

where \(\varOmega _2 = \{(i,j,k) | i + j/\chi \ge R/R^{T}_{b}, 0 \le k \le K_{\mathrm{max}} \ \text {or} \ R^{T}_{q} \le i + j/\chi < R/R^{T}_{b}, K^{T}_{b} \le k \le K_{\mathrm{max}}\}\).

As users are assumed to be uniformly distributed, the overall blocking probability is

$$\begin{aligned} P_{b} = (1-\rho \frac{\beta }{\alpha +\beta }) P_{b,1} + \rho \frac{\beta }{\alpha +\beta } P_{b,2} \text{. } \end{aligned}$$
(27)

We also compute the probability \(P_{s}\) that a new call from a DTU in the outer zone is served immediately (i.e., not queued):

$$\begin{aligned} P_{s} = \sum \nolimits _{(i,j,k) \in \varOmega _{s}} \theta _{i,j,k} \end{aligned}$$
(28)

where \(\varOmega _{s} = \{ (i,j,k) | i + j/\chi < R/R^{T}_{q}, 0 \le k \le K_{\mathrm{max}}\}\).

3.4 Mean service time of users in the system

Another performance metric we are interested in is the mean service time of users in the system, which includes only the transmission time when the users are served.

Let \(\overline{N_c}\) be the mean number of users being served in one cell. We have

$$\begin{aligned} \overline{N_c} = \sum \nolimits _{(i,j,k)\in \varOmega } (i+j) \ \theta _{i,j,k}\text{. } \end{aligned}$$
(29)

According to Little’s law, the mean service time \(\overline{T_c}\) in a given cell is

$$\begin{aligned} \overline{T_c} = \overline{N_c}/ \lambda _{ec} \end{aligned}$$
(30)

where \(\lambda _{ec} = \lambda (1-P_b) + \lambda _H\) is the arrival rate of callsFootnote 1.

Several handovers can happen during the data transmission phase. The mean service time \(\overline{T_s}\) in the system is thus different from the service time \(\overline{T_c}\) in a given cell. We have

$$\begin{aligned} \overline{T_s} = \sum _{n = 0}^{\infty } (n+1)p_H^n (1-p_H) \overline{T_c} = \frac{ \overline{T_c}}{1-p_H} \end{aligned}$$
(31)

where \(p_H = \lambda _H/(\lambda + \lambda _H)\) is the handover probability during the transmission phase.

3.5 Mean waiting time of queued DTUs in the system

We use a similar approach to compute the mean waiting time of queued DTUs in the system. We first compute the equivalent arrival rate \(\lambda _{eq}\) of calls for the queue in the outer zone as

$$\begin{aligned} \lambda _{eq} = \lambda \rho \frac{\beta }{\alpha +\beta } (1-P_{s}-P_{b,2}) + \lambda _W (1-P_{s}) \end{aligned}$$
(32)

By using Little’s law, we obtain the mean waiting time of queued DTUs in the considered cell as

$$\begin{aligned} \overline{W_c} = \frac{{\overline{K}}}{\lambda _{eq}} \end{aligned}$$
(33)

where \({\overline{K}}\) denotes the mean number of users in the queue and is given by

$$\begin{aligned} {\overline{K}} = \sum _{(i,j,k)\in \varOmega } k \ \theta _{i,j,k} \end{aligned}$$
(34)

We thus deduce the mean waiting time \(\overline{W_s}\) of queued DTUs in the system:

$$\begin{aligned} \overline{W_s}&= \sum _{n = 0}^{\infty } \Big \{ (n+1)[p_W(1-P_{s})]^n (1-p_W) \overline{W_c} \nonumber \\&\qquad + n[p_W(1-P_{s})]^{n-1}p_W P_{s} \overline{W_c} \Big \} = \frac{ \overline{W_c}}{1-p_W(1-P_{s})} \end{aligned}$$
(35)

where \(p_W = \lambda _W/(\lambda _W + \lambda \rho \frac{\beta }{\alpha +\beta })\) is the handover probability of queued DTUs.

4 Numerical results

The parameter values are listed in Table 3. The cell radius is 0.5 km and the inner-zone radius is 0.3215 km. This gives a proportion of users in the inner zone \(\alpha /(\alpha +\beta )=0.5\). The propagation exponent is assumed to be 3.3 and the bandwidth is 10 MHz. With (15) and (16), this gives a maximum speed rate \(R=16\) Mbit/s and a ratio between the bit rate in the outer zone and the rate in the inner zone \(\chi =0.25\). In addition, \(\lambda \) is chosen from 0.5 to 1.6 with interval 0.1.

Table 3 Parameter values

4.1 Blocking probability

4.1.1 Blocking probability against packet call arrival rate

The blocking probability against the packet call arrival rate is shown respectively in Figs. 4 and 5 for \(\rho = 0.2\) and 0.5. From both figures, we can see that the blocking probability for the ET scheduler is very close to the one for the RR scheduler. In addition, when \(\rho \) is small, the blocking probability is insensitive to the increase of queue length \(K^{T}_{b}\). Only when \(\rho \) is larger, the impact of \(K^{T}_{b}\) is obvious, see Fig. 6. Comparing Figs. 45 and 6, we can find that the blocking probability is lower if \(K^{T}_{b}\) is larger but having a large queue provides a noticeable gain only if the proportion of DTUs is high and at medium load (\( 0.8 \le \lambda \le 1.4\)), see Fig. 6.

Fig. 4
figure 4

Blocking probability of RR and ET schedulers under \(\rho = 0.2, K^{T}_{b} = 10, 15, 20\)

Fig. 5
figure 5

Blocking probability of RR and ET schedulers under \(\rho = 0.5, K^{T}_{b} = 10, 15, 20\)

Fig. 6
figure 6

Blocking probability of RR and ET schedulers under \(\rho = 0.8, K^{T}_{b} = 10, 15, 20\)

4.1.2 Blocking probability against velocity of users

In Sect. 4.1.1, we fix the velocity of users to be 1 m/s and obtain the results of blocking probability against packet call arrival rate under different \(\rho \). In this section, we aim to obtain the impact of users’ velocity on blocking probability and set \(\rho = 0.5\), \(K^{T}_{b} = 20\), \(v= 0.1, 1, 5, 10\) m/s respectively for both RR and ET schedulers.

The blocking probability under different velocities of users for the RR scheduler is shown in Fig. 7. We can see that the blocking probability decreases with the increase of users’ velocity, but the gap is not very large. Similarly, the blocking probability for the ET scheduler is given in Fig. 8, from which it can be seen that the blocking probability decreases little with the increase of users’ velocity. From both figures, we know that increasing users’ velocity can relatively decrease the blocking probability, but the impact of users’ velocity on blocking probability is very small.

Fig. 7
figure 7

Blocking probability of RR scheduler under different velocities of users

Fig. 8
figure 8

Blocking probability of ET scheduler under different velocities of users

4.1.3 Capacity increase

The results of blocking probability for the RR scheduler under \(K^{T}_{b} = 20\) and \(\rho = 0.2, 0.5, 0.8\) are shown in Fig. 9. We can see that when \(K^{T}_{b}\) is fixed, the blocking probability decreases with the increase of \(\rho \), as could be expected: having more delay-tolerant users allows more flexibility, and less blockings thanks to our queuing policy.

Fig. 9
figure 9

Blocking probability of RR scheduler under \(\rho = 0.2, 0.5, 0.8, K^{T}_{b} = 20\)

From Fig. 9, we compute the maximum arrival rates for different blocking probability targets, which represent medium to high load conditions (0.1%, 0.5%, 1%), and the relative capacity increase brought by having DTUs is shown in Table 4. The capacity increase is around 16% for \(\rho =0.2\), which is moderate, but the capacity can increase by 113% when \(\rho = 0.8\) and \(P_b = 0.1\%\).

Table 4 Capacity increase for different blocking probability targets

4.2 Mean waiting time of queued DTUs

The mean waiting time of queued DTUs under peak load for the RR scheduler is shown in Table 5 for different blocking probability targets. Here \(v=1\) m/s, \(K^{T}_{b} = 20\). In all cases, the waiting time is less than 13 s, which is quite acceptable.

Table 5 Mean waiting time of queued DTUs for the RR scheduler under peak load in seconds

We are also interested in the impact of users’ velocity on the mean waiting time of queued DTUs. We thus set \(\rho = 0.5, K^{T}_{b} = 20\), and \(v = 0.1, 1, 5, 10\) m/s to show the results for the RR and ET schedulers as in Tables 6 and 7 respectively. From both tables, we can see that the mean waiting time of queued DTUs decreases with the increase of users’ velocity. This is because when the velocity of DTUs increases, the DTUs have high probability to go from the outer zone to the inner zone and the probability to be served immediately also increases, leading to the decrease of mean waiting time. In addition, comparing the third columns of Tables 5 and 6, we can see that they are consistent.

Table 6 Mean waiting time of queued DTUs for the RR scheduler with different users’ velocities under peak load in seconds
Table 7 Mean waiting time of queued DTUs for the ET scheduler with different users’ velocities under peak load in seconds
Table 8 Mean service time of users for the RR scheduler under peak load in seconds
Table 9 Mean service time of users for the RR scheduler with different users’ velocities under peak load in seconds

4.3 Mean service time of users

The mean service time of users for the RR scheduler under peak load is shown in Table 8. Here \(v=1\) m/s, \(K^{T}_{b} = 20\). The maximum mean service time under these settings is about 1.4 s and is obtained for \(\rho = 0.5\) and \(P_b = 1\%\). As the mean size of files is 4 Mbit, the average bit rate is \(4/1.4 = 2.86\) Mbit/s, which is acceptable. Though the mean waiting time is higher for \(\rho = 0.8\) (compared with \(\rho = 0.5\)), the mean service time is relatively lower. This is a very interesting benefit of introducing DTUs: by limiting the number of users in the system who have a low rate, the average rate is higher, which is beneficial for both the system and users.

To show the impact of users’ velocity on the mean service time of users, we set \(\rho = 0.5, K^{T}_{b} = 20\), and \(v = 0.1, 1, 5, 10\) m/s. The results for the RR and ET schedulers are shown in Tables 9 and 10 respectively. From both figures, we can see that the mean service time of users under different users’ velocities changes slightly, which means that the impact of users’ velocity on the mean service time is very small. In addition, by comparing the third column of Tables 8 and 9, we can see that they are consistent.

Table 10 Mean service time of users for the ET scheduler with different users’ velocities under peak load in seconds

5 Conclusion

In this paper, we have proposed a model to analyze the benefit of having DTUs in cellular networks. We divide a cell into the inner zone where users have a higher SINR and the outer zone where users have a low SINR. When a packet call is from a DTU in the outer zone, it is queued if the current load is above a threshold. When the DTU moves to the inner zone or the load is below the threshold, its call is served. We then analyze the impact of such a policy on the system capacity. Numerical results show that when 20% of calls are DTUs, the capacity of cellular networks can increase about 16%, and when there is 80% DTUs, the capacity can even increase by 113% without affecting the blocking rate.

We have computed the mean waiting time, which is a first QoS indicator. However, users are generally sensitive to the occurrence of long delays. A next step of this work is thus to compute either the distribution of the waiting time or its standard deviation to estimate the last decile. Another extension is to consider an impatience threshold and to serve a user as soon as the experienced delay is higher than a threshold. Another possible extension of this work is to analyze the management of DTUs in a system that combines cellular networks and WiFi.