1 Introduction

Currently, the progress in micro-electromechanical systems (MEMS) technology has made it possible to develop compact sensor arrays that can detect, analyze and collect data from the environment in a more efficient and effective manner for remote communication. Wireless networks are referred to as wireless sensor networks (WSNs) [1] that do not require any kind of infrastructure and are made to share data while simultaneously monitoring physical or environmental conditions like temperature, noise, vibration, pressure, motion or pollution. Through the network, data can be monitored and analyzed at a primary location or sink [2]. A base station or sink serves as an interface between users and the network. By sending queries and collecting responses from the receiver, the necessary information is obtained from the network. There are typically hundreds of thousands of sensor nodes in a wireless sensor network. Due to a number of limitations, WSN enables new applications and necessitates unconventional protocol design paradigms [3]. The field has made significant efforts in research, standardization and industrial investment over the past decade. The majority of WSN research currently focuses on developing protocols and algorithms that are both energy and computationally efficient, and the scope of application is limited to simple monitoring and control applications. Although sensors are powered by batteries, they have limited processing speed and memory to detect or transmit data. As sensors are installed closer to the observed event, the probability of obtaining a large amount of accurately detected data increases [4]. WSNs have been increasingly used in a variety of security and surveillance-related applications ever since their inception tremendously. WSN, a collection of multi-resource strain sensor nodes, are broadly utilized in fields like ecological horticulture, far off tolerant observing, military reconnaissance, catastrophe forecast, checking and processing plant mechanization. The essential capability of sensor hubs in such applications is to assemble important natural information and send it to a base station (BS) or sink via multi-cast. This topological disadvantage causes neighboring nodes to drain energy significantly faster, which can create energy holes in the network [5]. WSN consist of multiple sensor nodes that send environmental data to a base station (BS) and collect it.

One of the most significant issues with WSNs is the need to maintain coverage. The batteries serve as the power source for the sensor nodes die when the battery is depleted; when multiple nodes fail, communication is lost and network coverage is interrupted [6]. In WSNs, coverage is ensured and maintained in two steps, with coverage hole detection carried out by a straightforward distributed system with low overhead. Nodes check to see if there are any holes around. At the point when the quantity of openings surpasses the, the openings with the most brief distance to the portable hubs and highest priority are chosen number of mobile nodes. Across the past decade, rapid advances in wireless communication and embedded micro-sensing technologies have led to smaller and cheaper wireless assemblies. Various nodal features related to existing work morphology and geometric distribution of connectivity are used to find gaps in coverage. However, the majority of algorithms call for a relatively high average node count degree, which results in high communication costs when searching for specific patterns [7]. It should be noted that no algorithm offers a technique for clustering nodes, despite the fact that all algorithms locate nodes near coverage gaps. Unfortunately, large communication overhead significantly shortens the life cycle of WSNs.

WSNs are used in advanced applications to detect hostile or unreachable locations such as environmental monitoring, rescue operations, pollution detection, outdoor surveillance, warfare, health care and home entertainment [8]. In addition, each node can collect, store, process and communicate information about the environment with neighboring nodes. In the monitoring area, see a coverage hole that is not covered by any sensor detection disk [9]. The objective is to provide effective coverage while minimizing energy consumption with the scalable firefly algorithm and mobile nodes. It involves evenly dividing the network area but smaller cells to facilitate physical coverage estimation [10]. To fill the gaps, a general research framework of reinforcement learning was hybridized with the branch and cut algorithm for network operators [11] which ensured the full coverage hole. The network operators need methods to combine and re-optimize antenna beams to improve network coverage during interference mitigation [12]. WSN hole detection (WHD) algorithm was developed to estimate the area of holes in ROIs where sensor nodes are distributed at random and to detect them. Based on energy consumption and designed to achieve quality of service (QoS), WHD average hole detection time [13]. Credible information coverage hole detection (CICHD) problem is solved and investigated based on credible information coverage (CIC) model [14]. WSNs can strictly monitor by FoI and whether WSNs can collect and transmit the necessary information are two important problems to be solved in WSN [15]. The sensor deployment schemes are blind spot centroid scheme (BCPS) and confused centroid scheme (TCPS) was used to solve the coverage problem in MWSNs [16].

Our contributions

For further enhancement, we propose an optimal cluster-based node position estimation as well as WSN coverage hole detection. The following is a discussion of the major contributions that this work relies on.

  1. 1.

    A modified Lyapunov optimization (MLO) algorithm is used for node position computation which ensures the detection of edge nodes in the network.

  2. 2.

    An optimal clustering technique is perform by using improved sand cat swarm optimization (ISCSO) algorithm to formulate efficient balanced clusters which computes coverage hole area in the network.

  3. 3.

    A hybrid deep reinforcement learning (Hyb-DRL) technique is further used for hole shape detection and hole size judgment within clusters, among clusters and along edges.

  4. 4.

    Furthermore, performance of the proposed approach is evaluated by implementing the proposed technique with the NS2 simulation tool, in which node density, node mobility and node sensing range are simulated.

The following is how the remainder of this paper is organized. Section 2 discusses the WSN coverage hole detection literature review. Section 3 covers our proposed work's problem methodology and network model. The appropriate mathematical model is then used to provide a comprehensive breakdown of the proposed work's workflow Sect. 4, in addition to the simulation outcomes and a comparison of the proposed and existing coverage hole detection methods. Section 6 shows the final section in the paper.

2 Related works

In this section, we will look at different methods for detecting coverage holes utilized for WSN environment. The summary of research gaps is listed in Table 1.

Table 1 Summary of research gaps

Wang et al. [17] have proposed a general research framework for problems with coverage gaps that have unknown characteristics to summarize related models like deployment models, detection models and network models using some fundamental ideas for coverage issues involving properties that are not clear. Improve the quality of coverage, extend the life of the network, and reduce sensor count when coverage issues are uncertain characteristics are the three goals that can be accomplished using these models. Implementation, planning or selection, movement or adjustment, and the various decision strategies are all implemented. Then, divide the solutions into traditional, heuristic, approximate, distributed, centralized and stochastic algorithms according to their various aspects. Phoemphon et al. [18] have proposed using enhanced particle swarm optimization (NS-IPSO) to segment sensor nodes in order to improve the precision with which the distances between unknown nodes and pairs of anchor nodes can be estimated to find potential sensor nodes that could be used to divide up the region's anchor nodes. These sensor nodes occur more frequently than the standard deviation of all sensor nodes because they have a shorter path between anchor nodes.

Jia et al. [19] have proposed in order to monitor partial discharges of high voltage (HV) equipment, a monitoring network was created. A strategy for installing ultrasonic sensors to monitor partial discharges is utilized in accordance with the three-dimensional characteristics of the devices. An adaptive angle adjustment-based virtual force algorithm is proposed. It has been demonstrated that applying this algorithm covers more than 94% of the target within a manageable integration range. A sensor collaboration scheme is proposed based on a standard security protocol, which verifies its energy saving and delay compression effects and guarantees the target acquisition rate. Xu et al. [20] have proposed the Gaussian kernel integration algorithm, the polynomial kernel integration algorithm and the polynomial graph-based semi-supervised manifold learning algorithm A low-dimensional physical space and a high-dimensional location data space can be linked with a polynomial mapping function be calculated using this method, which has a high discrimination rate and clear nonlinear feature mapping. The performance of this algorithm is then evaluated and compared to some related localization techniques under various signal noises, anchor nodes and communication ranges, respectively. Khalifa et al. [21] have proposed a distributed self-healing algorithm known as distributed hole detection and repair (DHDR) makes use of just existing network nodes to simultaneously find and fix holes. It accurately estimates the size and location of a cover hole is dynamically detected as it occurs. By exchanging data and coordinating their movements, it selects suitable nearby nodes that maximize coverage while consuming the least amount of energy. The selected nozzles move in such a way that they restore the hole's hollow area without affecting the connection or coverage that is already in place. Vishnupriya et al. [22] have developed the Rabin–Karp algorithm to separate the malicious sensor from the network. It validates the sensor's authenticity within the WSN and eliminates the eavesdropping attack. The cluster head (CH), who is chosen based on cooperation rank, data transmission rank and residual energy, receives maintenance data from sensor nodes. CH detection data is gathered with the help of a mobile node via intermediate channels. Thus, it reduces latency and usage of excess energy in the network. Roy et al. [23] have investigated the issues related to quality of service (QoS) and a proposed energy-proficient WSN versatile information gatherer convention for information assortment Numerous asset requirements of WSN hubs make it truly challenging to protect nature of administration, particularly because of energy restriction. Das et al. [24] have proposed a computational geometry-based and empty circle-based framework for estimating coverage holes and hole areas. The Delaunay triangle is constructed using the location information of sensor nodes to identify a coverage gap in a large-scale WSN. After locating the hole, the next objective is to estimate its area. The empty circle property can be used to determine whether or not a coverage hole exists in a specific ROI of a large-scale WSN. The coverage hole detection and area estimation (CHDAE) efficiency of the algorithm is also demonstrated through simulations. Christopher et al. [25] have come up with a low-latency Jellyfish dynamic routing protocol (JDRP) to avoid congestion and safeguard location privacy. Here, the transmission distance is calculated for each subsection, which selects the target area from the entire sensor field.

3 Problem methodology and network model

3.1 Problem methodology

Ma et al. [26] have proposed a computational geometry-based, distributed protocol for remotely detecting coverage holes in self-organized WSNs. Calculations rather than predictions regarding the presence of holes at the site are used to detect holes in the current card. Geometric sampling of just one or two neighbors from each node is used in the process of hole detection, which is straightforward but efficient. It is suggested that local node data serve as the foundation for a global coverage hole detection strategy. Each node's communication range is assumed to be equal to its detection range in contrast to other coverage hole detection protocols, allowing for greater energy savings during communication. After the deployment of the network, coverage gaps compromise this task. Coverage gaps are inevitable for a variety of reasons, including physical damage and depletion of sensor node energy. As a result, it is critical to have continuous coverage monitoring system as coverage holes can negatively affect network performance if left unattended. In general, coverage is considered as an important quality of service (QoS) measure to increase monitoring quality in the target area. Holes in WSNs can occur due to sensor placement issues or power/hardware failure. In such cases, the detection or data transmission may affect the WSN's normal operation. It can also reduce sensor detection coverage and network lifetime. Due to the fact that the precise location of sensors is frequently unknown, the hole detection problem in WSNs is particularly challenging.

Existing research [17,18,19,20,21,22,23,24,25,26] introduces a new node that joins the network or active nodes that are already in use to recover from coverage loss owing to the complications of conventional redundant hole detection methods such as extra time, low accuracy and high operating cost [19, 20]. Adding new nodes to the network when a coverage gap occurs not possible if the region of interest is in an unfavorable location [23]. Changing or changing the detection range of active nodes not only changes the network topology constantly introduces new coverage gaps and increases coverage overhead [22]. Maintaining network coverage, energy wireless sensor networks face significant difficulties in terms of lifetime and power consumption [17,18,19,20]. Since the sensor node battery cannot be replaced or recharged, the battery discharge lasts for the lifetime of the sensor node. The network coverage is reduced when some sensor nodes fail and disconnect decreases compromised [26]. Specifically, when a sensor node fails the main objective of the coverage and communication optimization model is to select some sensor nodes with maximum direct sensor node connectivity [25]. However, the existing methods do not allow obtaining a minimum selection of nodes, thereby alleviating the bottlenecks of traditional coverage methods. Hole detection is performed by taking one or two neighboring hops from each node using simple but effective geometric methods [26]. However, this method is not suitable for multi-hop routing and cannot be detected. To our knowledge, very few algorithms have suggested detecting coverage holes. Additionally, the majority of these protocols take into account a typical monitoring area, rather than the actual sensor location randomly aligned, it is more convenient to obtain a region of random arrangement. To overcome above problems by optimal cluster-based node position estimation and coverage hole detection approach using hybrid deep learning technique. The major objectives of our proposed approach are summarized as follows:

  1. 1.

    The joint optimization of node position estimation and coverage hole detection approach is used to avoid random deployment and node failure.

  2. 2.

    To design optimization algorithm to compute the exact location of sensor nodes in the network this ensures the edge nodes

  3. 3.

    To introduce clustering algorithm to detect the coverage holes in the network and formulate objective function

  4. 4.

    The hybrid deep learning technique finalizes the hole shape and hole size judgment for further decision-making which improves detection accuracy

3.2 Network model

We assume that n sensor nodes are designed to detect particular events and are arranged at random R in a rectangular field. Within its transmission range, every sensor hub can speak with different hubs and detect certain events. The sensor hubs are sufficiently thick and there are no disconnected sensor hubs. Sensor hubs realize their area utilizing GPS or other following strategies.

We also assume that every sensor hub realizes point by point data including the IDs and directions of its communication neighbors. This information can be collected through a beacon, a hello message or a series of topological changes expected. During network training, every hub learns its area data and gathers a rundown of one-bounce and two-jump neighbors. Figure 1 shows the organization model of our proposed way to deal with explore a multi-bounce remote organization with hubs with various handsets. The hubs of the objective locale are haphazardly disseminated inward hubs, and different hubs are consistently appropriated along the external limit of the objective district to guarantee total inclusion. Every hub does not know explicit area data, so a hub might be set as the default interior hub. The discovery range (Rs) is equivalent to the correspondence range (Rc), and every hub sends its area data through GPS or some other area data framework. In this organization model, regardless of whether a piece of the organization has an inclusion opening, the whole organization is completely associated. In any case, it is accepted that no separated sensor shows up in the organization on the grounds that the whole organization is completely associated, and the spotted circle between the two demonstrates twofold series correspondence.

Fig. 1
figure 1

Network model of proposed approach

4 Proposed methodology

In this segment, we portrays the functioning system of proposed optimal cluster-based node position estimation and coverage hole detection approach, which consists three fold process are node position estimation, coverage hole detection, hole shape detection and hole size judgment.

4.1 Node position estimation

In this stage, every hub gathers and decides the vital data about its current circumstance, which is utilized to set up the following stage, track down undesirable neighbors and lastly decide the convergence focuses with the opening. Most existing location algorithms can be classified as threshold-based or threshold-based, depending on whether the algorithm uses distance estimation or other information to calculate node locations. Therefore, it is necessary to develop new techniques, methods and methods to solve the localization and localization problem of wireless sensor nodes. In this work, the sensor node position estimation is depends on three important design constraints are received signal strength, interference range and distance between sensor nodes, sink node. Here, we noted that all the design metrics are considered as the time varying factors, so need to optimize it. For that reason, a modified Lyapunov optimization (MLO) algorithm is used and estimates the exact node position and ensures the detection of edge nodes in the network.

4.1.1 Design constraints for node position estimation

Received signal strength (RSS) is commonly used standard because it is easy to measure and directly related to service quality. RSS feeds and mobile terminals are closely related to its connection point. The power consumption of transmit and receive domains is derived from the basic output power model taking into account the power demand. The energy consumption of sensor nodes depends on the amount of data and the transmission spacing. The vigor ingestion of a node (n) is comparative with square of distance when the expansion distances (I) not exactly the starting distance. The total energy consumption of each sensor node is compute as follows.

$$E_{{{\text{total}}}} = S\,(n,\,I) + R\left( n \right)$$
(1)

where \(S\,(n,\,I)\) and \(R\,\left( n \right)\) are energy ingestion of transmission and acquire node.

$$S\left( {n,I} \right) = \left\{ {\begin{array}{*{20}c} {n \times E_{{{\text{elec}}}} + n \times \varepsilon_{{{\text{fs}}}} \times I^{2} ;\quad {\text{if}}\quad I < I_{0} } \\ {n \times E_{{{\text{elec}}}} + n \times \varepsilon_{{{\text{mp}}}} \times I^{4} ;\quad {\text{if}}\quad I \ge I_{0} } \\ \end{array} } \right.$$
(2)
$$R\,\left( n \right) = n \times E_{{{\text{elec}}}}$$
(3)

The power required to operate the transmitter or receiver circuit per unit area, denoted by S(n,I), determines the power consumption for both the escaped space and multi-path models, which is influenced by the source communication model and initial communication distance. It is important to note that all costs considered in this study are benign for energy consumption. The RSS metric is independent of distance and communication energy. When the node transmits packets with power, the received signal strength (RSS) is determined by the distance “I,” and is calculated as follows:

$${\text{RSS}} = \frac{{S\left( {n,I} \right)}}{{4\pi \,i_{i}^{2} }} + s_{{a,\,a_{1} /a_{2} }}$$
(4)

The sign strength of the ongoing example not entirely set in stone by the development, distance and relative speed, and the sample points are selected and controlled \(\Delta s_{1} = \Delta s_{2} = \Delta s\), but such points are not present in the sample field. Different reference points are used for the signal strength actually received from the potential nodes, where \(I_{{i_{1} }} ,\,I_{{i_{2} }}\) and \(I_{{i_{3} }}\) can be acquired and adapted distance is figured from the cosines regulations (A) as follows:

$$I_{{i_{1} }}^{2} = I_{{i_{2} }}^{2} + A_{1} A_{2}^{2} - 2I_{{i_{2} }} \cdot A_{1} A_{2} \cdot \cos \,\left( \alpha \right)$$
(5)
$$I_{{i_{3} }}^{2} = I_{{i_{3} }}^{2} + A_{1} A_{2}^{2} - 2I_{{i_{2} }} \cdot I_{1} I_{2} \cdot \cos \,\left( \beta \right)$$
(6)

The speed and distance of the sensor nodes are determining factors for reaching the target position (V), which describes as follows.

$$2A_{1} A_{2}^{2} = I_{{i_{1} }}^{2} + I_{{i_{2} }}^{2} - 2I_{{i_{3} }}^{2}$$
(7)
$$V = \sqrt {\frac{{2\,\left( {I_{{i_{1} }}^{2} + I_{{i_{2} }}^{2} - 2I_{{i_{3} }}^{2} } \right)}}{2\Delta s}}$$
(8)

The motion continuance \(S_{{a,\,a_{1} /a_{2} }}\) for sensor node from genuine spot to the impacted position or is expressed as the distance separated the hub's speed and it can secure by sign regulation as follows:

$$S_{{a,\,a_{1} /a_{2} }} = \frac{R \cdot \sin \vartheta }{{\sin \beta \cdot V}}$$
(9)

Finally, we get the following mobility (M) function as,

$$M = S_{{a,\,a_{1} /a_{2} }} = \frac{\Delta s \cdot R \cdot \sin \vartheta }{{\sin \beta \cdot \sqrt {\frac{{\,\left( {I_{{i_{1} }}^{2} + I_{{i_{2} }}^{2} - 2I_{{i_{3} }}^{2} } \right)}}{2}} }}$$
(10)

Interference range is compute from the sensor node coverage range. Assume there are two sensors (0, 0, 0) and (1, 0, 0). The awareness scope of the two sensors is characterized as follows.

$$y^{2} + x^{2} + z^{2} = r_{i}^{2}$$
(11)
$$(Y - Y_{ji} )^{2} + X^{2} + Z^{2} = R_{i}^{2}$$
(12)

The distance between where the two sensor congregations structure a three-layer focal point is equipped for estimating the state of the two circular covers.

$$G_{j} = \frac{{(R_{j} - R_{i} + Y_{ji} )(R_{j} - R_{i} + Y_{ji} )}}{{2Y_{ji} }}$$
(13)
$$G_{i} = \frac{{(R_{i} - R_{j} + Y_{ji} )(R_{j} + R_{i} - Y_{ji} )}}{{2Y_{ji} }}$$
(14)

The congestion rate is underneath and is between them.

$$v_{{{\text{overlap}}}} = v(R_{j} ,\,G_{j} ) + v(R_{i} ,\,G_{i} )$$
(15)

The updated solution is compute as follows:

$$v(R,\,G) = \frac{\pi }{3}G^{2} (3R - G)$$
(16)
$$v_{{{\text{overlap}}}} = \frac{\pi }{{12Y_{ji} }}(R_{j} - R_{i} - Y_{ji} )^{2} (Y_{ji}^{2} + 2Y_{ji} R_{i} - 3R_{i}^{2} + 2Y_{ji} R_{j} + 6R_{j} R_{i} - 3R_{i}^{2} )$$
(17)

In the event that the sensor span is no different for all sensors because of balance and change the awareness range as (\(R_{j} = R_{i} = R\))

$$v_{{{\text{overlap}}}} = \frac{\pi }{12}(2R - Y_{ji} )^{2} (Y_{ji} + 4R)$$
(18)

The typical relationship coefficient is communicated as follows:

$$\rho (j,\,i) = \frac{{v_{j}^{i} + v_{i}^{j} }}{v}$$
(19)

where \(v_{{{\text{overlap}}}} = v_{j}^{i} + v_{i}^{j}\) and update solution as follows

$$\rho (J,\,I) = \frac{1}{{16R^{3} }}({\text{cor}} - y_{ji} )^{2} (y_{ji} + 2 \times {\text{cor}})$$
(20)

where cor = 2R is the switch boundary.

$$\rho (j,\,i) = \left\{ {\begin{array}{*{20}l} {\frac{1}{{16R^{3} }}({\text{cor}} - y_{ji} )^{2} (y_{ji} + 2 \times {\text{cor}}),} \hfill & {\quad {\text{if}}\quad 0 \le y_{ji} < {\text{cor}}} \hfill \\ {0,} \hfill & {\quad {\text{if}}\quad y_{ji} \ge {\text{cor}}} \hfill \\ \end{array} } \right\}$$
(21)

The distance among sensors and sink hub is figure by utilizing the conveyance proportion calculation. The closeness of the connecting terminal is displayed as follows:

$$n_{q} = \frac{1}{{n_{{{\text{sos}}}} }}\sum\limits_{j = 1}^{{n_{{{\text{sos}}}} - 1}} {{\text{dist}}(n,j)}$$
(22)

where which gauges the distance among Tn and its neighbor Tn is without a doubt the quantity of touching Tn. Now, we applied modified Lyapunov optimization (MLO) algorithm for node position computation which ensures the detection of edge nodes in the network.

Lyapunov enhancement algorithm allows to the utilization of Lyapunov capability to control a unique framework ideally. Lyapunov capabilities are generally utilized in charge hypothesis to guarantee the stability of various types of systems. A multi-dimensional vector often describes the state of a system at a particular moment in time. In contrast, the MLO algorithm is inspired by the basic Lyapunov function, which is a nonnegative proportion of this multifaceted state. As a general rule, activity develops when the system changes to undesirable conditions. The stability of the system is obtained by performing control measurements, which cause the Lyapunov function to deviate in the negative direction toward zero. Here, Lyapunov capability of the line as. The persistent time Lyapunov float generator is characterized as,

$$\Delta (P) = \mathop {\lim }\limits_{\delta \to 0} \frac{E[v(P)(S + \delta )) - V(P(S))|P(S)]}{\delta }$$
(23)

By the going with lemma, the Lyapunov float ∆(Q) can similarly be gotten from the going with SDE. Following from the Lyapunov smoothing out framework, we add the discipline term to obtain the float notwithstanding discipline term, i.e., where v > 0 is control limit to control the power–defer trade-off. Then, we have the accompanying arrangement in regard to the float in addition to punishment term.

$$\Delta v(P) \le \Theta + VE[q_{{{\text{sos}}}} |P] - m = \sum\limits_{{}}^{m} {P_{m} E[U_{M} - {\mathbf{B}}_{M} |P]}$$
(24)

To apply Lyapunov advancement hypothesis, we initially change the drawn out typical imperatives into virtual sequences. Two virtual arrays A(s) and Q(s) can be defined below the battery and BER threshold, respectively.

$$A(s + 1) = \max \{ A(s) + \theta - a(s + 1),0\}$$
(25)
$$Q(s + 1) = \max \{ Q(s) + \phi (s) - \delta ,\,\,0\}$$
(26)

Virtual queues can be considered as signals to determine whether constraints have been encountered in previous time slots. At higher A(s), battery voltage is sacrificed due to earlier continuous transfer. Also, at the network level, each BS maintains a set of internal queues to store the current backlog of its users. Let \(Q_{NL} (s)\) represents the current backlog of lth user in nth BS. Then the evolution of the size of \(Q_{NL} (s)\) is given by

$$Q_{NL} (s + 1) = \max [Q_{NL} (s) + R_{NL} (s),\,0] + b_{NL} (s),$$
(27)

For all N ∈ n and L ∈ l(N), where is the transmission rate offered to the ith user of the nth BS in the time slot, we adopt the concept of strong stability, and the network is very stable.

$$\overline{P}_{NL} = \mathop {\lim \,\,\sup }\limits_{s \to \infty } \frac{1}{s}\sum\limits_{\tau = 1}^{s} {e\{ P_{NL} (\tau )\} < \infty ,\,\,\,\,N \in n,\,\,L \in l(N)}$$
(28)

There the waiting depends on the control policy, which is related to the random state of the channel and the control actions taken in response to those channel states. Intuitively, this expression means that a sequence is strongly stationary if its time mean is finite. The network is very stable if all individual sequences of the network are highly stable. Let \(Y_{N} (s)N \in n\) be virtual queues associated with constraint. We update the virtual queue \(Y_{N} (s)\) for all N ∈ n at each time slot as

$$Y_{N} (s + 1) = {\text{Max}}[Y_{N} (s) - y_{N}^{{{\text{out}}}} (s),\,\,0] + y_{N}^{{{\text{in}}}} (s)$$
(29)

Likewise, to ensure inequality constraint and define virtual queues \(X_{N} (s)N \in n\); and update \(X_{N} (s)\) for all n ∈ N according to the following dynamics

$$Y_{N} (s + 1) = {\text{Max}}[X_{N} (s) - x_{N}^{{{\text{out}}}} (s),\,\,0] + x_{N}^{{{\text{in}}}} (s)$$
(30)

Then, we define a quadratic Lyapunov function \(l(\Theta (s))\) as,

$$l(\Theta (s)) = \frac{1}{2}\left[ {\sum\limits_{N \in n} {\sum\limits_{L \in l(n)} {P_{NL} (s)^{2} + \sum\limits_{N \in n} {Y_{N} (s)^{2} + \sum\limits_{L \in l(n)} {X_{N} (s)^{2} } } } } } \right]$$
(31)

The Lyapunov function is a metric that measures network congestion. Intuitively, if is short, then all lines are short. If larger, then at least one row is larger. The drift of the Lyapunov function (i.e., the expected change from one point of the Lyapunov function to another) can be written as.

$$\Delta (\Theta (s)) = e\{ l(\Theta (s + 1)) - l(\Theta (s))|\Theta (s)\}$$
(32)

We use the introduced drift plus penalty minimization method to solve the problems. This control principle solves the problem by reducing the bottom drift and the upper bound on the penalty exposure.

$$\Delta (\Theta (s)) = - v\sum\limits_{N \in n} {e\{ \log (\mu_{N} (s))|\Theta (s)\} }$$
(33)

where V ≥ 0, subject to the constraint in each time slot. The working process of our proposed node position estimation using MLO is described in Algorithm 1.

figure a

4.2 Cluster-based coverage hole area detection

Here, we utilized optimization techniques to identify areas with no coverage by the deployed sensors. In this approach, the network is divided into clusters of nodes, and the coverage hole area within each cluster is computed. The goal is to detect and localize coverage holes as accurately as possible, which is important for maintaining network connectivity and optimizing the use of resources. The cluster-based approach is preferred over a centralized approach because it reduces the amount of data that needs to be transmitted to a central location, which in turn reduces the energy consumption of the nodes. Moreover, the distributed approach is more robust to node failures or network partitions, as each cluster can operate independently. The cluster-based coverage hole area detection technique involves several steps, including cluster formation, coverage hole detection and hole area computation.

Improved sand cat swarm optimization (ISCSO) is a metaheuristic optimization algorithm that is inspired by the hunting behavior of sand cats. ISCSO is used in this paper for cluster-based coverage hole area detection in wireless sensor networks. The goal of this algorithm is to generate optimal cluster heads with maximum coverage and minimum overlap. ISCSO uses a set of sand cat individuals, which move in search of the optimal solution. The algorithm initializes a population of sand cat individuals, and each individual represents a potential cluster head. These individuals move in the search space, which is defined by the network area. Each individual evaluates the fitness of the solution, which is measured by the coverage area and the overlap with other clusters. The algorithm uses a combination of exploration and exploitation strategies to find the optimal solution. The exploration strategy is based on the movement of sand cat individuals in the search space, while the exploitation strategy is based on the selection of the best individuals to generate new solutions. ISCSO has shown promising results in solving optimization problems, and it has been applied in various fields such as engineering, finance and image processing. In this paper, ISCSO is used to generate optimal clusters for coverage hole detection in wireless sensor networks, which improves the overall performance of the system. In the early stages of breeding, groups of cats can achieve strong universal optimization capabilities.

$$R_{K} = \frac{{1 - (\sqrt {C_{K} + 1} ) \cdot \,\,{\text{rand}} \cdot j}}{{C_{K} \cdot {\text{MaxIter}}}},\quad K = 1,\,2,\,j < \frac{{{\text{MaxIter}}}}{2}$$
(34)

The equations can be used to change the ability to search globally in the early stages of particle moving and to improve local refinement and resolution accuracy in subsequent iterations. Added a radius limits to the sensor nodes search location and position. When the distance \(y_{j}^{D}\) and \(h_{{{\text{best}}}}^{D}\) radius are less than the individual \(y_{j}^{D}\) to \(h_{{{\text{best}}}}^{D}\). However, when the distance \(y_{j}^{D}\) and \(q_{{{\text{best}}}}^{D}\) radius between them is small, deviate from \(q_{{{\text{best}}}}^{D}\). The value of the elements is determined by the following equation:

$$\begin{aligned} V_{j}^{D} (s + 1) & = \omega \cdot V_{j}^{D} (s) + C_{1} \cdot R_{1} \cdot (q_{{{\text{best}}}}^{D} (s) - y_{j}^{D} (s)) \\ & \quad + C_{2} \cdot R_{2} \cdot (h_{{{\text{best}}}}^{D} (s) - y_{j}^{D} (s)) + F \cdot f_{j} + E \cdot e_{j} , \\ \end{aligned}$$
(35)

If the underlying upsides of and are moderately little, add them to control the arrangement of negative qualities to stay away from negative numbers.

$$R_{1} = R_{2} = {\text{rand}},\quad R_{1} \le 0,\,\,R_{2} \ge 0.$$
(36)

Particles study each other to get the most enlightening data in their fields. The memory component of x is added to each finder, giving each particle a smaller memory load than before. Specifies more memory weights are used to improve the current level and historically optimal level of each query.

$$y_{j}^{D} (s + 1) = \frac{1}{2}(x \cdot y_{j}^{D} (s) - (1 - x) \cdot y_{j}^{D} (s - 1) + x \cdot V_{j}^{D} (s + 1) + (1 - x) \cdot V_{j}^{D} (s))$$
(37)

First, start with each cat approximately and compute the cost of the exercise. Finally, adjust its settings in tracing mode. After adding a section, we propose an advanced transfer learning model that applies three improved strategies to this small sketch data set. It consists of two parts: sample selection and sample refinement. The second term G is constructed so as not to contribute to the BCs, since \(x_{s} (y)\) satisfy them. This term f can be generated using the ISCSO algorithm and its weight and bias must be adjusted to solve the minimization problem. Fitness function for a given input y ISCSO algorithm defines,

$$n = \sum\limits_{j = 1}^{g} {\nu_{j} \sigma (w_{j} ),\,\,\,\,\,\,\,\,\,\,}$$
(38)
$$\,w_{j} = \sum\limits_{i = 1}^{N} {w_{ji} y_{i} + a_{j} }$$
(39)

\(w_{ji}\) input unit j represents the load that connects the hidden unit to j, \(\nu_{j}\) the input unit j represents the load that connects j to the output unit, \(a_{j}\) the hidden unit represents the dependence of j, and \(\sigma (w)\) is a sigmoidal transfer function (tansig). However, when the distance between \(x_{i}^{c}\) and \(p_{{{\text{best}}}}^{c}\) is less than the radius, just deviate it from \(p_{{{\text{best}}}}^{c}\). The value of fitness elements is distinct by,

$$U_{i}^{c} (s + 1) = \omega \cdot U_{i}^{c} (t) + D_{1} \cdot r_{1} \cdot (p_{{{\text{best}}}}^{c} (t) - x_{i}^{c} (t)$$
(40)

To find a small solution to the problem of exercise activity, which is defined as follows:

$$Fs_{{\text{a}}} = Fs_{\max } ;\;\;\;{\text{otherwise}},\,\,Fs_{{\text{a}}} = Fs_{\min }$$
(41)

The position mode in which the target is detected is called tracking mode. This process can be summarized in three steps. Update the speed of each measure according to the optimal level to transfer to the entire group of sand cats, i.e., the optimal solution currently found:

$$U_{i}^{c} (s + 1) = U_{i}^{c} (t) + r \cdot d(x_{{{\text{best}}}}^{c} (t) - x_{i}^{c} (t))\,\,\,\,C = 1,\,2,\,3 \ldots m$$
(42)

Verify the threshold conditions, if the speed is within the maximum speed limit. Update the sand cat's position as follows:

$$x_{i}^{c} (s + 1) = x_{i}^{c} (t) + U_{i}^{c} (t + 1))$$
(43)

This improvement not just works on the exhibition and incorporation of the calculation, yet additionally keeps a comprehension of appropriation strength. The functioning system of cluster formation and coverage hole detection using ISCSO is described in Algorithm 2.

figure b

4.3 Hole shape detection and hole size judgment

Hole shape detection refers to the process of identifying the shape or geometry of a coverage hole in a WSN. In other words, it involves determining the boundaries or outline of the area in which there is no coverage by the sensor nodes. Hole size judgment, on the other hand, is the process of estimating the size or area of the coverage hole detected in the WSN. This information is useful for optimizing the placement of additional sensor nodes or adjusting the transmission power levels to ensure complete coverage of the network. By accurately detecting the shape and size of the coverage hole, network managers can identify the best locations to deploy new nodes or reposition existing ones to improve network coverage and reliability. Additionally, this information can be used to optimize routing protocols to ensure that data is transmitted through the most reliable and efficient path in the network. Therefore, hole shape detection and size judgment are essential components of any coverage hole detection algorithm for WSN.

The long short-term memory (LSTM) is a type of recurrent neural network (RNN) that is well suited for processing and making predictions based on sequential data. In the context of reinforcement learning, the LSTM can be used as part of an agent's architecture to model and predict future states and rewards. In the case of determining the shape and size of holes in WSNs (wireless sensor networks), the Hyb-DRL (hybrid deep reinforcement learning) approach combines LSTM with other deep reinforcement learning techniques to address this problem. Specifically, the layer details of the Hyb-DRL model may involve the following components:

  • LSTM Layers: The LSTM layers are responsible for capturing and learning sequential patterns in the input data. These layers allow the model to retain information over a certain time window, making it suitable for processing time series data.

  • Deep Reinforcement Learning Layers: These layers typically include fully connected layers and other nonrecurrent layers that facilitate the training and decision-making process in reinforcement learning. They can help in learning the optimal policy for determining the shape and size of holes in WSNs based on the current state and potential rewards.

  • Output Layers: The output layers of the Hyb-DRL model provide the final predictions or decisions based on the input data and learned policies. In the context of determining the shape and size of holes in WSNs, the output layers may produce the desired configurations or parameters for the network layout.

It is important to note that the specific layer details of the Hyb-DRL model may vary depending on the implementation and specific requirements of the problem. The mentioned components provide a general overview of how LSTM and reinforcement learning can be combined to tackle the challenge of determining the shape and size of holes in WSNs.

In the context of coverage hole detection in WSNs, Hyb-DRL can be used to accurately detect the shape and size of coverage holes within clusters, among clusters and along edges. The Hyb-DRL algorithm is used to identify the shape and size of coverage holes by analyzing the sensor data and making decisions based on the reward signals. The deep neural network is trained using the Q-learning algorithm to optimize the decision-making process and improve the accuracy of the coverage hole detection. Hyb-DRL is particularly effective in situations where there is a large amount of data and complex decision-making processes involved. By combining the principles of deep learning and reinforcement learning, Hyb-DRL can identify complex patterns and make accurate decisions, making it a useful tool for coverage hole detection in WSNs. To update the gate's value affects how much information is brought in from the previous moment. We can determine the update fitness by:

$$z_{s} = \sigma (W_{z} \cdot [G_{s - 1} ,\,Y_{s} ] + A_{z} )$$
(44)

where \(w_{z}\) and \(a_{z}\) are the update gate's bias and weight matrix, respectively. "(X) = 1/[1 + exp(X)]" denotes the Sigmoid function, which serves as the gate-control signal by transforming the data into values between 0 and 1. The reset door is utilized to control the amount of the concealed layer data from the past second should be failed to remember you can sort it out by:

$$r_{s} = \sigma (W_{R} \cdot [G_{s - 1} ,\,Y_{s} ] + A_{R} )$$
(45)

where \(W_{R}\) and \(A_{R}\) are the reset gate's bias and weight matrix, respectively. The sigmoid will set the reset gate's output to 0 to erase previous moment's information about the hidden state.

$$\tilde{G}_{s} = \tan G(W_{g} \cdot [G_{s} \circ G_{s - 1} ,\,Y_{s} ] + A_{g} )$$
(46)

where \(W_{R}\) and \(A_{R}\) are, respectively, the candidate output state's bias and weight matrix, and tanh calls the function that scales the data between 0 and 1. Hyb-DRL output layer contains the ideal secret layer state, not entirely set in stone by and. The following is the mathematical expression:

$$G_{s} = (1 - z_{s} ) \circ G_{s - 1} + z_{s} \circ \tilde{G}_{s}$$
(47)

The larger \(z_{s}\) is, the higher level of reliance on \(G_{s}\) on \(\tilde{G}_{s}\) is, and ht1 plays a smaller role in determining the output.\((1 - z_{s} ) \circ G_{s - 1}\) The current node's information, ht, points to selective memory for the previous hidden state. The forward Hyb-DRL layer stores moment t and the input sequence's previous moment, while the subsequent moment is stored in the backward Hyb-DRL layer. The process of propagating hidden layers in Hyb-DRL classifier is,

$$\mathop{G}\limits^{\rightharpoonup} _{s} = f(\mathop{W}\limits^{\rightharpoonup}Y_{s} + \mathop{V}\limits^{\rightharpoonup}\mathop{\mathop{G}\limits^{\rightharpoonup}}\nolimits_{s - 1} + \vec{A})$$
(48)
$$\mathop{G}\limits^{\leftharpoonup} _{s} = f(\mathop{W}\limits^{\leftharpoonup}Y_{s} + \mathop{V}\limits^{\leftharpoonup}\mathop{\mathop{G}\limits^{\leftharpoonup}}\nolimits_{s - 1} + \mathop{A}\limits^{\leftharpoonup} )$$
(49)
$$X_{s} = \sigma (\mathop{G}\limits^{\rightharpoonup} _{s} ,\,\mathop{G}\limits^{\leftharpoonup} _{s} )$$
(50)

where \(\mathop{G}\limits^{\rightharpoonup} _{s}\) and \(\mathop{G}\limits^{\leftharpoonup} _{s}\) indicate the forward and in reverse estimation stowed away layer states, separately; mean, in both forward and reverse computations, the heaviness of the contribution as well as the previous state of the hidden layer; \(\vec{A}\) and \(\mathop{A}\limits^{\leftharpoonup}\), respectively, denote the forward and backward calculation bias. Qu-bit state is used in ensemble learning. A bit state is also frequently controlled by two quantum logic gates, a controlled NOT gate with two Qu bits and a cycle gate with one bit. The state of the Ith kv-bit neuron model in Mth sets is defined as follows based on these gates:

$$Z_{i}^{M} = f\left\{ {\frac{\pi }{2}\delta_{i}^{M} - \arg \left[ {\sum\limits_{K} {f(\theta_{K,i}^{M} )f(z_{K}^{M - 1} ) - f(\lambda_{i}^{M} )} } \right]} \right\}s$$
(51)

here \(z_{K}^{M - 1}\) is the controlled NOT gate's reversal parameter, Input of Kth neuron in (m−1) sets. This is the best step for cyclic gate phase parameters and threshold parameters. The output O layer of the network is denoted by m observed state from the jth neuron of the output layer. \(UNN_{i}\) is represented follows:

$${\text{UNN}}_{i} = |jm(Z_{i}^{0} )|^{2}$$
(52)

The best parameters are sought when training the multilayer quantum neural network.\(\theta_{K,i}^{M}\),\(\delta_{i}^{M}\) and \(\lambda_{i}^{M}\) that make the subsequent cost function smaller:

$$i(w) = \frac{1}{2}\sum\limits_{q} {\sum\limits_{i} {\{ U_{{d_{i} }} (Q) - UNN_{i} (w,q)\}^{2} } }$$
(53)

where \(U_{{d_{i} }}\) is the instruction signal for the ith neuron in the qth pattern. New members of the ensemble crossover in the real-coded crossover \(\omega_{C}\) are created by individuals with multiple parents.

$$\omega_{C} = \omega_{h} + \sum\limits_{j = 1}^{n + k} {\xi_{j} (\omega_{j} - \omega_{h} )}$$
(54)

The just generation gap model is the one that assumes that children replace parents in each generation alternation model used. The cost function's reciprocal defines the fitness function \(F(\omega_{j} )\), where the number of people is denoted by j. We consider the following discrete-time factory as the target system to be controlled in the design of a direct quantum neural network controller:

$$X(k + d_{q} ) = F_{q} [X(k),...,X(k - n_{q} + 1),\,u(k),...,u(k - m_{q} - d_{q} + 1)]$$
(55)

where x is the factory output, U is the factory input, factory commands, K is the model number and the factory idle time is a function describing it dynamic control variable. Algorithm 3 describes the working process involved in the coverage hole shape and hole size judgment.

figure c

5 Results and discussion

In this section, the performance of proposed optimal cluster-based node position estimation and coverage hole detection (named as OC-NP-CHD for results explanation purpose) approach is validated through different simulation scenarios and measures. Our proposed OC-NP-CHD approach is implement and simulated using NS-2.33 simulation tool. Besides, in order to justify the effectiveness of our proposed OC-NP-CHD approach in terms of simulation results, the reenactment results are contrasted and the benchmark inclusion opening location draws near, for example, DHC [27], PS [28], DCHD [29] and CG-DCHD [26].

5.1 Simulation setup

Table 2 provides a detailed description of the simulation setup and parameters used for validating the proposed OC-NP-CHD approach. The network size is set to 500 × 500 m2, which is a reasonable size for a wireless sensor network. The number of sensor nodes is varied from 600 to 1000 in increments of 100 to evaluate the performance of the proposed approach under different network densities. The number of simulation rounds is varied from 1000 to 5000 to obtain statistically significant results. The number of coverage holes is varied from 10 to 50 in increments of 10 to evaluate the ability of the proposed approach to detect coverage holes of different sizes. The sensing range and communication range of each sensor node are set to 20 m to simulate the radio coverage of the wireless sensor network. The IEEE 802.15.4 MAC/PHY specification is used for the simulation, and the ad hoc on-demand distance vector (AODV) routing protocol is used for routing packets. The initial energy of each sensor node is set to 10 J, which is a reasonable value for a wireless sensor node with a battery-powered energy source. The energy cost for control packets is set to 0.3 J to simulate the energy consumption of control packets in the network. The data rate is set to 250 Kbps, which is a reasonable value for a wireless sensor network. The simulation is run for a total time of 500 s, which provides sufficient time to evaluate the performance of the proposed approach.

Table 2 Simulation setup and parameters

5.2 Comparative analysis

In this section, the simulation results and comparative analysis of proposed and existing coverage hole detection approaches with respect to three different simulation scenarios, such as impact of node density, impact of hole density and impact of simulation rounds.

5.2.1 Impact of node density

In this scenario, we vary the number of nodes as 600, 700, 800, 900 and 1000 with the fixed network size as 500 × 500 m2 area. Table 3 describes the comparative analysis of proposed and existing coverage hole detection approaches with respect to impact of node density. The average energy consumption of our proposed OC-NP-CHD approach is 65.339%, 56.539%, 41.752% and 11.71% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 2 shows the average energy consumption results of proposed and existing coverage hole detection approaches with respect to impact of node density. The hole detection time of our proposed OC-NP-CHD approach is 32.174%, 24.458%, 14.761% and 2.207% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 3 shows the hole detection time results of proposed and existing coverage hole detection approaches with respect to impact of node density. The coverage of our proposed OC-NP-CHD approach is 9.474%, 7.368%, 5.263% and 45.433% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 4 shows the coverage results of proposed and existing coverage hole detection approaches with respect to impact of node density. The network lifetime of our proposed OC-NP-CHD approach is 19.175%, 16.779%, 14.382% and 11.985% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 5 shows the network lifetime results of proposed and existing coverage hole detection approaches with respect to impact of node density. The control packet overhead of our proposed OC-NP-CHD approach is 1.453%, 1.04%, 0.623% and 0.203% minimized compared to the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively.

Table 3 Results of proposed and existing coverage hole detection approaches with respect to impact of node density
Fig. 2
figure 2

Average energy consumption with node density

Fig. 3
figure 3

Hole detection time with node density

Fig. 4
figure 4

Coverage with node density

Fig. 5
figure 5

Network lifetime with node density

5.2.2 Impact of hole density

In this scenario, we vary the number of holes as 10, 20, 30, 40 and 50 with the fixed number of nodes as 1000 and network size as 500 × 500 m2 area. Table 4 describes the comparative analysis of proposed and existing coverage hole detection approaches with respect to impact of hole density. The average energy consumption of our proposed OC-NP-CHD approach is 65.339%, 56.539%, 41.752% and 11.71% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 6 shows the average energy consumption results of proposed and existing coverage hole detection approaches with respect to impact of hole density. The hole detection time of our proposed OC-NP-CHD approach is 32.174%, 24.458%, 14.761% and 2.207% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 7 shows the hole detection time results of proposed and existing coverage hole detection approaches with respect to impact of hole density. The coverage of our proposed OC-NP-CHD approach is 9.474%, 7.368%, 5.263% and 45.433% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 8 shows the coverage results of proposed and existing coverage hole detection approaches with respect to impact of hole density. The network lifetime of our proposed OC-NP-CHD approach is 19.175%, 16.779%, 14.382% and 11.985% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 9 shows the network lifetime results of proposed and existing coverage hole detection approaches with respect to impact of hole density. The control packet overhead of our proposed OC-NP-CHD approach is 1.453%, 1.04%, 0.623% and 0.203% minimized compared to the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively.

Table 4 Results of proposed and existing coverage hole detection approaches with respect to impact of hole density
Fig. 6
figure 6

Average energy consumption with hole density

Fig. 7
figure 7

Hole detection time with hole density

Fig. 8
figure 8

Coverage with hole density

Fig. 9
figure 9

Network lifetime with hole density

5.2.3 Impact of simulation rounds

In this scenario, we vary the number of holes as 10, 20, 30, 40 and 50 with the fixed number of nodes as 1000 and network size as 500 × 500 m2 area. Table 5 describes the comparative analysis of proposed and existing coverage hole detection approaches with respect to impact of simulation rounds. The average energy consumption of our proposed OC-NP-CHD approach is 65.339%, 56.539%, 41.752% and 11.71% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 10 shows the average energy consumption results of proposed and existing coverage hole detection approaches with respect to the impact of simulation rounds. The hole detection time of our proposed OC-NP-CHD approach is 32.174%, 24.458%, 14.761% and 2.207% lower than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. 

Table 5 Results of proposed and existing coverage hole detection approaches with respect to impact of simulation rounds
Fig. 10
figure 10

Average energy consumption with simulation rounds

Figure 11 shows the hole detection time results of proposed and existing coverage hole detection approaches with respect to impact of simulation rounds. The coverage of our proposed OC-NP-CHD approach is 9.474%, 7.368%, 5.263% and 45.433% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 12 shows the coverage results of proposed and existing coverage hole detection approaches with respect to impact of simulation rounds. The network lifetime of our proposed OC-NP-CHD approach is 19.175%, 16.779%, 14.382% and 11.985% higher than the existing benchmark approaches are CG-DCHD, DHC, PS and DCHD, respectively. Figure 13 shows the network lifetime results of proposed and existing coverage hole detection approaches with respect to impact of simulation rounds. The control packet overhead of our proposed OC-NP-CHD approach is 1.453%, 1.04%, 0.623% and 0.203% minimized compared to the existing benchmark approaches are CG-DCHD, DHC, PS andDCHD, respectively.

Fig. 11
figure 11

Hole detection time with simulation rounds

Fig. 12
figure 12

Coverage with simulation rounds

Fig. 13
figure 13

Network lifetime with simulation rounds

6 Conclusion

The computational geometry-based approaches are failed to detect coverage holes between sensors and boundary of the region. We have proposed an optimal cluster-based node position estimation and coverage hole detection (OC-NP-CHD) approach to solve the problems in the previous CG-DCHD approach. The MLO algorithm is used for computing the node positions in the WSN. Its impact is to ensure that the edge nodes in the network are optimally positioned. This helps in reducing the size of coverage holes in the network and improving the overall coverage of the WSN. The ISCSO algorithm is used for optimal clustering of the sensor nodes in the network. Its impact is to balance the clusters and efficiently compute the coverage hole area in the network. This helps in detecting coverage holes in a more accurate and efficient manner. The Hyb-DRL technique is used for detecting the shape and size of coverage holes within clusters, among clusters and along edges. To validate the performance of proposed CG-DCHD approach by using different simulation scenarios and measures. From the results we observed that the average energy consumption of our CG-DCHD approach is 43.835%, 32.674% and 26.164% lower compared to the existing benchmark approaches for node density, hole density and simulation rounds, respectively. The hole detection time of our CG-DCHD approach is 18.4%, 16.802% and 15.462% lower compared to the existing benchmark approaches for node density, hole density and simulation rounds, respectively. The coverage of our CG-DCHD approach is 16.885%, 14.977% and 12.219 higher compared to the existing benchmark approaches for node density, hole density and simulation rounds, respectively. The network lifetime of our CG-DCHD approach is 15.58%, 17.702% and 20.492% higher compared to the existing benchmark approaches for node density, hole density and simulation rounds, respectively. The control packet overhead of our CG-DCHD approach is 0.83%, 1.907% and 1.466% lower compared to the existing benchmark approaches for node density, hole density and simulation rounds, respectively. Energy consumption is a critical concern in WSNs. Future research could explore ways to optimize energy efficiency in the OC-NP-CHD approach. This may involve developing energy-aware algorithms, adaptive power management techniques or energy harvesting strategies to prolong the network's lifetime and reduce energy consumption.