Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Genomic structural variation, for example insertion and deletion of DNA, are common in the human population and have been linked to various diseases and conditions. The basic question scientists and clinicians want to answer is: given a DNA sample from a donor and a suitable reference genome, what structural variants does the donor have in comparison to the reference? Methods for identifying structural variants are continuously worked on, in terms of both experimental protocols and bioinformatic analysis. Short-read technologies are, despite their weaknesses, the primary data source because of the superior throughput/cost ratio. It is today important to improve accuracy of predictions and in particular to reduce the false-positive rate while retaining sensitivity. To that end, we have worked on improving the statistical analysis of paired reads, using paired-end (PE) or mate-pair (MP) libraries, for evaluating the significance of a detected insertion or deletion.

While aligned reads are important for identifying short variants and substitutions, larger variants and variants in repetitive regions where alignment is difficult are easier detected by paired reads spanning over the region. In PE and MP protocols, reads are from the ends of DNA fragments from the donor. PE libraries have short-range fragment lengths (up to 100 s bp), MP libraries are long range (1000 s bp), and they each have their own strengths and limitations. PE libraries often has superior coverage and narrow fragment length distribution while long range MP libraries can span larger insertions and, at similar read coverage, provide higher span coverage (the number of MP pairs separated by a random position) than PE libraries, which in theory can make up for the increased variation in individual fragment lengths by increasing statistical power from more observations.

Numerous structural variation algorithms using read pairs, and their fragment length, to detect variants have been proposed. Many tools use only discordant read pairs for downstream calling of variants, i.e., read pairs that align at a distance smaller than \(\mu - k\sigma \) or larger than \(\mu + k\sigma \) base pairs from each other, where \(\mu \) and \(\sigma \) are the mean and standard deviation of the fragment length distribution and \(k\in \mathbb {R}\) [2, 5, 11, 12, 20]. This restriction may reduce the computational demand, but it sacrifices sensitivity [17] by removing observations.

There are also tools with a statistical model/approach that utilizes all read pairs. CLEVER [17] finds insertions and deletions based on statistically significant deviation of the mean fragment length of all readsFootnote 1 over a position from \(\mu \). This method finds more and smaller variants compared to methods that use only discordant reads [17]. [9] models the number of discordant and concordant read pairs (classified by a cutoff) over a region as following a binomial distribution and finds inversions and deletions based on statistically significant accumulation of discordant read pairs over regions. However, any binary classification cutoff causes loss of information [8], thus statistical power, as they do not consider how much above or below the cutoff a fragment length isFootnote 2. Another approach is non-parametric testing of the distribution over a region, e.g., using the Kolmogorov-Smirnov test [15], but as [17] noted, this is computationally expensive. [10] presented a model to find the most likely common deletion length from several donor genomes with different fragment length distributions by maximizing the likelihood of observed fragment lengths given a deletion size and each of the distributions.

These methods however assumes that the probability of a fragment length being observed over a position/region follows the probability distribution of the full library fragment length distribution, which is not true [22]. Longer fragment lengths span more positions than shorter fragment lengths, so over any position in the genome there will be a bias towards read-pairs further apart than \(\mu \). This observation bias of fragment sizes has been investigated earlier in an assembly context, estimating the gap size between contigs [18, 21, 22]. The approaches given in [18, 21] are more general by using the exact (empirical) distribution over the fragment length, which also makes them computationally demanding. GapEst [22] assumes a normal fragment length distribution and derives an analytic expression for the likelihood of a gap size that scales very well, which opens up for other applications where this type of problem needs to be calculated for a large number of instances, e.g., structural variation detection. There is no previous work known to the authors on incorporating this model, or a similar one, to structural variation and investigating how it affects the balance between detecting deletions and insertions.

1.1 Contribution

We use the statistical model given in [22] and present it in the context of structural variation detection. The model provides a probability distribution for the fragment sizes we observe over a position (e.g., a potential breakpoint) or region. Given this distribution we derive a null-hypothesis distribution to detect variants. We show that the corrected null-hypothesis agrees with both simulated and biological data, while a commonly used null-hypothesis does not. We implement the null-hypothesis in the state-of-the-art fragment-length based variant caller CLEVER [17]. Although CLEVER uses constraints and assumptions that do not agree with our model, we show that the detection of insertions and deletions becomes more balanced and that the number of false positive calls decreases. This is a promising first result as we could only apply a part of our theory in CLEVER without a significant restructuring of the code. We also believe that this work is a step towards creating a statistical rigorous approach for read pair fragment lengths where we can detect indels to a much higher resolution than cutoff based ones.

In some places, we have to refer to an Appendix, which can be found in a bioRxiv (http://dx.doi.org/10.1101/036707) version of this paper.

2 Methods

We will review a model used to determine contig distances in scaffolding [22] and use it in the context of structural variation detection. Notation and assumptions are presented in Sect. 2.1. In Sect. 2.2 we present the probability distribution in a structural variation detection context. Section 2.4 discuss a commonly used null-hypothesis used for detecting variants with fragment length and derives an improved null-hypothesis using our model.

2.1 Notation and Assumption

We refer to our model as the Observed Fragment Length (OFL) model. This model carries no new concepts and makes the same assumptions as the Lander-Waterman model [14], but adds a variable and some constants. We only state it here for convenience of referencing to a model when we derive probabilities and a null-hypothesis. Read pairs are sampled independently and uniformly from the donor genome. Let G denote the length of the reference genome. Alignment of read pairs to the reference genome yields our observations: distance o between reads in a read pair, read length r, and number of allowed “inner”Footnote 3 softclipped bases s [16] in an alignment, see Fig. 1a. Read-pair distances x come from a library fragment length distribution f(x) (either given or estimated from alignments). We denote the mean and standard deviation of this distribution as \(\mu \) and \(\sigma \). Finally, a parameter \(\delta \) models the number of missing or added base pairs in the reference, compared to the donor sequence. That is, if the donor sequence contains an insertion, \(\delta \) is negative and we say that the donor sequence has \(\delta \) added bases. Similarly, if the donor sequence contains a deletion, \(\delta \) is positive and we say that the donor sequence has \(\delta \) deleted bases. For a given read-pair with fragment length x, let \(w_{G,p}(x)\) denote the probability that it spans over position p on a genome of size G. As we do not model that any two positions have different probabilities to be spanned over (reads are drawn uniformly), w will not depend on p and we omit it and refer to \(w_{G}(x)\) from now on.

Fig. 1.
figure 1

(a) Constants and variables in the OFL model. The figure illustrates the scenario of an insertion in the donor genome of length \(\delta \) at position p. Two reads (marked by arrows) of length r are at distance o from each other, with the left read partially aligned leaving s positions unaligned (softclipped). (b) Illustration of a full fragment size distribution f(x) from N(4000, 2000) (blue line), from which \(H_0\) is derived. The green dashed line shows the observed fragment distribution over any variant free position for fragments coming from f(x) for the simplified case when \(r,s,\delta =0\) (i.e., this is exactly the function xf(x)). Red lines indicate the \(\mu \pm 2\sigma \) quantiles of f(x). It is less likely to observe a smaller fragment size over any given position in the genome (see density of green distribution at red lines), as opposed to identical significance under f(x).

2.2 Probability Function over Observed Fragment Lengths

The distribution and probabilities derived in this section closely matches those given in [22] with the minor addition of the constant s. We restate the expressions in a structural variation context for clarity.

No Variant. First, we assume that donor and reference are identical, therefore \(\delta = 0\) at any position. Given the OFL model, the probability that we observe a read pair with fragment size x over a position p on a genome of length G is

$$\begin{aligned} {\begin{matrix} P(x | \delta = 0) = \frac{w_{G}(x)f(x) }{\sum _{x}w_{G}(x)f(x)} = \frac{ \frac{(x-2(r-s))}{G}f(x)}{\sum _{x} \frac{(x-2(r-s))}{G}f(x)}. \end{matrix}} \end{aligned}$$
(1)

Here f(x) is the probability to draw a fragment of length x from the full library and \(w_G(x)\) the probability that it spans over position p. The denominator is a normalization constant to make P a probability. It is assumed that \(x \ge 2(r-s)\). For example, if the read length is 100 and maximum allowed softclipped bases of an aligned read is 30 a read pair with fragment length 300 will have \(300 - 2(100 - 70) = 160\) possible placements where it spans position p. For simplicity, we omit the special case when p is near the end of a chromosome.

Modeling Variant at a Position. Let \(\delta \) be the unknown variant size. In this case we cannot observe the true fragment length x of read pairs. What we observe is instead \(o = x-\delta \) (see Fig. 1a). A modification of w(x) is needed as fragment sizes is now required to span \(\delta \) base pairs and have sufficiently many base pairs on each side to be mapped (\(2(r-s)\)). We have

$$\begin{aligned} w(x,\delta ) = \frac{1}{G} \max \{x- \delta - 2(r-s)+1,0\}. \end{aligned}$$

The 0 in the max function keeps the function weight to 0 in case we have no possible placing of a paired read over a variant. We can simplify this function to be expressed in o, as \(o= x-\delta \), and write \(w(o) = G^{-1}\max \{o-2(r-s)+1,0\}\).

We see that the function w is constant for any given observation and can therefore be interpreted as a “weight” function, hence the notation w.

2.3 Probability of Variant Size \(\delta \)

We can express the probability of \(\delta \) given observations as \(P(\delta | o)\). Lacking prior information about \(\delta \), we model it with the uniform distributionFootnote 4. Using Bayes theorem, we get

$$\begin{aligned} {\begin{matrix} P(\delta | o) = \frac{P(o | \delta ) P(\delta )}{P(o)} \propto P(o | \delta ) P(\delta ) \propto P(o | \delta ) \end{matrix}} \end{aligned}$$

where P(o) and \(P(\delta )\) are constant by the assumption of a uniform distribution. We now have

$$\begin{aligned} {\begin{matrix} P(o|\delta ) = \frac{w(o|\delta )f(\delta +o | \delta ) }{\sum _{t}w(t-\delta )f(t)} \end{matrix}} \end{aligned}$$
(2)

where the denominator is the sum of all possible fragment sizes that can be observed given \(\delta \) and f. We can now find the most likely \(\delta \) using maximum likelihood estimation (MLE) over (2). The time complexity for the MLE is \(O(n + \log t)\) Footnote 5 if \(f\sim N\) (with t continuous), where n is the number of observations [22]. Note that we implicitly get \(P(x|\delta )\) since \(P(x|\delta ) = P(o+\delta |\delta ) = P(o|\delta )\).

2.4 Null-Hypothesis and Statistical Testing

Let \(Y \doteq O | \delta =0\), that is, the random variable over observed fragment lengths given \(\delta =0\). Let \(\bar{y} \doteq \frac{\sum _{i=1}^n y_i}{n}\) be the sample mean of observed fragment lengths. Considering \(\bar{y}\) a random variable over experiments, it is commonly assumed that \(\bar{y} \sim N(\mu , \sigma /\sqrt{n})\), i.e., the distribution of the sample mean of f(x) under the central limit theorem (CLT), and this distribution is used as null-hypothesis [5, 17]. We call this null-hypothesis \(H_0\). Furthermore, the variant size \(\delta \) is estimated from observed fragment lengths o as \(\hat{\delta } = \mu - \bar{o}\) [5, 9, 10, 12, 17, 20]. At first glance this formula seems reasonable since we take the expected fragment size and subtract the mean of the observations, but it has strong limitations. One is that \(\hat{\delta }\) in this case has an upper bound of \(\mu -2(r-s)\) since \(o\ge 2(r-s)\). This equation implies that we can never span over a sequence longer than \(\mu -2(r-s)\). We use Eq. 2 to derive the correct mean and standard deviation of Y given the OFL model, denoted \(\mu _{p}\) and \(\sigma _p\) respectively. The derivation of \(\mu _{p}\) is similar to derivation of observed fragment size linking two contigs given in [22], and the derivation of \(\sigma _{p}\) is a special case of the derivation of the variance of observed fragment size linking two contigs given in [23]. See proof in AppendixFootnote 6.

Theorem 1

Given the OFL model, \(f \sim N(\mu ,\sigma )\), and \(\delta = 0\), we have \(\mu _{p} \approx \mu + \frac{\sigma ^2}{\mu - \left( 2(r-s)+1\right) }\) and \(\sigma _{p}\approx \sigma \sqrt{1 - \sigma ^2\left( \mu - (2(r-s)+1)\right) ^{-2}}\).

The null-hypothesis is that there is no variant, thus \(\delta = 0\). Under CLT, as n increases, we therefore have \(\bar{y} \sim N(\mu _p, \sigma _p/\sqrt{n})\). Notice that we can calculate \(\mu _p\) and \(\sigma _p\) without the assumption \(f \sim N(\mu ,\sigma )\) by using an empirical estimate of f(x) from aligned read pairs. Nevertheless, the closed expression formulas in Theorem 1 illustrates a basic feature of the model — larger variance increases the discrepancy between \(\mu \) and \(\mu _p\). It is also robust to non-normality, as we will see in Sect. 3.2. In case we have enough observations to motivate the Z-test, we perform a simple Z-test and obtain a p-value based on a two sided test (both deletions and insertions are tested for) using the z-statistic

$$\begin{aligned} {\begin{matrix} z=\frac{\bar{y}-\mu _p}{\sigma _p/\sqrt{n}}. \end{matrix}} \end{aligned}$$
(3)

We refer to the null-hypothesis test using (3) as \(H'_0\). Thus, we have derived a different distribution under the null-hypothesis which we advocate should be used instead of \(H_0\). In case we have few observations (more often over insertions), approximation with the Z-test is poor. To get an exact test we would need to derive the distribution of \(\sum _{i=1}^n Y_i\), for n observations \(y_i\) \(i\in [1,n]\). This could improve power to detect insertions, but we refrain from studying this in the present paper.

3 Results

We discuss why modeling bias contributes to making deletion calls more frequent than insertions calls in Sect. 3.1. In Sect. 3.2 we show that our corrected hypothesis agrees with biological data, and in Sect. 3.3 that how indel detection is affected in CLEVER when our null-hypothesis is inserted.

3.1 Bias Between Detection of Deletions and Insertions

As donor fragments need to span over insertions (\(\delta > 0\)), and this probability is \(w(x,\delta ) = \frac{1}{G} \max \{x- \delta - 2(r-s)+1,0\}\) according to the OFL model, it is less likely that such fragments will be observed, as \(\delta \) grows. We will therefore have a lower sample size over insertions in general. This naturally gives less power to detect an insertion compared to a deletion of similar size. However, methods using \(H_0\) have less power than necessary. Firstly, as \(\mu _p > \mu \), this gives too many significant upper quantile p-values (deletions) and too few significant lower quantile p-values (insertions). The difference in significance of observing a fragment of size \(\mu + 2\sigma \) compared to observing a fragment of size \(\mu - 2\sigma \) under \(H_0\), compared to the when observed under \(H'_0\) is seen in Fig. 1b. Secondly, the positive skew of the OFL distribution (Fig. 1b) makes a Z-test approximation less powerful compared to an exact test, especially for small sample sizes — as is more likely for insertions.

Table 1. Library information. Reads were aligned with BWA-MEM [16] version 0.7.12 with default parameters. Physical coverage is c, for all reads, and c (pp), for restricted proper pairs, i.e., read pairs that have both mates mapped in correct orientation and within a distance that depends on a statistical filtering of outliers based on the library distribution. The filtering bounds were roughly 10000, 6000, and 14000 bp for rhodo, plasm, and hs13 respectively. \(\mu \) and \(\sigma \) are the mean and standard deviation of the full fragment length distribution. True mean insert-size and standard deviation over a position on the genome, \(\mu _p\) and \(\sigma _p\) (calculated as the average over all positions in the genome) and predictions with closed formula, \(\hat{\mu }_p^c\) and \(\hat{\sigma }_p^c\), and exact calculation, \(\hat{\mu }_p^e\) and \(\hat{\sigma }_p^e\).
Fig. 2.
figure 2

(a) The fragment length distribution f(x) for the hs13 dataset and the red line is a best fit of a truncated normal distribution. f(x) deviates significantly from a normal distribution. Although the mean of f(x) is 2947, the average observed fragment length over position p (\(\mu _p\)) over all positions on hs13 shows that most values occur between 3000–4500 bp with the average around 3688 bp (d) — as approximately predicted from \(\hat{\mu }_p^e\) and \(\hat{\mu }_p^c\). Figures (b) and (c) shows the p-value distribution and CDF values from using \(H_0\) (i.e., using \(\mu \) and \(\sigma \)). Figures (e) and (f) shows the p-value distribution and CDF values from using \(H'_0\) (i.e., using \(\hat{\mu }_p^e\) and \(\hat{\sigma }_p^e\)).

3.2 Evaluating the Accuracy of \(H'_0\)

We evaluated the accuracy of our null-hypothesis on three mate pair libraries. We used a mate pair library from Rhodobacter sphaeroides from [21] denoted rhodo, a mate pair library from Plasmodium falciparum used in [13] denoted plasm, and mate-pair data from a human individual in the CEPH 1463 family-trioFootnote 7. For the human dataset we aligned the reads to the complete human genome, but limited analysis to chromosome 13. We call this dataset hs13. Table 1 shows information about the datasets. Recall from Sect. 2.4 that \(\mu _{p}\) and \(\sigma _{p}\) are the true mean and standard deviation of fragment lengths over a position that does not contain a variant. Let \(\hat{\mu }_{p}^c\) and \(\hat{\sigma }_{p}^c\) refer to the estimated quantity of \(\mu _{p}\) and \(\sigma _{p}\) from the closed formulas in Theorem 1. Similarly, let \(\hat{\mu }_{p}^e\) and \(\hat{\sigma }_{p}^e\) be the estimates of \(\mu _{p}\) and \(\sigma _{p}\) by using an empirical distribution of f(x) (estimated from a sample) and summing up the probabilities in Eq. 2 with \(\delta =0\). Estimates and observed values are shown in Table 1. It is our assumption that an overwhelming majority of positions are variant freeFootnote 8. Thus, we expect a model that fits data should give a uniformly distributed p-value distribution. Our observations are summarized below.

Predicting \(\mathbf {\mu _p}\) . \(\mu _p\) is estimated very well by both \(\hat{\mu }_{p}^e\) and \(\hat{\mu }_{p}^c\); compare Fig. 2d for hs13, and Appendix Figs. 3b and d for rhodo and plasm respectively with the estimated values in Table 1. Hence, testing \(\bar{o} = \hat{\mu }_p^e\) (or \(\hat{\mu }_p^c\)) as in \(H'_0\) introduce symmetrical cumulative distribution function (CDF) values, Fig. 2f, compared to CDF based on testing \(\bar{o} = \hat{\mu }\) where all values are distributed around 1.0 — suggesting significant deletions, see Fig. 2c.

Predicting \(\mathbf {\sigma _p}\) . The closed formula predictions of \(\sigma _{p}\) works best if f(x) is normal (plasm). For rhodo and hs13, \(\hat{\sigma }_{p}^e\) and \(\hat{\sigma }_{p}^c\) differs significantly and \(\hat{\sigma }_{p}^e\) should be used, compare \(\sigma _p\) with \(\hat{\sigma }_{p}^e\) and \(\hat{\sigma }_{p}^c\) in Table 1.

\(\mathbf {p}\) -values. The p-value distribution (ideally uniform) greatly improves with \(H'_0\) (Fig. 2e) compared to p-values obtained with \(H_0\) (Fig. 2b). Abnormalities in the p-values are most likely explained by: alignment artifacts (some regions are more difficult aligning to), fragment length bias [1, 19], coverage bias from GC-content, and in some cases, real variants, see evaluation of hs13 dataset in Sect. 5.7 of the Appendix. Similar p-value distributions are obtained on rhodo and plasm genome (data not shown) — that should not contain any variants — indicating that most of the enrichment of low p-values on hs13 is explained by any of the former three causes.

Table 2. Insertions and deletions called with CLEVER using \(H_0\) and \(H'_0\). Column \(\delta \) contains the size of 50 insertions and deletions, simulated on the reference genomes by either deleting or inserting a \(\delta \) bp sequence on the reference. A “0” indicates that the original biological dataset was used.

3.3 Implementing the Corrected Null-Hypothesis in CLEVER

In this section we illustrate as a proof-of-concept how the corrected hypothesis \(H'_0\) (with \(\hat{\mu }_p^e\) and \(\hat{\sigma }_p^e\)) balances the ratio between detected insertions and deletions. We applied \(H_0'\) in CLEVER (v 1.1). However, we want to emphasize that we did not tailor the statistical tests as needed to fit the assumptions made by their particular method. This limits the performance improvement. To further improve results with CLEVER, we would need to (1) implement exact tests for few observations — giving more power to detect insertions, (2) use the OFL-model for CLEVER’s discovery of positions to study, (3) based on our model adjust CLEVER’s methods to handle, e.g., heterozygous variants and controlling the false discovery rate. This would require additional modeling and significant restructuring of the code and we do not consider it here. Our aim here is only to illustrate how the simple adjustment of inserting \(H'_0\) instead of \(H_0\) in CLEVER has a significant impact on the output. We investigated how the replacement of \(H'_0\) instead of \(H_0\) changed variant calls from CLEVER on hs13, rhodo and plasm as well with ideal condition simulated data denoted sim-\(N(\cdot ,\cdot )\) (full simulated results in Appendix Sect. 5.6). For simulated variants, similarly to [17], a prediction is classified as a true positive (TP) if the breakpoint prediction is not further than one mean insert size (i.e., at most \(\mu - 2r\)) away from the true breakpoint. Otherwise it is classified as a false positive (FP). All variant calls on rhodo and plasm that are not from simulated variants are assumed to be false positives.

Because hs13 likely harbors true variants, we used annotated variants from dbVar [7], together with manual inspection in BamView [3], to assess if hits are true or false positives. For a deletion call in CLEVER with start and end coordinates \(p_s,p_e\) and a deletion in dbVar with coordinates \(q_s,q_e\), we let \(max\_del = \max (p_e-p_s, q_e-q_s)\) and \(overlap =\min \{0, \min (p_e,q_e) - \max (p_s,q_s)\}\). We let \(hit\_value =overlap/max\_del\) and a call is a hit if \(hit\_value > T\), where \(0<T<1\) is a threshold. Because dbVar contains a large amount of annotated variants from several individuals and CLEVER produces many calls under \(H_0\), roughly 173, 106 and 40 hits are expected by chance with \(T=0.25, 0.5, 0.75\) (estimation in Appendix Sect. 5.3), which is similar numbers to the observed hits from CLEVER: 226, 109 and 31 respectively under \(T=0.25, 0.5, 0.75\). We therefore further manually evaluated the hits produced with \(T=0.75\) by looking for coverage drop and accumulation of softclips near each breakpoint. This gave us Estimated True Positives (ETP) as a rough measure of the TP rate for hs13. Therefore, we report ETP and Total Calls (TC) for hs13 in Table 2, contrary to simple TP and FP for the other data sets where we have the ground truth.

Improvements. From Table 2 and Fig. 5 (Appendix) we see that CLEVER with \(H_0\) detects significantly more deletions than insertions of the same sizes. Using \(H'_0\), reduces this bias to some extent by increasing the detection of insertions across all data sets. CLEVER also returns significantly fewer false positive deletion calls with \(H'_0\), see rhodo Table 2 and sim-\(N(300,\cdot )\). Even though \(H_0\) have more sensitivity in calling deletions on hs13, the signal disappears in the overwhelming amount of total calls, compare ETP and TC for \(H_0\) and \(H'_0\) in Table 2.

Deterioration. A consequence of using \(H'_0\) is fewer deletion calls, which unfortunately also removes some true positive deletion calls (Table 2 and Appendix Fig. 5). It also increases the FP insertion calls on plasm (Table 2). We believe that calling variants with the plasm library carries additional difficulties due to its GC-poor genome sequence, such as positional fragment length bias [1, 19].

Additional evidence that most calls with \(H_0\) on hs13 are FPs are found by comparing statistics on CLEVER’s deletion calls (Appendix Fig. 6) and numbers reported in recent extensive studies [4, 6]. For example, [4] provide frequency distributions for both previously discovered and new deletions on single genomes. Roughly 250 deletions have lengths over 1000 bp (inspection of plot). The simplifying assumption that large-indel distribution is uniform over chromosomes gives around 8 expected deletionsFootnote 9 in size range 1000 bp. This approximate number, and the fact that almost all calls were removed when using \(H'_0\) corroborates, that the vast majority (in the order of \(>99\,\%\)) of calls with \(H_0\) are FPs — likely a consequence of using \(H_0 : \mu _p = 2410\) compared to the true value \(\mu _p = 3719\).

4 Conclusions

We stated a probability distribution of observed fragment length over a position or region and derived a new null-hypothesis for detecting variants with fragment length, which is sound and agrees with biological data. Applied in CLEVER, our null-hypothesis detects more insertions and reduces false positive deletion calls. Results could be further improved by deriving an exact distribution instead of a Z-test and updating CLEVER’s edge-creating conditions to agree with our model. The presented model, distribution, and null-hypothesis are general and could be used together with other information sources such as split reads, softclipped alignments, and read-depth information.