Abstract
Inspired by the classic Sauvola local image thresholding approach, we systematically study it from the deep neural network (DNN) perspective and propose a new solution called SauvolaNet for degraded document binarization (DDB). It is composed of three explainable modules, namely, Multi-Window Sauvola (MWS), Pixelwise Window Attention (PWA), and Adaptive Sauolva Threshold (AST). The MWS module honestly reflects the classic Sauvola but with trainable parameters and multi-window settings. The PWA module estimates the preferred window sizes for each pixel location. The AST module further consolidates the outputs from MWS and PWA and predicts the final adaptive threshold for each pixel location. As a result, SauvolaNet becomes end-to-end trainable and significantly reduces the number of required network parameters to 40K – it is only 1% of MobileNetV2. In the meantime, it achieves the State-of-The-Art (SoTA) performance for the DDB task – SauvolaNet is at least comparable to, if not better than, SoTA binarization solutions in our extensive studies on the 13 public document binarization datasets. Our source code is available at https://github.com/Leedeng/SauvolaNet.
This work was done prior to Amazon involvement of the authors.
This work was funded by The Science and Technology Development Fund, Macau SAR (File no. 189/2017/A3), and by University of Macau (File no. MYRG2018-00136-FST).
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Document binarization typically refers to the process of taking a gray-scale image and converting it to black-and-white. Formally, it seeks a decision function \(f_{\text {binarize}}(\cdot )\) for a document image \(\mathbf {D} \) of width W and height H, such that the resulting image \(\hat{\mathbf {B}}\) of the same size only contains binary values while the overall document readability is at least maintained if not enhanced.
Document binarization plays a crucial role in many document analysis and recognition tasks. It is the prerequisite for many low-level tasks like connected component analysis, maximally stable extremal regions, and high-level tasks like text line detection, word spotting, and optical character recognition (OCR).
Instead of directly constructing the decision function \(f_{\text {binarize}}(\cdot )\), classic binarization algorithms [17, 28] typically first construct an auxiliary function \(g(\cdot )\) to estimate the required thresholds \(\mathbf {T}\) as follows.
In global thresholding approaches [17], this threshold \(\mathbf {T}\) is a scalar, i.e. all pixel locations use the same threshold value. In contrast, this threshold \(\mathbf {T}\) is a tensor with different values for different pixel locations in local thresholding approache [28]. Regardless of global or local thresholding, the actual binarization decision function can be written as
where th(x, y) is the simple thresholding function and the binary state for a pixel located at i-th row and j-th column is determined as in Eq. (4).
Classic binarization algorithms are very efficient in general because of using simple heuristics like intensity histogram [17] and local contrast histogram [31]. The speed of classic binarization algorithms typical of the millisecond level, even on a mediocre CPU. However, simple heuristics also means that they are sensitive to potential variations [31] (image noise, illumination, bleed-through, paper materials, etc. ), especially when the relied heuristics fail to hold. In order to improve the binarization robustness, data-driven approaches like [33] learn the decision function \(f_{\text {binarize}}(\cdot )\) from data rather than heuristics. However, these approaches typically achieve better robustness by using much more complicated features, and thus work relatively slow in practice, e.g. on the second level [33].
Like in many computer vision and image processing fields, the deep learning-based approaches outperform the classic approaches by a large margin in degraded document binarization tasks. The state-of-the-art (SoTA) binarization approaches are now all based on deep neural networks (DNN) [22, 27]. Most of SoTA document binarization approaches [2, 19, 32] treat the degraded binarization task as a binary semantic segmentation task (namely, foreground and background classes) or a sequence-to-sequence learning task [1], both of which can effectively learn \(f_{\text {binarize}}(\cdot )\) as a DNN from data.
Recent efforts [2, 5, 9, 19, 30, 32, 34] focus more on improving robustness and generalizability. In particular, the SAE approach [2] suggests estimating the pixel memberships not from a DNN’s raw output but the DNN’s activation map, and thus generalizes well even for out-of-domain samples with a weak activation map. The MRAtt approach [19] further improves the attention mechanism in multi-resolution analysis and enhances the DNN’s robustness to font sizes. DSN [32] apply multi-scale architecture to predict foreground pixel at multi-features levels. The DeepOtsu method [9] learns a DNN that iteratively enhances a degraded image to a uniform image and then binarized via the classic Otsu approach. Finally, generative adversarial networks (GAN) based approaches like cGANs [34] and DD-GAN [5] rely on the adversarial training to improve the model’s robustness against local noises by penalizing those problematic local pixel locations that the discriminator uses in differentiating real and fake results.
As one may notice, both classic and deep binarization approaches have pros and cons: 1) the classic binarization approaches are extremely fast, while the DNN solutions are not; 2) the DNN solutions can be end-to-end trainable, while the classic approaches can not. In this paper, we propose a novel document binarization solution called SauvolaNet – it is an end-to-end trainable DNN solution but analogous to a multi-window Sauvola algorithm. More precisely, we re-implement the Sauvola idea as an algorithmic DNN layer, which helps SauvolaNet attain highly effective feature representations at an extremely low cost – only two Sauvola parameters are needed. We also introduce an attention mechanism to automatically estimate the required Sauvola window sizes for each pixel location and thus could effectively and efficiently estimate the Sauvola threshold. In this way, the SauvolaNet significantly reduces the total number of DNN parameters to 40K, only 1% of the MobileNetV2, while attaining comparable performance of SoTA on public DIBCO datasets. Figure 1 gives the high-level comparisons of the proposed SauvolaNet to the SoTA DNN solutions.
The rest of the paper is organized as follows: Sect. 2 briefly reviews the classic Sauvola method and its variants; Sect. 3 proposes the SauvolaNet solution for degraded document binarization; Sect. 4 presents Sauvola ablation studies results and comparisons to SoTA methods; and we conclude the paper in Sect. 5.
2 Related Sauvola Approaches
The Sauvola binarization algorithm [28] is widely used in main stream image and document processing libraries and systems like OpenCVFootnote 1 and Scikit-ImageFootnote 2. As aforementioned, it constructs the binarization decision function (3) via the auxiliary threshold estimation function \(g_{\text {Sauvola}}\), which has three hyper-parameters, namely, 1) w: the square window size (typically an odd positive integer [4]) for computing local intensity statistics; 2) k: the user estimated level of document degradation; and 3) r: the dynamic range of input image’s intensity variation.
where \(\theta \,=\,\{w, k, r\}\) indicates the used hyper-parameters. Each local threshold is computed w.r.t. the 1st- and 2nd-order intensity statistics as shown in Eq. (6),
where \(\mu [i,j]\) and \(\sigma [i,j]\) respectively indicate the mean and standard deviation of intensity values within the local window as follows.
It is well known that heuristic binarization methods with hyper-parameters could rarely achieve their upper-bound performance unless the method hyper-parameters are individually tuned for each input document image [12], and this is also the main pain point of Sauvola approach.
Many efforts have been made to mitigate this pain point. For example, [14] introduces a multi-grid Sauvola variant that analyzes multiple scales in the recursive way; [13] proposes a hyper-parameter free multi-scale binarization solution called Sauvola MS [2] by combining Sauvola results of a fixed set of window sizes, each with its own empirical k and r values; [8] improves the classic Sauvola by using contrast information obtained from pre-processing to refine Sauvola ’s binarization; [12] estimates the required window size w in Sauvola by using the stroke width transform matrix. Table 1 compares these approaches with the proposed SauvolaNet , and it is clear that only SauvolaNet is end-to-end trainable.
3 The SauvolaNet Solution
Figure 2 describes the proposed SauvolaNet solution. It learns an auxiliary threshold estimation function from data by using a dual-branch design with three main modules, namely Multi-Window Sauvola (MWS), Pixelwise Window Attention (PWA), and Adaptive Sauvola Threshold (AST).
Specifically, the MWS module takes a gray-scale input image \(\mathbf {D}\) and leverages on the Sauvola to compute the local thresholds for different window sizes. The PSA module also takes \(\mathbf {D}\) as the input but estimates the attentive window sizes for each pixel location. The AST module predicts the final threshold for each pixel location \(\mathbf {T}\) by fusing the thresholds of different window sizes from MWS using the attentive weights from PWA. As a result, the proposed SauvolaNet is analogous to a multi-window Sauvola , and models an auxiliary threshold estimation function \(g_{\texttt {SauvolaNet}}\) between the input \(\mathbf {D}\) and the output \(\mathbf {T}\) as follows,
Unlike in the classic Sauvola ’s threshold estimation function (5), SauvolaNet is end-to-end trainable and doesn’t require any hyper-parameter. Similar to Eq. (3), the binarization decision function \(f_{\texttt {SauvolaNet}}\) used in testing as shown below
and the extra thresholding process (i.e. (4)) is denoted as the Pixelwise Thresholding (PT) in Fig. 2. Details about these modules are discussed in later sections.
3.1 Multi-window Sauvola
The MWS module can be considered as a re-implementation of the classic multi-window Sauvola analysis in the DNN context. More precisely, we first introduce a new DNN layer called Sauvola (denoted as \(g_{\text {Sauvola}}(\cdot )\) in the function form), which has the Sauvola window size as input argument and Sauvola ’s hyper-parameter s and r as trainable parameters. To enable multi-window analysis, we use a set of Sauvola layers, each corresponding to one window size in (11).The selection of window are verified in Sect. 4.2.
Figure 3 visualizes all intermediate outputs of SauvolaNet , and Fig. 3-(b*) show predicted Sauvola thresholds based on these window sizes, and Fig. 3-(c*) further binarize the input image using corresponding thresholds. These results again confirm that satisfactory binarization performance can be achieved by Sauvola when the appropriate window size is used.
It is worthy to emphasize that Sauvola threshold computing window-wise mean and the standard deviation (see (6)) is very time-consuming when using the traditional DNN layers (e.g. , AveragePooling2D), especially for a big window size (e.g. , 31 or above). Fortunately, we implement our Sauvola layer by using integral image solution [29] to reduce the computational complexity to O(1).
3.2 Pixelwise Window Attention
As mentioned in many works [12, 13], one disadvantage when using Sauvola algorithm is the tuning of hyper-parameters. Among all three hyperparameters, namely, the window size w, the degradation level k, and the input deviation r, w is the most important. Existing works typically decompose an input image into non-overlapping regions [13, 18] or grids [14] and apply each a different window size. However, existing solutions are not suitable for DNN implementation for two reasons: 1) non-overlapping decomposition is not a differentiable operation; and 2) processing regions/grids of different sizes are hard to parallelize.
Alternatively, we adopt the widely-accepted attention mechanism to remove the dependency on the user-specified window sizes. Specifically, the proposed PWA module is a sub-network that takes an input document image \(\mathbf {D}\) and predicts the pixel-wise attention on all predefined window sizes. It conceptually follows the multi-grid method introduced by DeepLabv3 [3] while using a fixed dilation rate at 2. Also, we use the InstanceNormalization instead of the common BatchNormalization to mitigate the overfitting risk caused by a small training dataset. The detailed network architecture is shown in Fig. 2.
Sample result of PWA can be found in Fig. 3-(e). As one can see, the proposed PWA successfully predicts different window sizes for different pixels. More precisely, it prefers \(w=39\) (see Fig. 3-(b5)) and \(w=15\) (see Fig. 3-(b2)) for background and foreground pixels, respectively; and uses very large window sizes, e.g. , \(w=63\) (i.e. Fig. 3-(b8)) for those pixels on text borders.
3.3 Adaptive Sauvola Threshold
As one can see from Fig. 2, the MWS outputs a Sauvola tensor \(\mathbf {S}\) of size \(H\times W\times N\), where N is the number of used window sizes (and we use \(N=8\), see Eq. (11)), the PWA outputs an attention tensor \(\mathbf {A}\) of the same size as \(\mathbf {S}\) and the attention sum for all window sizes on each pixel location is always 1, namely,
The AST applies the window attention tensor \(\mathbf {A}\) to the window-wise initial Sauvola threshold tensor \(\mathbf {S}\) and compute the pixel-wise threshold \(\mathbf {T}\) as below
Fig. 3-(g) shows the adaptive threshold \(\mathbf {T}\) when using the sample input Fig. 3-(a). By comparing the corresponding binarized result (i.e. Fig. 3-(h)) with those of single window’s results (i.e. Fig. 3-(c*)), one can easily verify that the adaptive threshold \(\mathbf {T}\) outperforms any individual threshold result in \(\mathbf {S}\).
3.4 Training, Inference, and Discussions
In order to train SauvolaNet , we normalize the input \(\mathbf {D}\) to the range of (0, 1) (by dividing 255 for uint8 image), and employ a modified hinge loss, namely
where \(\mathbf {B}\) is the corresponding binarization ground truth map with binary values −1 (foreground) and +1 (background); \(\mathbf {T}\) is SauvolaNet ’s predicted thresholds as shown in Eq. (9); and \(\alpha \) is a parameter to empirically control the margin of decision boundary and only those pixels close to the decision boundary will be used in gradient back-propagation. Throughout the paper, we always use \(\alpha =16\).
We implement SauvolaNet in the TensorFlow framework. The used training patch size is 256\(\times \)256, and the data augmentations are random crop and random flip. The training batch size is set to 32, and we use Adam optimizer with the initial learning rate of 1\(e-\)3. During inference, we use \(f_{\texttt {SauvolaNet}}\) instead of \(g_{\texttt {SauvolaNet}}\) as shown in Fig. 2-(a). It only differs from the training in terms of one extra the thresholding step (4) to compare SauvolaNet predicted thresholds with the original input to obtain the final binarized output.
Unlike in most DNNs, each module in SauvolaNet is explainable: the MWS module leverages the Sauvola algorithm to reduce the number of required network parameters significantly, and the PWA module employs the attention idea to get rid of the Sauvola ’s disadvantage of window size selection, and finally two branches are fused in the AST module to predict the pixel-wise threshold. Sample results in Fig. 3 further confirm that these modules work as expected.
4 Experimental Results
4.1 Dataset and Metrics
In total, 13 document binarization datasets are used in experiments, and they are {(H-)DIBCO 2009 [7] (10), 2010 [23] (10), 2011 [20] (16), 2012 [24] (14), 2013 [25] (16), 2014 [16] (10), 2016 [26] (10), 2017 [27] (20), 2018 [21] (10); PHIDB [15] (15), Bickely-diary dataset [6] (7), Synchromedia Multispectral dataset [10] (31), and Monk Cuper Set [9] (25)}. The braced numbers after each dataset indicates its sample size, and detailed partitions for training and testing will be specified in each study. For evaluation, we adopt the DIBCO metrics [16, 20, 21, 24,25,26,27] namely, F-Measure (FM), psedudo F-Measure (\(F_{ps}\)), Distance Reciprocal Distortion metric (DRD) and Peak Signal-to-Noise Ratio (PSNR).
4.2 Ablation Studies
To simplify discussion, let \(\theta \) be the set of parameter settings related to a studied Sauvola approach f. Unless otherwise noted, we always repeat a study about f and \(\theta \) on all datasets in the leave-one-out manner. More precisely, each score reported in ablation studies is obtained as follows
where \(m(\cdot )\) indicates a binarization metric, e.g. FM; and \(\hat{\mathbf {B}}^x_\theta =f^{\mathbb {X}-x}_{\theta }(\mathbf {D})\) indicates the predicted binarized result for a given image \(\mathbf {D}\) using the solution \(f^{\mathbb {X}-x}_{\theta }\) that is trained on dataset \({\mathbb {X}-x}\) using the setting \(\theta \). More precisely, the inner summation of Eq. (15) represents the average score for the model \(f_{\theta }^{\mathbb {X}-x}\) over all testing samples in x, and that the outer summation of Eq. (15) further aggregates all leave-one-out average scores, and thus leaves the resulting score only dependent on the used method f with setting \(\theta \).
Does Sauvola With Learnable Parameters Work Better? Before discussing SauvolaNet , one must-answer question is whether or not re-implement the classic Sauvola algorithm as an algorithmic DNN layer is the right choice, or equivalently, whether or not Sauvola hyper-parameters learned from data could generalize better in practice. If not, we should leverage on existing heuristic Sauvola parameter settings and use them in SauvolaNet as non-trainable.
To answer the question, we start from one set of Sauvola hyper-parameters, i.e. \(\theta =\{w, k, r\}\), and evaluate the corresponding performance of single window Sauvola , i.e. \(g_{\text {Sauvola} | \theta }\) under four different conditions, namely, 1) non-trainable k and r; 2) non-trainable k but trainable r; 3) trainable k but non-trainable r; and 4) trainable k and r. We further repeat the same experiments for three well-known Sauvola hyper-parameter settings OpenCV (see footnote 1) (w = 11, k = 0.5, r = 0.5), Scikit-Image (see footnote 2) (w = 15, k = 0.2, r = 0.5) and PythresholdFootnote 3 (w = 15, k = 0.35, r = 0.5).
Table 2 summarizes the performance scores for single-window Sauvola with different parameter settings. Each row is about one \(score(\theta , f_{\text {Sauvola}})\), and the three mega rows represent the three initial \(\theta \) settings. As one can see, three prominent trends are: 1) the heuristic Sauvola hyper-parameters (i.e. the non-trainable k and r setting) from the three open-sourced libraries don’t work well for DIBCO-like dataset; 2) allowing trainable k or r leads to better performance, and allowing both trainable gives even better performance; 3) the converged values of trainable k and r are different for different window sizes. We therefore use trainable k and r for each window size in the Sauvola layer (see Sec. 3.1).
Does Multiple-Window Help? Though it seems that having multiple window sizes for Sauvola analysis is beneficial, it is still unclear that 1) how effective it is comparing to a single-window Sauvola , and 2) what window sizes should be used. We, therefore, conduct ablation studies to answer both questions.
More precisely, we first conduct the leave-one-out experiments for the single-window Sauvola algorithms for different window sizes with trainable k and r. The resulting \(score(w, f_\text {Sauvola})\) are presented in the upper-half of Table 3. Comparing to the best heuristic Sauvola performance attained by Scikit-Image in Table 2, these results again confirm that Sauvola with trainable k and r works much better. Furthermore, it is clear that \(f_\text {Sauvola}\) with different window sizes (except for \(w=7\)) attain similar scores, possibly because there is no single dominant font size in the 13 public datasets.
Finally, we conduct the ablation studies of using multiple window sizes in SauvolaNet in the incremental way, and report the resulting \(score(\mathbb {W}, f_\texttt {SauvolaNet})\)s in the lower-half of Table 3. It is now clear that 1) multi-window does help in SauvolaNet ; and 2) the more window sizes, the better performance scores. As a result, we use all eight window sizes in SauvolaNet (see Eq. (11)).
4.3 Comparisons to Classic and SoTA Binarization Approaches
It is worthy to emphasize that different works [2, 9, 11, 13, 19, 34] use different protocols for document binarization evaluation. In this section, we mainly follow the evaluation protocol used in [9], and its dataset partitions are: 1) training: (H)-DIBCO 2009 [7], 2010 [23], 2012 [24]; Bickely-diary dataset [6]; and Synchromedia Multispectral dataset [10], and for testing: (H)-DIBCO 2011 [20], 2014 [16], and 2016 [26]. We train all approaches using the same evaluation protocol for fairly comparison. As a result, we focus on those recent and open-sourced DNN based methods, and they are SAE [2], DeepOtsu [9], cGANs [34] and MRAtt [19]. In addition, heuristic document binarization approaches Otsu [17], Sauvola [28] and Howe [11] are also included. Finally, Sauvola MS [13], a classic multi-window Sauvola solution is evaluated to better gauge the performance improvement from the heuristic multi-window analysis to the proposed learnable analysis.
Table 4, 5 and 6 reports the average performance scores of the four evaluation metrics for all images in each testing dataset. When comparing the three Sauvola based approaches, namely, Sauvola, Sauvola MS, and SauvolaNet , one may easily notice that the heuristic multi-window solution Sauvola MS does not necessarily outperform the classic Sauvola. However, the SauvolaNet , again a multi-window solution but with all trainable weights, clearly beat both by large margins for all four evaluation metrics. Moreover, the proposed SauvolaNet solution outperforms the rest of the classic and SoTA DNN approaches in DIBCO 2011. And SauvolaNet is comparable to the SoTA solutions in H-DIBCO 2014 and DIBCO 2016. Sample results are shown in Fig. 4. More importantly, the SauvolaNet is super lightweight and only contains 40K parameters. It is much smaller and runs much faster than other DNN solutions as shown in Fig. 1.
5 Conclusion
In this paper, we systematically studied the classic Sauvola document binarization algorithm from the deep learning perspective and proposed a multi-window Sauvola solution called SauvolaNet . Our ablation studies showed that the Sauvola algorithm with learnable parameters from data significantly outperforms various heuristic parameter settings (see Table 2). Furthermore, we proposed the SauvolaNet solution, a Sauvola -based DNN with all trainable parameters. The experimental result confirmed that this end-to-end solution attains consistently better binarization performance than non-trainable ones, and that the multi-window Sauvola idea works even better in the DNN context with the help of attention (see Table 3). Finally, we compared the proposed SauvolaNet with the SoTA methods on three public document binarization datasets. The result showed that SauvolaNet has achieved or surpassed the SoTA performance while using a significantly fewer number of parameters (1% of MobileNetV2) and running at least 5x faster than SoTA DNN-based approaches.
References
Afzal, M.Z., Pastor-Pellicer, J., Shafait, F., Breuel, T.M., Dengel, A., Liwicki, M.: Document image binarization using LSTM: a sequence learning approach. In: International Workshop on Historical document Imaging and Processing, pp. 79–84 (2015)
Calvo-Zaragoza, J., Gallego, A.J.: A selectional auto-encoder approach for document image binarization. Pattern Recognit. 86, 37–47 (2019)
Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587 (2017)
Cheriet, M., Moghaddam, R.F., Hedjam, R.: A learning framework for the optimization and automation of document binarization methods. Comput. Vis. Image Underst. 117(3), 269–280 (2013)
De, R., Chakraborty, A., Sarkar, R.: Document image binarization using dual discriminator generative adversarial networks. IEEE Signal Process. Lett. 27, 1090–1094 (2020)
Deng, F., Wu, Z., Lu, Z., Brown, M.S.: BinarizationShop: a user-assisted software suite for converting old documents to black-and-white. In: Proceedings of the 10th Annual Joint Conference on Digital Libraries, pp. 255–258 (2010)
Gatos, B., Ntirogiannis, K., Pratikakis, I.: ICDAR 2009 document image binarization contest (DIBCO 2009). In: International Conference on Document Analysis and Recognition, pp. 1375–1382. IEEE (2009)
Hadjadj, Z., Meziane, A., Cherfa, Y., Cheriet, M., Setitra, I.: ISauvola: improved Sauvola’s algorithm for document image binarization. In: Campilho, A., Karray, F. (eds.) ICIAR 2016. LNCS, vol. 9730, pp. 737–745. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-41501-7_82
He, S., Schomaker, L.: DeepOtsu: document enhancement and binarization using iterative deep learning. Pattern Recognit. 91, 379–390 (2019)
Hedjam, R., Nafchi, H.Z., Moghaddam, R.F., Kalacska, M., Cheriet, M.: ICDAR 2015 contest on multispectral text extraction (MS-TEx 2015). In: International Conference on Document Analysis and Recognition, pp. 1181–1185. IEEE (2015)
Howe, N.R.: Document binarization with automatic parameter tuning. Int. J. Doc. Anal. Recognit. 16(3), 247–258 (2013)
Kaur, A., Rani, U., Josan, G.S.: Modified Sauvola binarization for degraded document images. Eng. Appl. Artif. Intell. 92, 103672 (2020)
Lazzara, G., Géraud, T.: Efficient multiscale Sauvola’s binarization. Int. J. Doc. Anal. Recognit. 17(2), 105–123 (2014)
Moghaddam, R.F., Cheriet, M.: A multi-scale framework for adaptive binarization of degraded document images. Pattern Recognit. 43(6), 2186–2198 (2010)
Nafchi, H.Z., Ayatollahi, S.M., Moghaddam, R.F., Cheriet, M.: An efficient ground truthing tool for binarization of historical manuscripts. In: International Conference on Document Analysis and Recognition, pp. 807–811. IEEE (2013)
Ntirogiannis, K., Gatos, B., Pratikakis, I.: ICFHR2014 competition on handwritten document image binarization (H-DIBCO 2014). In: International Conference on Frontiers in Handwriting Recognition, pp. 809–813. IEEE (2014)
Otsu, N.: A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 9(1), 62–66 (1979)
Pai, Y.T., Chang, Y.F., Ruan, S.J.: Adaptive thresholding algorithm: efficient computation technique based on intelligent block detection for degraded document images. Pattern Recognit. 43(9), 3177–3187 (2010)
Peng, X., Wang, C., Cao, H.: Document binarization via multi-resolutional attention model with DRD loss. In: International Conference on Document Analysis and Recognition, pp. 45–50. IEEE (2019)
Pratikakis, I., Gatos, B., Ntirogiannis, K.: ICDAR 2011 document image binarization contest (DIBCO 2011). In: International Conference on Document Analysis and Recognition, pp. 1506–1510 (2011)
Pratikakis, I., Zagori, K., Kaddas, P., Gatos, B.: ICFHR 2018 competition on handwritten document image binarization (H-DIBCO 2018). In: International Conference on Frontiers in Handwriting Recognition, pp. 489–493 (2018)
Pratikakis, I., Zagoris, K., Karagiannis, X., Tsochatzidis, L., Mondal, T., Marthot-Santaniello, I.: ICDAR 2019 competition on document image binarization (DIBCO 2019). In: International Conference on Document Analysis and Recognition, pp. 1547–1556 (2019)
Pratikakis, I., Gatos, B., Ntirogiannis, K.: H-DIBCO 2010-handwritten document image binarization competition. In: International Conference on Frontiers in Handwriting Recognition, pp. 727–732. IEEE (2010)
Pratikakis, I., Gatos, B., Ntirogiannis, K.: ICFHR 2012 competition on handwritten document image binarization (H-DIBCO 2012). In: International Conference on Frontiers in Handwriting Recognition, pp. 817–822. IEEE (2012)
Pratikakis, I., Gatos, B., Ntirogiannis, K.: ICDAR 2013 document image binarization contest (DIBCO 2013). In: International Conference on Document Analysis and Recognition, pp. 1471–1476. IEEE (2013)
Pratikakis, I., Zagoris, K., Barlas, G., Gatos, B.: ICFHR2016 handwritten document image binarization contest (H-DIBCO 2016). In: International Conference on Frontiers in Handwriting Recognition, pp. 619–623. IEEE (2016)
Pratikakis, I., Zagoris, K., Barlas, G., Gatos, B.: ICDAR2017 competition on document image binarization (DIBCO 2017). In: International Conference on Document Analysis and Recognition, vol. 1, pp. 1395–1403. IEEE (2017)
Sauvola, J., Pietikäinen, M.: Adaptive document image binarization. Pattern Recognit. 33(2), 225–236 (2000)
Shafait, F., Keysers, D., Breuel, T.M.: Efficient implementation of local adaptive thresholding techniques using integral images. In: Document recognition and retrieval XV, vol. 6815, p. 681510. International Society for Optics and Photonics (2008)
Souibgui, M.A., Kessentini, Y.: De-gan: A conditional generative adversarial network for document enhancement. IEEE Trans. Pattern Anal. Mach. Intell. (2020)
Su, B., Lu, S., Tan, C.L.: Robust document image binarization technique for degraded document images. IEEE Trans. Image Process. 22(4), 1408–1417 (2012)
Vo, Q.N., Kim, S.H., Yang, H.J., Lee, G.: Binarization of degraded document images based on hierarchical deep supervised network. Pattern Recognit. 74, 568–586 (2018)
Wu, Y., Natarajan, P., Rawls, S., AbdAlmageed, W.: Learning document image binarization from data. In: International Conference on Image Processing, pp. 3763–3767 (2016)
Zhao, J., Shi, C., Jia, F., Wang, Y., Xiao, B.: Document image binarization with cascaded generators of conditional generative adversarial networks. Pattern Recognit. 96, 106968 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, D., Wu, Y., Zhou, Y. (2021). SauvolaNet: Learning Adaptive Sauvola Network for Degraded Document Binarization. In: Lladós, J., Lopresti, D., Uchida, S. (eds) Document Analysis and Recognition – ICDAR 2021. ICDAR 2021. Lecture Notes in Computer Science(), vol 12824. Springer, Cham. https://doi.org/10.1007/978-3-030-86337-1_36
Download citation
DOI: https://doi.org/10.1007/978-3-030-86337-1_36
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-86336-4
Online ISBN: 978-3-030-86337-1
eBook Packages: Computer ScienceComputer Science (R0)