Abstract
In a low-light condition, the quality of a captured image may be much poorer than that obtained in a normal environment. As an effective preprocessing step, many enhancement algorithms have been proposed to improve the performance of a computer vision task. In most existing algorithms, a image is often enhanced as a whole. As a result, the image may be over-enhanced or under-enhanced due to different degree of exposure in local area. Aiming at this issue, in this paper, we propose a low-light image enhancement algorithm based on image fusion technology. In the proposed method, a fusion strategy is devised by considering the exposure extent of local area. The weight matrix for image fusion is first calculated. Then, the pixel with insufficient exposure is selected according to the adaptive threshold. Next, the multi-exposure images can be synthesized by using the estimated optimal exposure rate. Finally, we use the input image to fuse with the enhanced image for slightly under-exposed images to get the enhancement image, while the severely under-exposed images can be enhanced by fusing a reflection map based on retinex with the enhanced image. Experimental results show that our method can obtain enhancement results with less color and lightness distortion compared to several state-of-the-art methods.
The work was supported by a grant from National Natural Science Foundation of China (No. 61370109), a key project of support program for outstanding young talents of Anhui province university (No. gxyqZD2016013), a grant of science and technology program to strengthen police force (No. 1604d0802019), and a grant for academic and technical leaders and candidates of Anhui province (No. 2016H090).
Access provided by CONRICYT-eBooks. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
In many cases, the quality of the captured images may be affected by environment, equipment and technology. Many useful information of these images may be hidden. On the one hand, images with high visibility are critical to computer vision technology including target recognition and tracking, medical research, can improve the accuracy of algorithms. On the other hand, the clear pictures have good visual effects, which are valuable images, such as space and archaeology. Therefore, the study in low-light image enhancement technology has important value and significance.
Many classic algorithms to deal with low-light images were proposed from different perspectives. Low illumination image intuitive sense is caused by the uneven distribution of pixel level. The most intuitive approach is to stretch the contrast of the image directly. However, this method of stretching contrast may lead to excessive brightness in the local area. Histogram equalization (HE) strategies [1, 2], contextual and variational contrast enhancement (CVC) [3], by seeking a layered difference representation of 2D histograms (LDR) [4], although it is possible to avoid pixel overflow and over-boost by forcing the pixel value in the range of [0,1]. However, the method only focuses on improving the contrast of the image, unable to solve the problem of color saturation. Retinex theory is a method based on human vision systems, which is widely used in image enhancement. Naturalness preserved enhancement algorithm (NPEA) was proposed in [5], the algorithm by non-uniform enlarge the contrast of brightness figure. Though solving the problem of low-light, the method may suffer from under-enhancement and color distortion. In [6], the bilateral filter can be used to generate different filter effects according to different pixel location and gray value. However, this method may appear artifacts in the edge regions of high contrast. The structure-aware smoothing model was proposed in [7], can construct the brightness map rapidly through different weighting schemes. But the algorithm will suffer color supersaturated, causing the enhanced images to be unnatural. Based on the dark channel algorithm in [8,9,10,11], the low-illumination image operates in a similar way to the fog image after inversion, and the enhancement of the images are realized by using the dark channel theory. The work proposed in [12] has applied deep learning used in image processing. Though the method has achieved good results, the global atmospheric light estimation is a challenge. The image fusion method based on camera model was proposed in [13, 14]. This algorithm can make the image enhancement more natural. However, the method is not ideal for images treatment with severe underexposure.
In this paper, we proposed different solutions to different degrees of low-light images. Experiments on a number of challenging images are conducted to reveal the advantages of our method, which can obtain results with less contrast and lightness distortion in comparison with other state-of-the-art methods.
2 Methodology
2.1 Framework Construction
In many cases, the quality of the image is not only affected by the illumination, but the exposure rate of the camera is also the cause of poor visibility. Under the same conditions, we get more information about the better exposure, as shown in Fig. 1. In [14], the mapping function of a series of different exposure rate images is obtained through the input graph, which is called the luminance transformation function (BTF).
where g is the brightness transformation function, and k is the exposure rate. By varying k, we can get a series of images P with different exposure rates.
Many algorithms have a good effect on the low illumination image. However, they may cause over-enhancement, color distortion and unnatural to some extent. We found that images fusion technology can coordinate multiple sensor information in the same scene and help to obtain the information of target or scene more accurately and comprehensively. Therefore, in order to solve the above mentioned problems, we construct an image fusion frame.
where E is the output, W is the weight, the operator \( \circ \) represents element-wise multiplication. In order to achieve a better enhancement, in different situations, the choice of L is different. It is specifically introduced in Sect. 2.4.
2.2 Weight Matrix and Reflection Map Estimation
In the fusion framework, the selection of weight is the key to retain the advantages of each part and avoid weaknesses. To avoid over-enhancement, the building of the weight depends on the illumination. The weighting matrix is calculated as in [13]:
where T is the brightness map of the input image, \( \upmu \) is a parameter that controls the degree of enhancement, the value range of which is [0, 1]. In order to avoid over-enhancement, our algorithm selection μ = 0.5.
The Initialization of the Illumination Map
To preserve the brightness of the image, we initialize the brightness image T as:
for each individual pixel x. illumination map T is solved by optimization.
Optimized Illumination Map
In order to better preserve the edge information of the image, we chose the guide filter to optimize the brightness image T, which was proposed in [15].
where the input image P denote the guide image, \( \hat{T} \) is the rough estimate of the illumination image, \( \varepsilon \) is the regularization parameter, and \( w_{k} \) is the local area of the center of the pixel point k in the guide image. Parameters of \( a_{k} \) and \( b_{k} \) can be calculated by the linear ridge regression method.
Through the processing of the above, we can get further optimized illumination map T:
where \( |w| \) is the number of pixels in the local area \( w_{k} \).
Reflection Estimation
Based on the Retinex theory, we can obtain the reflection map R of the scene through the illumination map T
where R and T are the reflectance map and the illumination map, respectively.
2.3 Exposure Rate Estimation
Adaptive Threshold Selection
As we mentioned earlier, globally enhanced methods may lead to over-enhancement in some regions. Therefore, we proposed the adaptive threshold to filter the pixels. The pixels below the threshold were extracted, and the pixels above the threshold were ignored.
where Q contains all the lower pixels. In this article, the selection of threshold \( Th \), we use the \( Otsu \) threshold method, which can according to the image adaptive selection of the corresponding threshold.
The Determination of Exposure Rate
Intuitively, well-exposed images transfer to us more information than under-exposed images. Image entropy is a statistical form of feature, which reflects the average amount of information in the image. As in [13], the image entropy is calculated as follows:
where \( B = \sqrt[3]{{Q_{r} \circ Q_{g} \circ Q_{b} }} \), \( Q_{r} ,Q_{g} \) and \( Q_{b} \) are the red, green, and blue channel of Q respectively. \( p_{i} \) is the \( i \)-th bin of the histogram of B and N is the number of bins. The optimal k can make the maximum amount of information. As proposed in [13], we obtain the optimal exposure rate by maximizing the image entropy:
2.4 Fusion Strategy
For Slightly Under-Exposed Images
In many cases, a poorly quality image is only partially under-exposed, and the other part is still well. In order to better use this part of the good information, we select the original map to fuse with the image that enhances the exposure rate.
where, P is the input image.
For Severely Under-Exposed Images
In the case of severely underexposed images, the above method is not satisfactory. However, images captured in low-light conditions are often of low visibility, which cause a lot of information to be hidden in the dark. To visualize the hidden information, we propose the following fusion method:
where R is the reflection of the scene.
3 Experiments
In this section, a public dataset with a standard comparison graph (Robust Patch-Based HDR Reconstruction of Synamic Scenes) is used to visualize the vividly and intuitively of our proposed algorithm. We compare our method with other seven algorithms including CVC, HE, LDR, WAHE, Dong, LIME and CAIP in our experiments.
As can be seen from the comparison of Fig. 2, our algorithm is superior to algorithms of CVC, HE, LDR, WAHE and Dong, which can enhance the picture and keep the picture color undistorted. Compared with the LIME method, our algorithm is more natural and more consistent with human visual perception. Figure 3 shows the enhanced effect of our algorithm.
The relative order of image brightness can not only indicate the direction of light source but also the relative change of brightness. The natural degree of image enhancement is closely related to the relative order of brightness in different local regions [5, 7]. In other words, the relative order of the brightness of different regions of the picture should remain unchanged before and after the enhancement. The originally relatively bright region is still brighter after enhancement, and the originally darker region still maintains a darker brightness after enhancement. Therefore, we use the luminance order error (LOE) as a measure of the brightness fidelity performance of the algorithm. The definition of LOE is as follows:
where m is the number of pixels and \( \oplus \) is the exclusive-or operator. \( {\mathbf{Q}} \) is an enhancement image and \( {\mathbf{Q}}_{r} \) is a reference image. If \( p \ge q \), then the function \( U(p,q) \) returns 1, otherwise it returns 0. In addition, \( {\mathbf{Q}}(x) \) and \( {\mathbf{Q}}_{r} (x) \) respectively are the maximum values of R, G and B at the x position of enhanced image and reference image. The smaller the \( LOE \), the better the natural degree of brightness. As a result, the effect of image enhancement is better (Table 1).
4 Conclusions
In this paper, we propose a fusion framework and adaptive enhancement algorithm. This model can better solve the problem of insufficient to enhance or increase, can better keep the naturalness of the image. The selection of weights is the key of fusion, and we use the guide filter to optimize the illumination diagram, thus obtaining more accurate weight. Moreover, different fusion schemes make our algorithm more applicable. The experimental results demonstrated that the proposed method is an effective and feasible low-light image enhancement algorithm.
References
Pisano, E.D., Zong, S., Hemminger, B.M., et al.: Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms. J. Digit. Imaging 11(4), 193 (1998)
Abdullah-Al-Wadud, M., Kabir, M.M., Dewan, A.A., Chae, O.: A dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 53(2), 593–600 (2007)
Celik, T., Tjahjadi, T.: Contextual and variational contrast enhancement. IEEE Trans. Image Process. 20(12), 3431–3441 (2011)
Lee, C., Lee, C., Kim, C.-S.: Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 22(12), 5372–5384 (2013)
Wang, S., Zheng, J., Hu, H.M., Li, B.: Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 22(9), 3538–3578 (2013)
Elad, M.: Retinex by two bilateral filters. In: Kimmel, R., Sochen, N.A., Weickert, J. (eds.) Scale-Space 2005. LNCS, vol. 3459, pp. 217–229. Springer, Heidelberg (2005). https://doi.org/10.1007/11408031_19
Guo, X., Li, Y., Ling, H.: Lime: low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 26(2), 982–993 (2017)
He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. In: IEEE Computer Vision and Pattern Recognition, pp. 1956–1963 (2009)
Narasimhan, S.G., Nayar, S.K.: Contrast restoration of weather degraded images. IEEE Computer Society (2003)
Meng, G., Wang, Y., Duan, J., Xiang, S., Pan, C.: Efficient image dehazing with boundary constraint and contextual regularization. In: IEEE International Conference on Computer Vision, pp. 617–624 (2013)
Dong, X., Wang, G., Pang, Y., Li, W., Wen, J., Meng, W., Lu, Y.: Fast efficient algorithm for enhancement of low lighting video. In: IEEE International Conference on Multimedia and Expo, pp. 1–6 (2011)
Ren, W., Liu, S., Zhang, H., Pan, J., Cao, X., Yang, M.-H.: Single image dehazing via multi-scale convolutional neural networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016, Part II. LNCS, vol. 9906, pp. 154–169. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_10
Ying, Z., Li, G., Ren, Y., Wang, R., Wang, W.: A new image contrast enhancement algorithm using exposure fusion framework. In: Felsberg, M., Heyden, A., Krüger, N. (eds.) CAIP 2017, Part II. LNCS, vol. 10425, pp. 36–46. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-64698-5_4
Ying, Z., Li, G., Ren, Y., Wang, R., Wang, W.: A new low-light image enhancement algorithm using camera response model. In: International Conference on Computer Vision Workshop (2017)
He, K., Sun, J., Tang, X.: Guided image filtering. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part I. LNCS, vol. 6311, pp. 1–14. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15549-9_1
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Wang, YM., Sun, ZL., Han, FQ. (2018). An Effective Low-Light Image Enhancement Algorithm via Fusion Model. In: Huang, DS., Gromiha, M., Han, K., Hussain, A. (eds) Intelligent Computing Methodologies. ICIC 2018. Lecture Notes in Computer Science(), vol 10956. Springer, Cham. https://doi.org/10.1007/978-3-319-95957-3_42
Download citation
DOI: https://doi.org/10.1007/978-3-319-95957-3_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-95956-6
Online ISBN: 978-3-319-95957-3
eBook Packages: Computer ScienceComputer Science (R0)