DOI QR코드

DOI QR Code

Reflectance estimation for infrared and visible image fusion

  • Gu, Yan (North Night Vision Technology Corp., Ltd.) ;
  • Yang, Feng (North Night Vision Technology Corp., Ltd.) ;
  • Zhao, Weijun (North Night Vision Technology Corp., Ltd.) ;
  • Guo, Yiliang (North Night Vision Technology Corp., Ltd.) ;
  • Min, Chaobo (The College of Internet of Things Engineering, HoHai University)
  • Received : 2020.11.15
  • Accepted : 2021.07.03
  • Published : 2021.08.31

Abstract

The desirable result of infrared (IR) and visible (VIS) image fusion should have textural details from VIS images and salient targets from IR images. However, detail information in the dark regions of VIS image has low contrast and blurry edges, resulting in performance degradation in image fusion. To resolve the troubles of fuzzy details in dark regions of VIS image fusion, we have proposed a method of reflectance estimation for IR and VIS image fusion. In order to maintain and enhance details in these dark regions, dark region approximation (DRA) is proposed to optimize the Retinex model. With the improved Retinex model based on DRA, quasi-Newton method is adopted to estimate the reflectance of a VIS image. The final fusion outcome is obtained by fusing the DRA-based reflectance of VIS image with IR image. Our method could simultaneously retain the low visibility details in VIS images and the high contrast targets in IR images. Experiment statistic shows that compared to some advanced approaches, the proposed method has superiority on detail preservation and visual quality.

Keywords

1. Introduction

Infrared (IR) images reflecting the thermal radiation emitted by objects are influenced little by illumination variations and disguises [1]. Thus, IR imaging has been applied widely, such as medical imaging [2], night vision [3] and remote sensing [4], etc. However, it typically suffers from the loss of image details and textures. In contrast, visible (VIS) images could easily capture the information of the reflectance in target scenes. Hence, VIS images have rich details and conform well to the human visual system. Nevertheless, VIS images are sensitive to scene variations such as illumination, smoke, rain and disguises so that the targets in VIS images maybe not salient. Image fusion is put forward to produce a single fused image with both abundant details of VIS images and salient target areas of IR images, harnessing adequately of complementary information from IR and VIS images. Therefore, our major concern of this work is IR and VIS image fusion.

In recent years, numerous VIS and IR image fusion approaches have been proposed as the demand is burgeoning. Generally, fusion algorithms could be classified into six categories [1]: sparse representation, subspace, multi-scale transform, neural network, saliency and hybrid models. Studies of sparse representation [5,6] think that an image is a linearly combination of sparse basis in an over-complete dictionary. As a result, a source image can be resolved into two kinds of layers: detail and base, and following that, the different layers of two to-be-fused images are fused by different fusion regulations. Subspace-based methods [7,8] can extract the intrinsic structures of the source image by projecting high-dimensional image into low-dimensional subspace, in order to remove redundant information from the fused image. Multi-scale transformation [9] are actively employed for image fusion, making an assumption that a source image could be resolved into a variety of layers. The different layers can be fused by certain particular fusion tactic in order to obtain the final fusion. Some transform models, such as pyramid [10], wavelet [11], curvelet [12], Gaussian filters [13] and non-subsampled contourlet transform (NSCT) [14], are often applied for multi-scale decomposition. With the springing up of deep learning techniques, neural network, imitating the perceptual behaviour system of the human brain, performs well on image fusion because of strong adaptability, fault tolerance and noise immunity [15,16]. Saliency-based methods [17,18] could enhance the visual effect of the fused image by using human visual attention. The sixth category is hybrid methods [19,20] combining the superior properties of various fusion algorithms. Take an instance, Yin et al. [21] proposed a hybrid strategy for IR and VIS image fusion based on sparse representation and shift-invariant dual-tree complex shearlet transformation. Recently, with the development of deep learning, some relative fusion approaches have emerged. Convolutional neural network [22,23] and generative adversarial network [24] have been applied successfully.

Essentially, the abovementioned approaches pay special attention to the extraction of detail information from IR and VIS images to ensure rich textures of the fused outcome. However, as for IR images, people are more interested in salient targets than details. Therefore, the high-contrast property in IR images is more valuable to image fusion. If only textures are considered for image fusion, the benefit of the thermal radiation in IR images will disappear. Several methods are proposed for this issue, such as gradient transfer fusion (GTF) [25], fusion generative adversarial network [26], image fusion via detail preserving adversarial learning [1]. The main idea of these methods is that the detail information from VIS images is fused with the intensity information of the thermal radiation in IR images. Thus, the performance of detail preservation directly determines fusion performance. However, the visual quality of VIS images is greatly influenced by illumination conditions. In the bright regions of VIS images, the textures of objects are clear so that they are easily captured and remained in the fused images. Nevertheless, the detail information of dark regions suffer from low contrast, blurry edges and high noise. These raise a tough challenge for detail preservation and will blur the targets in fusion results.

Retinex decomposition [27], regarding the captured picture as the outcome of illumination and reflectance layers, is often used for reflectance estimation to recover details from the captured images. The Retinex model has been applied well to image enhancement, dehazing, fusion and denoising [28]. However, since the details of dark regions in the source images are easily lost in reflectance recovery using the Retinex model [29], the performance of detail preservation on image fusion is limited.

In this work, IR and VIS image fusion via evaluating reflectance with dark region approximation (DRA) is introduced to overcome the abovementioned challenges. At first, the DRA is proposed for the Retinex model to preserve more details of those dark regions in VIS images when processing reflectance estimation. After that, derived from Retinex model with DRA, reflectance estimation could be formulated as an unconstrained optimization problem through Gaussian field criterion and then solved by the quasi-Newton method. Finally, DRA-based reflectance estimated from VIS image is fused with IR image by a strategy called Gamma correction. Qualitative and quantitative comparisons of the method in this paper and 3 advanced strategies are conducted on a public dataset. Unlike the counterparts, our fusion work can illuminate the dark regions of VIS images and produce the fused images with highlighted details that have even low visibility in the source images.

The contributions of this paper lie in two aspects:

1) Reflectance estimation on the DRA-version is developed to exploit detail information from the bright and dark regions of VIS images via two different imaging models respectively. Consequently, our fused images have abundant details and high visibility.

2) Because reflectance estimation based on the DRA is formulated as an unconstrained optimization problem that can be solved by gradient-based numerical optimization technology, our method has a straightforward structure and is easy to be implemented in real applications.

The rest of this paper is unfolded as following. Section 2 presents some basic theories and implementation details of our method. Experiment results and discussion are shown in Section 3. Finally, the concluding remarks are given in Section 4.

2. Methodology

In this part, we describe the basic theory of the DRA and present the implementation details of the proposed method. Fig. 1 shows the framework of our approach.

E1KOBZ_2021_v15n8_2749_f0001.png 이미지

Fig. 1. Framework of the proposed method. Firstly, utilizing DRA, the reflectance is estimated from the VIS image. The IR image is then fused with the DRA-based reflectance layer of the VIS image by a fusion strategy based on Gamma correction.

2.1 Dark Region Approximation (DRA)

In the Retinex model, one VIS image denoted by 𝐋 is the element-wise multiplication of two layers, namely an illumination layer 𝐓 and a reflectance layer 𝐑 

\(\mathbf{L}=\mathbf{R} \circ \mathbf{T}\)       (1)

Eq. (1) could be expressed in logarithmic domain, given as

\(\hat{\mathbf{L}}=\widehat{\mathbf{R}}+\widehat{\mathbf{T}}\)       (2)

in which

\(\begin{aligned} &\hat{\mathbf{L}}=\log (\mathbf{L}) \\ &\hat{\mathbf{R}}=\log (\mathbf{R}) \\ &\widehat{\mathbf{T}}=\log (\mathbf{T}) \end{aligned}\),thought as a map describing detailed structures of the object.

In some previous-proposed Retinex image enhancement algorithm, like single-scale Retinex (SSR) [30] and multi-scale Retinex (MSR) [31], the reflectance \(\widehat{\mathbf{R}}\) is \(\hat{\mathbf{L}}\) minus its Gaussian-filtered value \(\widehat{\mathrm{T}}\)(that is , \(\widehat{\mathbf{R}}\) = \(\hat{\mathbf{L}}\) - \(\widehat{\mathrm{T}}\)). Whereas, the decomposition of VIS image into bright and dark regions gives rise to practical problems about Retinex model so that above expression \(\widehat{\mathbf{R}}\) = \(\hat{\mathbf{L}}\) - \(\widehat{\mathrm{T}}\) changes to

\(\widehat{\mathbf{R}}=\widehat{\mathbf{R}}_{L}+\widehat{\mathbf{R}}_{H}=\left(\hat{\mathbf{L}}_{L}-\widehat{\mathbf{T}}_{L}\right)+\left(\hat{\mathbf{L}}_{H}-\widehat{\mathbf{T}}_{H}\right)\)       (3)

\(\widehat{\mathbf{R}}\),\(\hat{\mathbf{L}}\) and \(\widehat{\mathrm{T}}\) with certain subscript means these are relative region of \(\widehat{\mathbf{R}}\),\(\hat{\mathbf{L}}\) and \(\widehat{\mathrm{T}}\) respectively. Subscript 𝐋 means the dark region while H indicates the bright region. Thanks to good illumination, the discrepancy between \(\hat{\mathbf{L}}_{H}\) and \(\widehat{\mathbf{T}}_{H}\) are apparent. That is to say, the structural details \(\widehat{\mathbf{R}}_{H}\) could be remained well in \(\widehat{\mathbf{R}}\). However, due to the weak intensity of \(\hat{\mathbf{L}}_{L}\), the intensity of \(\widehat{\mathrm{T}}_{L}\) merely stays at a low level. As a result, there is no significant discrepancy between \(\hat{\mathbf{L}}_{L}\) and \(\widehat{\mathrm{T}}_{L}\) in comparison with \(\widehat{\mathbf{R}}_{H}\), which demonstrates that the detail loss of dark regions (, losing \(\widehat{\mathbf{R}}_{L}\), ) is more likely to happen in \(\widehat{\mathbf{R}}\). Fig. 2 is a descriptive graph of Retinex model based reflectance estimation. It could be seen that details in the blue box (a bright region) is more distinguishable and apparent than those in the red box (a dark region). Correspondingly, the direct reflectance estimation in logarithmic space loses the details we need [26].

E1KOBZ_2021_v15n8_2749_f0002.png 이미지

Fig. 2. Descriptive graph of Retinex model-based reflectance estimation. \(\mathbf{T}^{\mathrm{G}}\) , a Gaussian-filtered form of 𝐋, is the illumination layer. An estimated reflectance \(\mathbf{R}^{\mathrm{G}}\) is \(\hat{\mathbf{L}}-\log \left(\mathrm{T}^{\mathrm{G}}\right)\). The red and blue box is dark and bright area separately.

Observing the dark regions, then we put forward DRA: assuming there is few illuminations in the dark area i.e. \(\widehat{\mathrm{T}}_{L}=\mathbf{0}\). Then Eq. (3) is modified as Retinex model in DRA version as (4):

\(\widehat{\mathbf{R}}=\widehat{\mathbf{R}}_{L}+\widehat{\mathbf{R}}_{H} \approx \hat{\mathbf{L}}_{L}+\left(\hat{\mathbf{L}}_{H}-\widehat{\mathbf{T}}_{H}\right)\)       (4)

With DRA’s participation, the intensity distribution in dark areas could be thought as the reflectance in VIS images. Examining Eq. (3) and (4), it looks obvious that \(\hat{\mathbf{L}}_{L}>\hat{\mathbf{L}}_{L}-\widehat{\mathbf{T}}_{L}\), indicating the details of dark regions in has been enhanced. The DRA-based Retinex model reconciles the details of bright and dark regions in \(\widehat{\mathbf{R}}\).

But how to distinguish \(\hat{\mathbf{L}}_{L}\) and \(\hat{\mathbf{L}}_{H}\)? This hanging matter prevents model (4) from being utilized directly on reflectance recovery. In the following section, we will solve this issue.

2.2 DRA-based reflectance estimation

2.2.1 Objective function

It is common sense that illuminating an object partly arouses reflectance so that we defined the reflectance \(\widehat{\mathbf{R}}\) as a map function \(\psi\) relying on the illumination layer \(\widehat{\mathrm{T}}\) and then (2) is rewritten as

\(\hat{\mathbf{L}}=\psi(\widehat{\mathbf{T}})+\widehat{\mathbf{T}}\)     (5)

With Eq. (5), adopting Gaussian field criterion to construct an objective function for solving the map function \(\psi\). The objective function based on Gaussian-field is given by

\(\min _{\psi} E(\psi)=\min _{\psi}-\sum_{i=1}^{M} \exp \left\{-\frac{\left[\hat{\mathbf{L}}_{i}-\left(\psi\left(\widehat{\mathbf{T}}_{i}\right)+\widehat{\mathbf{T}}_{i}\right)\right]^{2}}{2 \sigma^{2}}\right\},\)       (6)

𝜎 is a range parameter, 𝑀 is pixels’ quantity in \(\hat{\mathbf{L}}\)\(\hat{\mathbf{L}}_{i}\) and \(\widehat{\mathbf{T}}_{i}\) is the intensity of the 𝑖-th pixel in \(\hat{\mathbf{L}}\) and \(\widehat{\mathrm{T}}\), separately. The objective function emphasizes the closeness between \(\hat{\mathbf{L}}\) - \(\widehat{\mathrm{T}}\) and map function𝜓(\(\widehat{\mathrm{T}}\)). With the merits of continuous differentiability and convenient computation, Gaussian field criterion is a good distance measurement. Otherwise, this criterion in (6) lays a solid grounding on the distinguishment of the two regions in single-channel images, about which would be talked later.

2.2.2 Reflectance model

In our research, reflectance is seen as one kind of transformation of illumination. It is supposed that a regular pattern of the transformation may exist in VIS image. Therefore, the previous mentioned map function 𝜓(, that is, the reflectance model) could be given as (7) 

\(\psi_{N}\left(\widehat{\mathrm{T}}_{i}\right)=\sum_{n=1}^{N}\left(\sum_{k=0}^{n} \alpha_{n k} x_{i}^{k} y_{i}^{n-k}+\beta_{n} \widehat{\mathbf{T}}_{i}^{n}\right)\),       (7)

𝛼nk and 𝛽n: reflectance parameters.

𝑁: order of the reflectance model.

[𝑥𝑖, 𝑦𝑖]: coordinate vector of the 𝑖th pixel.

The first term of (7) represents the𝑡𝑡 ℎ spacial distribution of the reflectance. Another term expresses the regular pattern of the transformation from illumination to reflectance.

Fundamentally, the reflectance model is a mixture of diverse polynomials depending on intensity values and coordinate vectors, whose highly nonlinearity helps to represent the complex pattern in reflectance.

Write (7) in the matrix form:

\(\psi_{N}\left(\widehat{\mathbf{T}}_{i}\right)=\mathbf{C}^{N} \mathbf{P}_{i}^{N}=\left[\mathbf{c}_{c}^{N} \mid \mathbf{c}_{d}^{N}\right]\left[\mathbf{p}_{c i}^{N} \mid \mathbf{p}_{d i}^{N}\right]^{\mathrm{T}}\) ,      (8)

where \(\mathbf{c}_{c}^{N}=\left[\alpha_{10}, \alpha_{11}, \alpha_{20}, \alpha_{21}, \alpha_{22}, \cdots, \alpha_{N N}\right]\) is the \(1 \times N_{p}\left(N_{p}=N(N+3) / 2\right)\) dimensional vector including every 𝛼nk of the first term in Eq. (7).\(\mathbf{c}_{d}^{N}=\left[\beta_{1}, \beta_{2}, \cdots, \beta_{N}\right]\) is the \(1 \times N\) dimensional vector including every 𝛽of the second term in (7). Thus \(\mathbf{C}^{N}\) is the \(1 \times\left(N_{p}+N\right)\) dimensional reflectance parameter vector. \(\mathbf{p}_{c i}^{N}\) is the \(1 \times N_{p}\) dimension vector including all \(x_{i}^{k} y_{i}^{n-k}\) and \(\mathbf{p}_{d i}^{N}\) is the \(1 \times N\) -dimension vector including every \(\widehat{\mathbf{T}}_{i}^{n}\). Therefore, \(\mathbf{P}_{i}^{N}\) is the \(\left(N_{p}+N\right) \times 1\) dimensional polynomial vector of the 𝑖-th pixel in \(\widehat{\mathrm{T}}\). \([\cdot]^{\mathrm{T}}\) is matrix transposition.

Substituting (8) into (6), the optimization function would be

\(\min _{\mathbf{C}^{N}} E\left(\mathbf{C}^{N}\right)=\min _{\mathbf{C}^{N}}-\sum_{i=1}^{M} \exp \left\{-\frac{\left(\hat{\mathbf{L}}_{i}-\mathbf{C}^{N} \mathbf{P}_{i}^{N}-\widehat{\mathbf{T}}_{i}\right)^{2}}{2 \sigma^{2}}\right\} .\)       (9)

2.2.3 Optimization

It could be observed that Eq. (9) is continuously differentiable to \(\mathbf{C}^{N}\), the reflectance parameter. Owing to the proposed model in the form of polynomial, it is not hard to derive (9) correspondingly as below:

\(\frac{\partial E\left(\mathbf{C}^{N}\right)}{\partial \mathbf{C}^{N}}=\frac{1}{\sigma^{2}} \sum_{i=1}^{M} \mathbf{P}_{i}^{N}\left(\mathbf{C}^{N} \mathbf{P}_{i}^{N}+\widehat{\mathbf{T}}_{i}-\hat{\mathbf{L}}_{i}\right) \exp \left\{-\frac{\left(\hat{\mathbf{L}}_{i}-\mathbf{C}^{N} \mathbf{P}_{i}^{N}-\widehat{\mathbf{T}}_{i}\right)^{2}}{2 \sigma^{2}}\right\}\) (10)

With the derivation(10), it could be accepted that solving \(\mathbf{C}^{N}\), the optimal parameter, with gradient-based numerical optimization approaches, for instance, quasi-Newton method [32]. Prior to that, the acquisition of the illumination layer 𝐓 is still a hard nut to crack in estimating reflectance. It is an ill-posed problem to estimate the illumination and reflectance layers from the captured image in the meantime [33], which objective function (9) and derivation (10) both fail to solve, making the illumination layer determination must be earlier than reflectance estimation.

2.2.4 Illumination layer calculation

According to objective function (9), it is obvious that we deem reflectance estimation to a fitting problem. The captured image \(\hat{\mathbf{L}}\) and the illumination \(\widehat{\mathrm{T}}\) are known data. The reflectance model (7) is a fitting function while the objective function is a fitting criterion aiming at making the reflectance model approximate to \(\hat{\mathbf{L}}\) - \(\widehat{\mathrm{T}}\). Since this model is made up of polynomials, it could illustrate the regular pattern of the reflectance, \(\widehat{\mathbf{R}}\) = \(\hat{\mathbf{L}}\) - \(\widehat{\mathrm{T}}\). In some previous-proposed Retinex image enhancement algorithm like SSR and MSR, the illumination layer is often gotten from the Gaussian-filtered captured image. If the objective function is optimized by Gaussian-filtered illumination layer acquired by \(\widehat{\mathrm{T}}^{\mathrm{G}}=\log \left(\mathrm{T}^{\mathrm{G}}\right)\), the reflectance estimation might have similar defect to direct-estimated reflectance ce as in Section 2.1.

The penalty curve plotted in Fig. 3 shows the Gaussian criterion is pretty tolerant to large \(x\left(\text { e.g. } 1-\exp \left\{-x^{2} /\left(2 \sigma^{2}\right)\right\}\right)\), indicating the property of poor response to large x, with which coarse blur is i2 ntrodu2 ced to obtain the illumination and to estimate reflectance by DRA-based Retinex model (4) then.

E1KOBZ_2021_v15n8_2749_f0003.png 이미지

Fig. 3. Penalty curve of Gaussian criterion

To be frankly, the process of coarse blur is pretty straightforward. Make the pixel value zero if the parity of its and coordinates numbers are the same, and then the Gaussian-based illumination layer \(\widehat{\mathbf{T}}^{\mathrm{G}}\) is further degraded (as shown in Fig. 4). \(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\) is the pixel value in the coarse-blur-based illumination layer \(\widehat{\mathbf{T}}^{C}\). When optimizing the objective function (9) with \(\widehat{\mathbf{T}}_{i}^{\mathrm{C}} \neq 0\) pixels, we fit the reflectance model to the differences between \(\hat{\mathbf{L}}\) and \(\widehat{\mathbf{T}}^{\mathrm{G}}\). On condition that \(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\) is non-zero, the reflectance model is more willing to show reflectance’s regular pattern of brigC ht area with Retinex model.

E1KOBZ_2021_v15n8_2749_f0004.png 이미지

Fig. 4. DRA-based reflectance estimating. \(\mathrm{T}^{\mathrm{C}}\) is the coarse-blurred illumination layer from \(\mathbf{T}^{\mathrm{G}}\). The red rectangle means a dark region and the blue rectangle means a bright region.

At the same time, if \(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\) = 0, the term \(\left[\hat{\mathbf{L}}_{i}-\left(\psi\left(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\right)+\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\right)\right]^{2}\)in the objective function would be \(\left[\hat{\mathbf{L}}_{i}-\psi\left(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\right)\right]^{2}\). The objective function responses weakly to large numbers of \(\left[\hat{\mathbf{L}}_{i}-\psi\left(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}\right)\right]^{2}\)in Fig. 3. Thus, we could fit the reflectance to the weak intensities of the captured picture by arranging a suitable value of the range parameter 𝜎. This is to say, when \(\widehat{\mathbf{T}}_{i}^{\mathrm{C}}=0\), the reflectance model is prone to show the intensity distribution’s regular pattern in the dark regions.

With Eq. (9), we could initially set a soft threshold to make decomposition of a single-channel picture into bright and dark areas. Followed by modelling the reflectance layer of these areas by \(\widehat{\mathbf{R}}_{H}=\hat{\mathbf{L}}_{H}-\widehat{\mathbf{T}}_{H}\) and \(\widehat{\mathbf{R}}_{L}=\hat{\mathbf{L}}_{L}\) respectively. Ending with reconciling the reflectance’s regular patterns in bright and dark areas with the reflectance model. Therefore, the objective function attaching the coarse-blurred illumination layer estimates the optimal reflectance model and it is this model that could express the regular patterns with \(\widehat{\mathbf{R}}=\hat{\mathbf{L}}_{L}+\left(\hat{\mathbf{L}}_{H}-\widehat{\mathbf{T}}_{H}\right)\) , DRA-based Retinex model.

After determining the optimal reflectance model, DRA-based reflectance layer \(\mathbf{R}^{C}\) is

\(\mathbf{R}^{\mathrm{C}}=\exp \left\{\psi_{N}^{\mathrm{C}}(\hat{\mathbf{L}})\right\}\),       (11)

where \(\psi_{N}^{\mathrm{C}}\) is the N-dimensional optimal reflectance model.

Making a comparison between the DRA-estimated reflectance estimation in Fig. 4 and the direct-estimated reflectance in Fig. 2, it is salient that DRA-based reflectance makes details in bright regions more clear and illuminates those in dark regions as well.

Fig. 5 presents several DRA-based results of reflectance estimation of VIS images. We could see that the DRA-based reflectance layers have abundant details and good contrast both in the dark and the bright regions. Some details hard to observe in the source images are more salient in the DRA-based reflectance layers.

E1KOBZ_2021_v15n8_2749_f0005.png 이미지

Fig. 5. DRA-based results of reflectance estimation from VIS images. From top to bottom: VIS images and the corresponding DRA-based reflectance layers. The regions pointed by the arrows are the dark regions.

2.3 Fusion based on Gamma correction

According to the DRA-based reflectance, a strategy based on Gamma correction is developed to IR and VIS image fusion, which could be formulated as

\(\mathbf{F}=\mathbf{L}_{i r} \diamond\left(\mathbf{R}_{v i s}^{\mathrm{C}}+\mathbf{t}\right)\),       (12)

where the operator ⋄ denotes element-wise power. 𝐅 denotes the final fusion result. \(\mathbf{L}_{i r}\) denotes IR image and \(\mathbf{R}_{v i s}^{\mathrm{C}}\) denotes the DRA-based reflectance of VIS image. The intensity ranges of IR and VIS images have to be scaled into [0,1]. 𝐭 is a threshold matrix which has the same size with \(\mathbf{R}_{v i s}^{\mathrm{C}}\) and the every element is defined as \(\mathbf{t}_{i}=1-\operatorname{otsu}\left(\mathbf{R}_{\text {vis }}^{\mathrm{C}}\right)\), where otsu(\(\mathbf{R}_{v i s}^{\mathrm{C}}\)) represents the threshold calculated from \(\mathbf{R}_{v i s}^{\mathrm{C}}\) by OTSU algorithm [34].

By using the Eq. (12), the details estimated from VIS images by the DRA-based Retinex model are fused with IR images, because the Gamma value of each pixel in IR images is evaluated according to \(\mathbf{R}_{v i s}^{\mathrm{C}}\). Moreover, according to the threshold matrix , the Gamma values of the pixels with low intensity in VIS images are greater than 1, while those of the pixels with high intensity are less than 1. As a result, the contrast of the final fusion can be enhanced further. The process of our method is outlined in Algorithm 1.

Algorithm 1. Image fusion using DRA reflectance estimation

3. Experiment

3.1 Parameter setting

Algorithm 1 tells us we need 2 input parameters: the range parameter 𝜎 and the order 𝑁, determining the width of the response range of Gaussian fields criterion in the objective ​​​​function (9) and the nonlinearity of the reflectance model, respectively. Here, after plenty of trials, we got the optimal setting is 𝜎 = 0.6 and 𝑁 = 4. In addition, because we map the grey range of the input image into , the optimal 𝜎 and 𝑁 would be similar on different samples.

3.2 Dataset

The dataset tested in this work contains 30 pairs of IR and VIS images from TNO dataset1 containing aligned pairs of IR and VIS image. To explore the fusion performance of the proposed way, we compare it with 3 fusion methods, namely, two-scale image fusion based on visual saliency (TSIFVS) [35], GTF [25] and image fusion based on gradient transfer optimization model (GTOM) [36]. All of these methods are programmed in Matlab and parameters set goes along with original papers. Each experiment is managed on a computer with the configuration of Windows 10 OS (64 bits) with Intel Core i7-9700K CPU @3.60GHz and 16GB RAM. The average runtime of our method is 2.42 seconds for a pair of IR and VIS images with 620 450 pixels.

3.3 Qualitative× comparison

We pick up 30 pairs of IR and VIS images from TNO dataset as the testing set of qualitative comparison. Five typical pairs are selected for qualitative illustration from the testing set as Fig. 6.

E1KOBZ_2021_v15n8_2749_f0006.png 이미지

Fig. 6. Qualitative fusion results of 5 typical pairs of IR and VIS image from TNO dataset. The blue&yellow boxes emphasize the area of emerging marked differences between our result and the others’. From left: VIS images, IR images, our proposed method, TSIFVS, GTF and GTOM.

The original IR and VIS images are located at the first two columns of Fig. 6. The fusion results of our method are shown in the 3rd row, while the rest of Fig. 6 correspond to the results of the other approaches. It is apparent that all methods could to some extent fuse the information from the original IR and VIS images. However, from the qualitative comparison we can see that compared to the competitors, our method have an advantage on detail preservation. For instance, in our fusion images of Fig. 6(a) and (b), the brightness of the dark regions in the VIS images has substantial growth. Hence, the hard-visible details in the original VIS images, such as shrubs and trees, are preserved and enhanced in our fusion results. From Fig. 6(c) we can see that the texture of the jeep has high saliency in the image fusion of our method. As for Fig. 6(d), the targets, such as vehicle, signpost and street lamp, have good visibility in our fusion result. On the contrary, the detail of VIS images is not well preserved by TSIFVS, GTF and GTOM. Undoubtedly, the superiority of our approach on preserving details thanks to the DRA-based reflectance estimated from VIS images. Furthermore, exploiting the details information of the dark areas of VIS images via DRA-based reflectance estimation is very helpful to improve the visibility of fused images.

In the boxes of Fig. 6(e), the details in the IR image are more salient than those in the VIS image. From the fused images of the competitors, we can see that although the content of IR and VIS images are fused, details from the IR image become rather blurry. This is because the fused image suffers from the low saliency of the VIS image. It is obvious that our method could handle this problem. This demonstrates that the DRA-based reflectance estimated from VIS images could retain the salient information of the source images. It is also proved that the fusion model (12) could simultaneously preserve the textural detail information of IR and VIS images.

Fig. 7 shows the fusion outcome of 2 pairs of IR and VIS image pairs with high noise. We can see that noise affects us more slightly than other competitors. The reason is that DRA-estimated reflectance is able to limit noise via the regularity of the reflectance model (7).

E1KOBZ_2021_v15n8_2749_f0007.png 이미지

Fig. 7. Fusion results on 2 pairs of IR and VIS image pairs with high noise. We have zoomed some regions in the box in and put them at the corner of each subplot. From left: VIS images, IR images, our proposed method, TSIFVS, GTF and GTOM.

3.4 Quantitative comparison

In this work, six metrics are introduced to quantitative evaluation. They are: Entropy (EN) [37]: information amount in an image.

Correlation coefficient (CC) [38]: the degree of linear correlation between the source image and the fused image.

Standard deviation (SD) [39]: the contrast distribution of an image.

Spatial frequency (SF) [40]: the richness of the textural details of an image.

Visual information fidelity (VIF) [41]: the information fidelity of the fused image.

Structural similarity index measure (SSIM) [42]: the structural similarity between the source image and the fused image.

In terms of these 6 metrics, the larger the value, the better the performance would be.

The results of the six metrics are reported in Fig. 8. It is clear that our experiment has the best SD, EN, CC, VIF and SF for most pairs of image. Moreover, the mean values of these 5 metrics of our method respectively are the highest among all these approaches. The highest SD indicates that the contrast of our fused images is the best among the approaches. The highest EN demonstrates that the information of our fusion results are more abundant than those of the competitors. The greatest CC suggests that our fused images have strong correlation with the original images. The highest VIF demonstrates that our fused images have good visual quality and conforms well to the human visual system. The best SF means that there are rich details and textures in our fused images. However, SSIM of our method is relatively low because our method have good performance on detail preservation. Some barely visible details in the source images could be clearly maintained in our fused images. Moreover, our method could simultaneously exploit the textural detail information from IR and VIS images. Thus, some areas in our fused images might look neither like VIS images nor like IR images, resulting in low structural similarity. A representative example can be seen in Fig. 6(a), in which the details of the shrubs have low intensity in the VIS image and are not visible in the IR image; however, in order to preserve the details in dark regions, the brightness of the shrubs in our fused image is increased greatly. Hence, the area of the shrubs in our fused image looks neither like VIS images nor like IR images. Similar situation also occurs in the trees of Fig. 6(b) and the jeep of Fig. 6(c).

E1KOBZ_2021_v15n8_2749_f0008.png 이미지

Fig. 8. Quantitative comparison of 6 evaluation metrics on TNO dataset. Three advanced methods, TSIFVS, GTF and GTOM, are used for comparison. The numbers in the legends are the average values of evaluation metrics.

4. Conclusion

In this paper, a reflectance estimation way with dark region approximation (DRA) has been proposed to IR and VIS image fusion. The performance of Retinex model has been better taking advantages of DRA, especially that of detail preservation of dark regions of VIS images. DRA-based reflectance of VIS image with the Gaussian field criterion is evaluated by the quasi-Newton method. Then, it is fused with IR image by a strategy based on Gamma correction to generate the final fusion result. Qualitative and quantitative comparisons unveil the superiority of our method over other advanced approaches on preserving details and visual quality.

Admittedly, the proposed method has done well in our experiments evaluated from many respects, however, the high computation complexity of it is still a problem, increasing the difficulty in practical applications. Accordingly, we will focus on simplifying the structure of our method and improving the time cost of reflectance estimation in future.

Acknowledgements

The authors gratefully acknowledge the financial supports from the National Natural Science Foundation of China (No.61901157).

References

  1. J. Ma, P. Liang, W. Yu, C. Chen, X. Guo, J. Wu and J. Jiang, "Infrared and visible image fusion via detail preserving adversarial learning," Inf. Fusion, vol. 54, pp. 85-98, Jul. 2019. https://doi.org/10.1016/j.inffus.2019.07.005
  2. F. An, X. Zhou and D. Lin, "Multiscale self-coordination of bidimensional empirical mode decomposition in image fusion," KSII Trans. Int. Inf. Syst., vol. 9, no. 4, pp. 1441-1456, Apr. 2015. https://doi.org/10.3837/tiis.2015.04.010
  3. W. Jeong, B. Han, H. Yang and Y. Moon, "Real-time visible-infrared image fusion using multi-guided filter," KSII Trans. Int. Inf. Syst., vol. 13, no. 6, pp. 3092-3107, Jun. 2019. https://doi.org/10.3837/tiis.2019.06.018
  4. C. Deng, Z. Wang, X. Li, H. Li and C. C. Cavalcante, "An improved remote sensing image fusion algorithm based on IHS transformation," KSII Trans. Int. Inf. Syst., vol. 11, no. 3, pp. 1633-1649, Mar. 2017. https://doi.org/10.3837/tiis.2017.03.021
  5. Y. Liu, X. Chen, R. K. Ward and Z. J. Wang, "Image fusion with convolutional sparse representation," IEEE Signal Process. Lett., vol. 23, no. 12, pp. 1882-1886, Dec. 2016. https://doi.org/10.1109/LSP.2016.2618776
  6. J. Wang, J. Peng, X. Feng, G. He and J. Fan, "Fusion method for infrared and visible images by using non-negative sparse representation," Infrared Phys. Technol., vol. 67, pp. 477-489, Nov. 2014. https://doi.org/10.1016/j.infrared.2014.09.019
  7. D. P. Bavirisetti, G. Xiao and G. Liu, "Multi-sensor image fusion based on fourth order partial differential equations," in Proc. of Int. Conf. on Inf. Fusion, pp. 701-709, July 10-13, 2017.
  8. W. Kong, Y. Lei and H. Zhao, "Adaptive fusion method of visible light and infrared images based on non-subsampled shearlet transform and fast non-negative matrix factorization," Infrared Phys. Technol., vol. 67, pp. 161-172, Nov. 2014. https://doi.org/10.1016/j.infrared.2014.07.019
  9. S. Li, X. Kang, J. Hu, "Image fusion with guided filtering," IEEE Trans. Image Process., vol. 22, no. 7, pp. 2864-2875, Jul. 2013. https://doi.org/10.1109/TIP.2013.2244222
  10. A. Toet, "Image fusion by a ratio of low-pass pyramid," Pattern Recognit. Lett., vol. 9, no. 4, pp. 245-253, May. 1989. https://doi.org/10.1016/0167-8655(89)90003-2
  11. P. Hill, M. E. Al-Mualla and D. Bull, "Perceptual image fusion using wavelets," IEEE Trans. Image Process., vol. 26, no. 3, pp. 1076-1088, Mar. 2017. https://doi.org/10.1109/TIP.2016.2633863
  12. M. Choi, R. Kim, M. Nam and H. Kim, "Fusion of multispectral and panchromatic satellite images using the curvelet transform," IEEE Geosci.Remote Sens. Lett., vol. 2, no. 2, pp. 136-140, Apr. 2005. https://doi.org/10.1109/LGRS.2005.845313
  13. Z. Zhou, B.Wang, S. Li and M. Dong, "Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters," Inf. Fusion, vol. 30, pp. 15-26, Jul. 2016. https://doi.org/10.1016/j.inffus.2015.11.003
  14. F. Meng, M. Song, B. Guo, R. Shi and D. Shan, "Image fusion based on object region detection and non-subsampled contourlet transform," Comput. Electr. Eng., vol. 62, pp. 375-383, Aug. 2017. https://doi.org/10.1016/j.compeleceng.2016.09.019
  15. W. Kong, L. Zhang and Y. Lei, "Novel fusion method for visible light and infrared images based on NSST-SF-PCNN," Infrared Phys. Technol., vol. 65, pp. 103-112, Jul. 2014. https://doi.org/10.1016/j.infrared.2014.04.003
  16. T. Xiang, L. Yan and R. Gao, "A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain," Infrared Phys. Technol., vol. 69, pp. 53-61, Mar. 2015. https://doi.org/10.1016/j.infrared.2015.01.002
  17. X. Zhang, Y. Ma, F. Fan, Y. Zhang and J. Huang, "Infrared and visible image fusion via saliency analysis and local edge-preserving multi-scale decomposition," JOSA A., vol. 34, no. 8, pp. 1400-1410, Aug. 2017. https://doi.org/10.1364/JOSAA.34.001400
  18. J. Zhao, Y. Chen, H. Feng, Z. Xu and Q. Li, "Infrared image enhancement through saliency feature analysis based on multi-scale decomposition," Infrared Phys. Technol., vol. 62, pp. 86-93, Jan. 2014. https://doi.org/10.1016/j.infrared.2013.11.008
  19. Y. Liu, S. Liu and Z.Wang, "A general framework for image fusion based on multi-scale transform and sparse representation," Inf. Fusion, vol. 24, pp. 147-164, Jul. 2015. https://doi.org/10.1016/j.inffus.2014.09.004
  20. W. Zhao, H. Lu and D. Wang, "Multisensor image fusion and enhancement in spectral total variation domain," IEEE Trans. Multimedia, vol. 20, no. 4, pp. 866-879, Apr. 2018. https://doi.org/10.1109/tmm.2017.2760100
  21. M. Yin, P. Duan,W. Liu and X. Liang, "A novel infrared and visible image fusion algorithm based on shift-invariant dual-tree complex shearlet transform and sparse representation," Neurocomputing, vol. 226, pp. 182-191, Feb. 2017. https://doi.org/10.1016/j.neucom.2016.11.051
  22. Y. Liu, X. Chen, H. Peng, Z. Wang, "Multi-focus image fusion with a deep convolutional neural network," Inf. Fusion, vol. 36, pp. 191-207, Jul. 2017. https://doi.org/10.1016/j.inffus.2016.12.001
  23. H. Li, X.-J. Wu, J. Kittler, "Infrared and visible image fusion using a deep learning framework," in Proc. of International Conference on Pattern Recognition, pp. 2705-2710, Nov. 2018.
  24. X. Mao, Q. Li, H. Xie, R.Y. Lau, Z. Wang, S.P. Smolley, "Least squares generative adversarial networks," in Proc. of IEEE International Conference on Computer Vision, pp. 2813-2821, Dec. 2017.
  25. J. Ma, C. Chen, C. Li and J. Huang, "Infrared and visible image fusion via gradient transfer and total variation minimization," Inf. Fusion, vol. 31, pp. 100-109, Sep. 2016. https://doi.org/10.1016/j.inffus.2016.02.001
  26. J. Ma, W. Yu, P. Liang, C. Li and J. Jiang, "FusionGAN: a generative adversarial network for infrared and visible image fusion," Inf. Fusion, vol. 48, pp. 11-26, Aug. 2019. https://doi.org/10.1016/j.inffus.2018.09.004
  27. E. H. Land, "The Retinex theory of color vision," Sci. Amer., vol. 237, no. 6, pp. 108-128, Dec. 1977. https://doi.org/10.1038/scientificamerican1277-108
  28. J. Qu, Y. Li, Q. Du and H. Xia, "Hyperspectral and panchromatic image fusion via adaptive tensor and multi-scale retinex algorithm," IEEE Access, vol. 8, pp. 30522-30532, Feb. 2020. https://doi.org/10.1109/ACCESS.2020.2972939
  29. X. Fu, D. Zeng, Y. Huang, X. Zhang, and X. Ding, "A weighted variational model for simultaneous reflectance and illumination estimation," in Proc. of IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 2782-2790, Jun. 2016.
  30. D. J. Jobson, Z. Rahman, and G. A. Woodell, "Properties and performance of a center/surround Retinex," IEEE Trans. Image Process., vol. 6, no. 3, pp. 451-462, Mar. 1997. https://doi.org/10.1109/83.557356
  31. D. J. Jobson, Z. Rahman, and G. A. Woodell, "A multiscale Retinex for bridging the gap between color images and the human observation of scenes," IEEE Trans. Image Process., vol. 6, no. 7, pp. 965-976, Jul. 1997. https://doi.org/10.1109/83.597272
  32. M. H. Loke and R. D. Barker, "Rapid least-squares inversion of apparent resistivity pseudosections by a quasi-Newton method," Geophysical Prospecting, vol. 44, no. 1, pp. 131-152, Jan. 1996. https://doi.org/10.1111/j.1365-2478.1996.tb00142.x
  33. X. Ren, W. Yang, W. Cheng and J. Liu, "LR3M: robust low-light enhancement via low-rank regularized Retinex model," IEEE Trans. Image Process., vol. 29, pp. 5862-5876, Apr. 2020. https://doi.org/10.1109/TIP.2020.2984098
  34. N. Otsu, "A threshold selection method from gray-level histograms," IEEE Trans. Syst. Man Cy., vol. 9, no. 1, pp. 62-66, Jan. 1979. https://doi.org/10.1109/TSMC.1979.4310076
  35. D. P. Bavirisetti and R. Dhuli, "Two-scale image fusion of visible and infrared images using saliency detection," Infrared Phys. Technol., vol. 76, pp. 52-64, May. 2016. https://doi.org/10.1016/j.infrared.2016.01.009
  36. R. Yu, W. Chen and D. Zhou, "Infrared and visible image fusion based on gradient transfer optimization model," IEEE Access, vol. 8, pp. 50091-50106, Mar. 2020. https://doi.org/10.1109/ACCESS.2020.2979760
  37. J. W. Roberts, J. Van Aardt and F. Ahmed, "Assessment of image fusion procedures using entropy, image quality, and multispectral classification," J. Appl. Remote Sens., vol. 2, no. 1, pp. 023522, Jan. 2008. https://doi.org/10.1117/1.2945910
  38. M. Deshmukh and U. Bhosale, "Image fusion and image quality assessment of fused images," Int. J. Image Process., vol. 4, no. 5, pp. 484-508, Dec. 2010.
  39. Y. Rao, "In-fibre bragg grating sensors," Meas. Sci. Technol., vol. 8, no. 4, pp. 355, 1997. https://doi.org/10.1088/0957-0233/8/4/002
  40. A. M. Eskicioglu and P. S. Fisher, "Image quality measures and their performance," IEEE Trans. Commun., vol. 43, no. 12, pp. 2959-2965, Dec. 1995. https://doi.org/10.1109/26.477498
  41. Y. Han, Y. Cai, Y. Cao and X. Xu, "A new image fusion performance metric based on visual information fidelity," Inf. fusion, vol. 14, no. 2, pp. 127-135, Apr. 2013. https://doi.org/10.1016/j.inffus.2011.08.002
  42. Z. Wang and A. C. Bovik, "A universal image quality index," IEEE Signal Process. Lett., vol. 9, no 3, pp. 81-84, Mar. 2002. https://doi.org/10.1109/97.995823