arxiv:1904.00198v1 [cs.cv] 30 mar 2019 · multi-focus image fusion, a technic to generate an...

6
BOUNDARY AWARE MULTI-FOCUS IMAGE FUSION USING DEEP NEURAL NETWORK Haoyu Ma, Juncheng Zhang, Shaojun Liu, Qingmin Liao Shenzhen Key Laboratory of Information Science and Technology, Shenzhen Engineering Laboratory of IS&DRM, Department of Electronics Engineering, Graduate School at Shenzhen, Tsinghua University, Shenzhen, 518055, China {hy-ma17, zjc16, liusj14}@mails.tsinghua.edu.cn; [email protected] ABSTRACT Since it is usually difficult to capture an all-in-focus image of a 3D scene directly, various multi-focus image fusion meth- ods are employed to generate it from several images focus- ing at different depths. However, the performance of existing methods is barely satisfactory and often degrades for areas near the focused/defocused boundary (FDB). In this paper, a boundary aware method using deep neural network is pro- posed to overcome this problem. (1) Aiming to acquire im- proved fusion images, a 2-channel deep network is proposed to better extract the relative defocus information of the two source images. (2) After analyzing the different situations for patches far away from and near the FDB, we use two net- works to handle them respectively. (3) To simulate the reality more precisely, a new approach of dataset generation is de- signed. Experiments demonstrate that the proposed method outperforms the state-of-the-art methods, both qualitatively and quantitatively. Index TermsImage fusion, multi-focus fusion, convo- lutional neural network, deep learning 1. INTRODUCTION When capturing an image of a 3D scene, it is difficult to take an image where all the objects are focused since the depth- of-field is limited. However, in many image processing tasks, it is more convenient and effective to use all-in-focus images as the input. Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths, is effective to address this prob- lem. For decades, a large number of multi-focus image fusion algorithms have been proposed. Most of them can be broadly categorized into two groups, i.e., transform domain-based al- gorithms [1-7] and spatial domain-based algorithms [8-15]. The transform domain-based algorithms are usually based on multi-scale transform (MST) theories, such as the Lapla- cian pyramid (LP) [1], wavelet transform [2, 3], curvelet transform (CVT) [4], and non-subsampled contourlet trans- form (NSCT) [5]. These methods usually decompose the source images first, then extract and fuse the features of the images, and finally reconstruct the fused images. In addition to MST methods, some feature space-based methods have been proposed in recent years, such as independent compo- nent analysis (ICA) [6], sparse representation (SR) [7] and NSCT-SR [8]. Most of the transform domain methods are efficient, but the fusion images are usually indistinct. The spatial domain-based algorithms can be further di- vided into block-based, region-based and pixel-based fusion algorithms. The block-based algorithms usually divide an im- age into blocks, measure their spatial frequency as well as sum modified-Laplacian [9], and then fuse the image blocks. In these algorithms, such as morphology-based fusion [10], the size of the image block has a great impact and is hard to decide. The region-based algorithms [11] are based on seg- mentation of input images and usually highly depend on seg- mentation accuracy. Recently, several pixel-based algorithms, including guided filtering (GF) [12] and dense SIFT (DSIFT) [13], have been proposed. They can achieve improved results, however, usually suffer from the blocking effect. For all of these traditional fusion methods, both transform domain-based and spatial domain-based, the defocus level de- scriptors and the comparison rules need to be designed man- ually. However, the situations are quite complex for real photos, and the results of these methods are usually imper- fect. In [14], CNN is used to extract the defocus level de- scriptors and the comparison rules in a data-driven way. Un- fortunately, the results of existing neural network based ap- proaches [14, 15] are still unsatisfied, especially for the areas near the focused/defocused boundary (FDB), therefore much post processing is employed to ease this problem. The rea- sons are explained as follows. Firstly, the network structures used in [14, 15] could be improved for the fusion task. Sec- ondly, the situations for areas far away from and near the FDB are quite different, therefore, it is unwise using a single net- work to tackle these two situations together. Thirdly, training datasets used by [14, 15] are inconsistent with the reality. In this paper, a boundary aware multi-focus image fu- sion approach using deep neural network is proposed to over- come the insufficiency, together with a new dataset genera- tion method. The contributions are threefold. (1) Compared arXiv:1904.00198v1 [cs.CV] 30 Mar 2019

Upload: others

Post on 17-Jul-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

BOUNDARY AWARE MULTI-FOCUS IMAGE FUSION USING DEEP NEURAL NETWORK

Haoyu Ma, Juncheng Zhang, Shaojun Liu, Qingmin Liao

Shenzhen Key Laboratory of Information Science and Technology,Shenzhen Engineering Laboratory of IS&DRM, Department of Electronics Engineering,

Graduate School at Shenzhen, Tsinghua University, Shenzhen, 518055, China{hy-ma17, zjc16, liusj14}@mails.tsinghua.edu.cn; [email protected]

ABSTRACT

Since it is usually difficult to capture an all-in-focus image ofa 3D scene directly, various multi-focus image fusion meth-ods are employed to generate it from several images focus-ing at different depths. However, the performance of existingmethods is barely satisfactory and often degrades for areasnear the focused/defocused boundary (FDB). In this paper,a boundary aware method using deep neural network is pro-posed to overcome this problem. (1) Aiming to acquire im-proved fusion images, a 2-channel deep network is proposedto better extract the relative defocus information of the twosource images. (2) After analyzing the different situations forpatches far away from and near the FDB, we use two net-works to handle them respectively. (3) To simulate the realitymore precisely, a new approach of dataset generation is de-signed. Experiments demonstrate that the proposed methodoutperforms the state-of-the-art methods, both qualitativelyand quantitatively.

Index Terms— Image fusion, multi-focus fusion, convo-lutional neural network, deep learning

1. INTRODUCTION

When capturing an image of a 3D scene, it is difficult to takean image where all the objects are focused since the depth-of-field is limited. However, in many image processing tasks,it is more convenient and effective to use all-in-focus imagesas the input. Multi-focus image fusion, a technic to generatean all-in-focus image from several images of the same scenefocusing on different depths, is effective to address this prob-lem. For decades, a large number of multi-focus image fusionalgorithms have been proposed. Most of them can be broadlycategorized into two groups, i.e., transform domain-based al-gorithms [1-7] and spatial domain-based algorithms [8-15].

The transform domain-based algorithms are usually basedon multi-scale transform (MST) theories, such as the Lapla-cian pyramid (LP) [1], wavelet transform [2, 3], curvelettransform (CVT) [4], and non-subsampled contourlet trans-form (NSCT) [5]. These methods usually decompose thesource images first, then extract and fuse the features of the

images, and finally reconstruct the fused images. In additionto MST methods, some feature space-based methods havebeen proposed in recent years, such as independent compo-nent analysis (ICA) [6], sparse representation (SR) [7] andNSCT-SR [8]. Most of the transform domain methods areefficient, but the fusion images are usually indistinct.

The spatial domain-based algorithms can be further di-vided into block-based, region-based and pixel-based fusionalgorithms. The block-based algorithms usually divide an im-age into blocks, measure their spatial frequency as well assum modified-Laplacian [9], and then fuse the image blocks.In these algorithms, such as morphology-based fusion [10],the size of the image block has a great impact and is hard todecide. The region-based algorithms [11] are based on seg-mentation of input images and usually highly depend on seg-mentation accuracy. Recently, several pixel-based algorithms,including guided filtering (GF) [12] and dense SIFT (DSIFT)[13], have been proposed. They can achieve improved results,however, usually suffer from the blocking effect.

For all of these traditional fusion methods, both transformdomain-based and spatial domain-based, the defocus level de-scriptors and the comparison rules need to be designed man-ually. However, the situations are quite complex for realphotos, and the results of these methods are usually imper-fect. In [14], CNN is used to extract the defocus level de-scriptors and the comparison rules in a data-driven way. Un-fortunately, the results of existing neural network based ap-proaches [14, 15] are still unsatisfied, especially for the areasnear the focused/defocused boundary (FDB), therefore muchpost processing is employed to ease this problem. The rea-sons are explained as follows. Firstly, the network structuresused in [14, 15] could be improved for the fusion task. Sec-ondly, the situations for areas far away from and near the FDBare quite different, therefore, it is unwise using a single net-work to tackle these two situations together. Thirdly, trainingdatasets used by [14, 15] are inconsistent with the reality.

In this paper, a boundary aware multi-focus image fu-sion approach using deep neural network is proposed to over-come the insufficiency, together with a new dataset genera-tion method. The contributions are threefold. (1) Compared

arX

iv:1

904.

0019

8v1

[cs

.CV

] 3

0 M

ar 2

019

Page 2: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

Fig. 1. The block diagram of the proposed boundary aware multi-focus fusion method.

with existing networks for fusion [14, 15], a 2-channel struc-ture is designed to better extract the relative defocus informa-tion of the two source images and improve the fusion results.(2) Instead of a single network, two networks are employedto tackle the different situations for areas far away from andnear the FDB respectively. As a consequence, the FDB ofthe fusion image is more clear for the proposed method. (3)To simulate the reality better, especially the situation for areanear the FDB, a more reasonable approach of dataset gener-ation is presented. Based on these three improvements, theproposed method can obtain pleased fusion results. Experi-ments demonstrate that the proposed method outperforms thestate-of-the-art methods, both qualitatively and quantitatively.

2. PROPOSED FUSION METHOD

Our method consists of 3 steps: initial score map generation,score map refinement, as well as post processing and imagefusion. The block diagram of our method is shown in Fig. 1.Firstly, a 2-channel network is proposed to generate an initialscore map. Secondly, the initial score map is refined by twoother networks that are specially designed for areas near andfar away from the FDB. Thirdly, some simple post processingis employed to generate the decision map from the refinedscore map. Then the fusion result is obtained from the twosource images according to the decision map.

2.1. Initial Score Map Generation

Pixel-wised multi-focus image fusion can be viewed as a clas-sification problem, and CNN is the classic model for imageclassification. Therefore it is natural to begin with a simpleCNN. In order to get better fusion results, a network deeperthan [14, 15] is needed. However, when a network is deep-ened, the degradation problem would arise, because the op-timization is not similarly easy for all systems. The residuallearning framework fixes this problem and achieves preciseclassification [16]. Therefore, it is employed here.

Since only the comparative information of input imagesis desired for multi-focus image fusion task, variety of struc-tures might be applied. Tang [15] used a single input net-work, which paid little attention to the relativity between thesource images. Using the two source images together as in-put should be a better choice. Furthermore, for multi-focusimage fusion, two source images are taken from the same po-sition, meaning that they are perfectly aligned in nature. Thesame result is expected to gain when the order of two sourceimages is changed. Therefore, for pixel-wise image fusion,the 2-channel model is more effective and flexible than thesiamese model [17]. As it’s demonstrated in Fig. 2, the twoinput source image patches are simply taken as one imagepatch with two channels.

Fig. 2. The input patch of the 2-channel network.

2.2. Refinement

The area near the FDB of the fusion results is unsatisfied evenusing the proposed network. There are two reasons.

Firstly, the situations are quite different for patches faraway from and near the FDB. For the patches far away fromthe FDB, the patches are totally focused or defocused. Conse-quently, the sharpness of the patch can be used as a metric togenerate the score map. On the contrary, for patches near theFDB, both focused area and defocused area exist. Therefore,

Page 3: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

the sharpness of the patch is not suitable anymore, and theposition of the center pixel becomes even more vital, sincethe proposed method is pixel-wise. For example, the patchshould be thought as focused if the center pixel is in the fo-cused region, even if the defocused area is larger in the wholepatch. In conclusion, it is unwise to handle these two differentsituations using a single network.

Secondly, the number of patches near the FDB is muchfewer than that far away from the FDB, as the training patchesare usually chosen from training images randomly. The im-balance of the training patches will also lead to the bad per-formance near the FDB when using a single network.

Therefore, taking the defocus level comparison task apartas two relatively independent one might be a better choice.Specifically, based on the initial score map, each patch can beclassified to either far away from or near the FDB. Then theyare processed by the normal refinement net and the boundaryrefinement net, respectively, to obtain a better score map. Itis worth to point out that the initial net already has a satis-fied performance over the patches far away from the bound-ary. Therefore, the initial fusion net can serve as the normalrefinement net directly without any loss in performance. Ul-timately, the flow path of the proposed method is simplified.After the initial score map generation, the patches near theFDB are processed by the boundary refinement net. Then theinitial focus scores of these position are replaced.

The classification of patches near the FDB is described asfollows. Based on the initial score map, the average of focusscore (FS) at the centre pixel (m,n) of a patch is calculatedin a surrounding window of size (2l+1)×(2l+1) as follows.

FS(m,n) =1

(2l + 1)2

∑i=m+l

i=m−l

∑j=n+l

j=n−lFS(i, j). (1)

If the average focus score satisfies

0.2 < FS(m,n) < 0.8, (2)

the patch is taken as near the FDB.

2.3. Post Processing and Image Fusion

After the refinement, some simple post processing such asbinarization and small region removal [14] are still needed.Specifically, binarization is applied to the focus score mapto obtain a decision map. Small regions inside the focusedor defocused area are also cleared. These small tricks con-tribute some improvement to the final result, making it morevisually comfortable. We intend to finish the main job bythe networks, therefore, more post processing used in exist-ing works, such as guided filter [14], is not needed.

Once the decision map is obtained, the fusion image(ImgF ) can be directly generated from the decision map(DM ) and the two source images (ImgS) as follows.

ImgF = DM · ImgSA + (1−DM) · ImgSB . (3)

(a) Source image A (b) Source image B

(c) After Initial Net (d) After Refinement

(e) Decision Map (f) Fusion Result

Fig. 3. The fusion maps after each step and the fusion result.

The score maps after each step are shown in Fig. 3. Asshown in Fig. 3(c), the score map after the initial fusion net isdecent, but imperfections exist near the FDB. Fig. 3(d) showsthat the refinement net fixes these defects, and the result nearthe FDB becomes better. Then the post processing removesthe small regions inside, and the decision map is given in Fig.3(e). The fusion image is shown in Fig. 3(f).

3. EXPERIMENTS

In this section, firstly a new approach to generate trainingdataset for deep-learning based multi-focus image fusion isproposed. Then the network details and the settings of train-ing are discussed. Next, the comparative experiments withexisting approaches are set. Finally, the results of proposedfusion method are shown and discussed.

3.1. Proposed Dataset Generation Method

A good training dataset should represent the normal and com-prehensive situations of the task. In other fusion methodsbased on machine learning [14, 15], several training imagesgeneration approaches have been used. Unfortunately, noneof them simulates the reality that the FDB usually coincides

Page 4: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

with the object boundary [18].Considering the reality, the best choice is to use real pho-

tos. However there are only few multi-focus fusion sourceimages, and the ground truth needs to be labeled manually.Therefore, a feasible method is to generate artificial train-ing images that are similar to the reality yet easy to obtain.The foreground images dataset with ground truth is used andsome images without obvious defocus are chosen as the back-ground dataset. Both the original foreground (FG) and thebackground (BG) images are processed by Gaussian filtersfor the blurred images firstly. Then the source image pairs areobtained according to the ground truth (GT ) pixel by pixel.

ImgSA = FGOri ·GT +BGBlu · (1−GT ), (4)ImgSB = FGBlu ·GT +BGOri · (1−GT ). (5)

Specifically, Alpha Matting dataset [19], which contains27 images with ground truth, are used as foreground dataset,and 700 background images from COCO 2017 dataset [20]are used as background dataset. The foreground and back-ground images are combined one by one, therefore, 18,900pairs of source image are obtained. For each source imagepair, we randomly select 10 patches pairs. A training pair willbe labeled 1 if the center pixel is focused in the source imageA and defocused in the source image B, otherwise labeled 0.We then exchange the channels of a patch pair and label it theopposite, so the changing of input order will lead to the sameresult. Each source image pair is augmented by rotating andflipping. Finally, there are 2,268,000 training samples in total.

Additionally, the foreground images from Alpha Mattingare larger than the background images from COCO dataset,so the foreground images are resized half at the very start.

3.2. Network Details and Training Settings

In our implementation, typical Res56 networks are employed.The influences of network depth and the size of patches aretested. The performance will be improved if a deeper net-work or larger patches are used. Nonetheless, there is a trade-off between performance with time and memory cost. Thechosen 56-layer networks with input patch size of 64*64*2,have produced satisfied results.

The same structure is applied to the initial fusion net andthe boundary refinement net, and the training is carried out ondifferent sample sets. The initial fusion net is trained on thefull sample set, while the boundary refinement net is trainedonly on the samples that are near the FDB. Only the patcheshaving at least 10 percent focused pixels and 10 percent defo-cused pixels is chosen.

As for network settings, the normal setups of ResNet onCIFAR-10 [16] are used, since CIFAR-10 has similar inputas our approach. The stochastic gradient descent (SGD) isapplied to train the models, with softmax loss function. 15percents of the training examples are randomly handout for

verification. The initial learning rate is set to 0.001, and re-duced after 80, 120, 160, 180 epochs. 200 epochs are usedin total, and the batch size is 128. Therefore, there are about15000 iterations during one single epoch in total.

3.3. Comparison Settings

We compare the proposed approach with 6 other multi-focusfusion methods, including NSCT [5], SR [7], NSCT-SR [8],GF [12], DSIFT [13], and CNN [14].

The comparison is conducted on 24 pairs of multi-focusimages: 20 pairs from dataset ”Lytro”, the mostly mentioneddataset for image fusion; the others are ”clock”, ”book”, ”lab”and ”flower”, which are commonly used too.

A few objective metrics for fusion image quality as-sessment are used to evaluate the results: normalized mu-tual information (QNMI ) [21], gradient-based fusion perfor-mance (QG) [22], Yang’s metric (QY ) [23], Chen-Blum met-ric (QCB) [24].

3.4. Experimental Results and Analysis

Fig. 4 shows the visual comparison on the ”clock”. Comparedwith other approaches, the proposed method performs well ingeneral. Moreover, both the front clock’s edge and the leftside of number ”8” in the behind clock are clear in our result,as shown in the enlarged squares.

Fig. 5 shows another example, ”lytro-05”. The differencemaps (DM ) is made to show the enhanced visual result:

DM = α · | ImgSB − ImgF | . (6)

Here α is set to 10, showing a more straightforward result.The difference maps are made into gray scale and the pseudo-color transformation is then applied. The difference map isexpected to be red in the area where the source image is fo-cused, and blue where the source image is blurred. Sourceimage B is blurred on the iron grid and clear on the rest area.Compared with existing approaches, the difference map ofour fusion result is distinct near the FDB. There is less incor-rect division in the dark part of difference map, and the brightarea is more continuous and clear than the others too.

The quantitative comparisons are shown in Table 1 andthe results of proposed method without and with refinementare put in the last two columns. The larger evaluation meansbetter fusion result for all the four quality metrics. The shownresults are the average values over the 24 pairs of images, andthe best average result of the compared methods is bold. Thenumber of image pairs that one method beats all the othermethods is shown in the parentheses. In addition, the withoutrefinement version is ignored in the above comparisons.

The proposed method without refinement already outper-forms existing methods, then the refinement improves the re-sults, especially theQCB . It means that both the modificationof the network structure and the refinement are effective. As

Page 5: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

(a) Source image A (b) Source image B (c) NSCT (d) SR

(e) NSCT-SR (f) GF (g) DSIFT (h) CNN (i) The proposed method

Fig. 4. The fusion results of different methods on the ”clock”.

(a) Source image A (b) Source image B (c) NSCT (d) SR

(e) NSCT-SR (f) GF (g) DSIFT (h) CNN (i) The proposed method

Fig. 5. The difference maps of different fusion methods with source image B.

can be seen, the proposed approach markedly outperforms theother fusion methods on all the four quality metrics.

4. CONCLUSIONS

In this paper, we present a boundary aware multi-focus fu-sion approach base on deep neural networks. The proposedmethod utilizes residual networks, and a 2-channel model is

applied to extract more useful information directly from thesource images. Moreover, independent refinement networksare employed after the initial net, to deal with the differentsituations for area near and far away from the FDB, respec-tively. Furthermore, a new way to generate training samplesis also proposed to better approximate the reality. Based onall these improvements, the proposed method obtains promis-ing results and outperforms the state-of-the-art methods, both

Page 6: arXiv:1904.00198v1 [cs.CV] 30 Mar 2019 · Multi-focus image fusion, a technic to generate an all-in-focus image from several images of the same scene focusing on different depths,

Table 1. The quantitative comparison of different fusion methods on 24 widely used image pairs. The results of the proposedmethod without/with refinement are shown in the last two columns. The best average result of the compared methods is in bold.The number of image pairs that one method beats all the other methods is shown in the parentheses. Particularly, the withoutrefinement version is ignored in this comparison.

Metrics NSCT SR NSCT-SR GF DSIFT CNNProposed Method

without/with RefinementQMI 0.9407 (0) 1.0807 (0) 0.9651 (0) 1.1001 (0) 1.1902 (1) 1.1561 (0) 1.2000 1.2002 (23)QG 0.6789 (0) 0.6944 (0) 0.6823 (0) 0.7104 (0) 0.7162 (0) 0.7155 (5) 0.7187 0.7188 (18)QY 0.9549 (0) 0.9663 (0) 0.9576 (0) 0.9775 (1) 0.9841 (2) 0.9851 (6) 0.9867 0.9871 (16)QCB 0.7423 (0) 0.7641 (0) 0.7499 (0) 0.7848 (0) 0.8005 (7) 0.8000 (1) 0.8028 0.8042 (16)

qualitatively and quantitatively.

5. ACKNOWLEDGEMENT

This work was supported by the National Natural ScienceFoundation of China under Grant No. 61771276.

6. REFERENCES

[1] P. J. Burt and E. H. Adelson, “The laplacian pyramid as acompact image code,” IEEE Transactions on Communications,vol. 31, no. 4, pp. 532–540, 1983.

[2] J. J. Lewis, R. J. O’Callaghan, S. G. Nikolov, D. R. Bull, andN. Canagarajah, “Pixel- and region-based image fusion withcomplex wavelets,” Inf. Fusion, vol. 8, no. 2, pp. 119–130,2007.

[3] Y. P. Liu, J. Jin, Q. Wang, Y. Shen, and X. Q. Dong, “Regionlevel based multi-focus image fusion using quaternion waveletand normalized cut,” Signal Processing, vol. 97, pp. 9–30,2014.

[4] F. Nencini, A. Garzelli, S. Baronti, and L. Alparone, “Remotesensing image fusion using the curvelet transform,” Inf. Fu-sion, vol. 8, no. 2, pp. 143–156, 2007.

[5] Qiang Zhang and Bao long Guo, “Multifocus image fusionusing the nonsubsampled contourlet transform,” Signal Pro-cessing, vol. 89, no. 7, pp. 1334 – 1346, 2009.

[6] N. Mitianoudis and T. Stathaki, “Pixel-based and region-basedimage fusion schemes using ica bases,” Inf. Fusion, vol. 8, no.2, pp. 131–142, 2007.

[7] B. Yang and S. T. Li, “Multifocus image fusion and restorationwith sparse representation,” IEEE Transactions on Instrumen-tation and Measurement, vol. 59, no. 4, pp. 884–892, 2010.

[8] Y. Liu, S. P. Liu, and Z. F. Wang, “A general framework forimage fusion based on multi-scale transform and sparse repre-sentation,” Inf. Fusion, vol. 24, pp. 147 – 164, 2015.

[9] H. Wei and Z. L. Jing, “Evaluation of focus measures in multi-focus image fusion,” Pattern Recognition Letters, vol. 28, no.4, pp. 493–500, 2007.

[10] I. De and B. Chanda, “Multi-focus image fusion using amorphology-based focus measure in a quad-tree structure,” Inf.Fusion, vol. 14, no. 2, pp. 136–146, 2013.

[11] M. Li, W. Cai, and Z. Tan, “A region-based multi-sensor imagefusion scheme using pulse-coupled neural network,” PatternRecognition Letters, vol. 27, no. 16, pp. 1948–1956, 2006.

[12] S. T. Li, X. D. Kang, and J. W. Hu, “Image fusion with guidedfiltering,” IEEE Transactions on Image Processing, vol. 22, no.7, pp. 2864–2875, 2013.

[13] Y. Liu, S. P. Liu, and Z. F. Wang, “Multi-focus image fusionwith dense sift,” Inf. Fusion, vol. 23, pp. 139–155, 2015.

[14] Yu Liu, Xun Chen, Hu Peng, and Zengfu Wang, “Multi-focusimage fusion with a deep convolutional neural network,” Inf.Fusion, vol. 36, pp. 191–207, 2017.

[15] Han Tang, Bin Xiao, Weisheng Li, and Guoyin Wang, “Pixelconvolutional neural network for multi-focus image fusion,”Information Sciences, vol. 433-434, pp. 125–141, 2018.

[16] K. M. He, X. Y. Zhang, S. Q. Ren, and J. Sun, “Deep residuallearning for image recognition,” in CVPR, 2016, pp. 770–778.

[17] S. Zagoruyko and N. Komodakis, “Learning to compare imagepatches via convolutional neural networks,” in CVPR, 2015,pp. 4353–4361.

[18] S. Liu, F. Zhou, and Q. Liao, “Defocus map estimation froma single image based on two-parameter defocus model,” IEEETransactions on Image Processing, vol. 25, no. 12, pp. 5943–5956, 2016.

[19] C. Rhemann, C. Rother, J. Wang, M. Gelautz, P. Kohli, andP. Rott, “A perceptually motivated online benchmark for imagematting,” in CVPR, 2009, pp. 1826–1833.

[20] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays,Pietro Perona, Deva Ramanan, Piotr Dollar, and C. LawrenceZitnick, “Microsoft coco: Common objects in context,” inECCV, 2014, pp. 740–755.

[21] M. Hossny, S. Nahavandi, and D. Creighton, “Comments on’information measure for performance of image fusion’,” Elec-tronics Letters, vol. 44, no. 18, pp. 1066–1067, 2008.

[22] C. S. Xydeas and V. Petrovic, “Objective image fusion per-formance measure,” Electronics Letters, vol. 36, no. 4, pp.308–309, 2000.

[23] Cui Yang, Jian-Qi Zhang, Xiao-Rui Wang, and Xin Liu, “Anovel similarity based quality metric for image fusion,” Inf.Fusion, vol. 9, no. 2, pp. 156 – 160, 2008.

[24] Yin Chen and Rick S. Blum, “A new automated quality assess-ment algorithm for image fusion,” Image and Vision Comput-ing, vol. 27, no. 10, pp. 1421 – 1432, 2009.