11
IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 1 Structure-preserving Guided Retinal Image Filtering and Its Application for Optic Disc Analysis Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon Wing Kee Wong, Jiang Liu Abstract—Retinal fundus photographs have been used in the diagnosis of many ocular diseases such as glaucoma, pathological myopia, age-related macular degeneration and diabetic retinopa- thy. With the development of computer science, computer aided diagnosis has been developed to process and analyse the retinal images automatically. One of the challenges in the analysis is that the quality of the retinal image is often degraded. For example, a cataract in human lens will attenuate the retinal image, just as a cloudy camera lens which reduces the quality of a photograph. It often obscures the details in the retinal images and posts challenges in retinal image processing and analysing tasks. In this paper, we approximate the degradation of the retinal images as a combination of human-lens attenuation and scattering. A novel structure-preserving guided retinal image filtering (SGRIF) is then proposed to restore images based on the attenuation and scattering model. The proposed SGRIF consists of a step of global structure transferring and a step of global edge-preserving smoothing. Our results show that the proposed SGRIF method is able to improve the contrast of retinal images, measured by histogram flatness measure, histogram spread and variability of local luminosity. In addition, we further explored the benefits of SGRIF for subsequent retinal image processing and analysing tasks. In the two applications of deep learning based optic cup segmentation and sparse learning based cup-to-disc ratio (CDR) computation, our results show that we are able to achieve more accurate optic cup segmentation and CDR measurements from images processed by SGRIF. Index Terms- retinal image processing, segmentation, computer aided diagnosis I. I NTRODUCTION Retinal fundus photographs have been widely used by clini- cians to diagnose and monitor many ocular diseases including glaucoma, pathological myopia, age-related macular degen- eration, and diabetic retinopathy. Since manual assessment of the images is tedious, expensive and subjective, computer aided diagnosis methods [1] have been developed to analyse the images automatically for optic disc segmentation [2]–[5], Manuscript received March 19, 2018; revised May 12; accepted May 16, 2018. This work was supported in part under grant Y80002RA01 by Cixi institute of biomedical engineering, Chinese Academy of Sciences, China and Ningbo 3315 Innovation team grant Y61102DL03. Asterisk indicates corresponding author. *J. Cheng is with Cixi Institute of Biomedical Engineering, Chi- nese Academy of Sciences, China. (email: [email protected], [email protected]) Z. Li is with Institute for Infocomm Research, Agency for Science, Technology and Research, Singapore. (email: [email protected]). . Z. Gu is with Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences, and Shanghai University, China (email: [email protected]). H. Fu and D. W. K. Wong are with Ocular Imaging (iMED) Department in Institute for Infocomm Research, Agency for Science, Technology and Re- search, Singapore (email: [email protected], [email protected]). J. Liu is with Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences, China (email: [email protected]) (a) (b) (c) (d) Fig. 1. Retinal images: (a)-(b) from eyes with cataract and (c)-(d) from eyes without cataract optic cup segmentation or cup to disc ratio assessment [6]– [9], retinal vessel detection [10], [11], glaucoma detection [12], [13], diabetic retinopathy detection [14], [15], age-related macular degeneration detection [16]–[19], and pathological myopia detection [20]. Image quality is an essential factor for proper development and validation of the algorithms [21]. Very often, low quality images lead to poor performance of automatic analysis. Therefore, it is necessary to restore the image for better analysis. There are many factors that may affect the quality of retinal images. In retinal imaging, an illumination light passes through the lens of human eye to reach the retina. Ideally, the light will be reflected back to the fundus camera by the retina to form the retinal images. However, the human eye is not a perfect optical system and the light received by the fundus camera is often attenuated along the path of the light. This can be serious when the lens of human eye is affected by diseases such as cataracts. Cataract is caused by clouding of human-lens in the eye. Since a retinal image is captured through the lens of human eye, a cataract will lead to attenuation and scattering of the light travelling through the human-lens, just as a cloudy camera lens reducing the quality of a photograph. The process of retina images by cataractous human-lens is referred to as clouding and the processing to remove the effect is referred to as declouding. Meanwhile, cataract accounts for 33% of blindness worldwide [22] and its global prevalence in adults over 50 years of age was about 47.8% in 2002 [23]. The arXiv:1805.06625v2 [cs.CV] 22 May 2018

Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 1

Structure-preserving Guided Retinal Image Filteringand Its Application for Optic Disc AnalysisJun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon Wing Kee Wong, Jiang Liu

Abstract—Retinal fundus photographs have been used in thediagnosis of many ocular diseases such as glaucoma, pathologicalmyopia, age-related macular degeneration and diabetic retinopa-thy. With the development of computer science, computer aideddiagnosis has been developed to process and analyse the retinalimages automatically. One of the challenges in the analysis is thatthe quality of the retinal image is often degraded. For example,a cataract in human lens will attenuate the retinal image, just asa cloudy camera lens which reduces the quality of a photograph.It often obscures the details in the retinal images and postschallenges in retinal image processing and analysing tasks. Inthis paper, we approximate the degradation of the retinal imagesas a combination of human-lens attenuation and scattering. Anovel structure-preserving guided retinal image filtering (SGRIF)is then proposed to restore images based on the attenuationand scattering model. The proposed SGRIF consists of a step ofglobal structure transferring and a step of global edge-preservingsmoothing. Our results show that the proposed SGRIF methodis able to improve the contrast of retinal images, measured byhistogram flatness measure, histogram spread and variability oflocal luminosity. In addition, we further explored the benefits ofSGRIF for subsequent retinal image processing and analysingtasks. In the two applications of deep learning based optic cupsegmentation and sparse learning based cup-to-disc ratio (CDR)computation, our results show that we are able to achieve moreaccurate optic cup segmentation and CDR measurements fromimages processed by SGRIF.

Index Terms- retinal image processing, segmentation,computer aided diagnosis

I. INTRODUCTION

Retinal fundus photographs have been widely used by clini-cians to diagnose and monitor many ocular diseases includingglaucoma, pathological myopia, age-related macular degen-eration, and diabetic retinopathy. Since manual assessmentof the images is tedious, expensive and subjective, computeraided diagnosis methods [1] have been developed to analysethe images automatically for optic disc segmentation [2]–[5],

Manuscript received March 19, 2018; revised May 12; accepted May 16,2018. This work was supported in part under grant Y80002RA01 by Cixiinstitute of biomedical engineering, Chinese Academy of Sciences, Chinaand Ningbo 3315 Innovation team grant Y61102DL03. Asterisk indicatescorresponding author.

*J. Cheng is with Cixi Institute of Biomedical Engineering, Chi-nese Academy of Sciences, China. (email: [email protected],[email protected])

Z. Li is with Institute for Infocomm Research, Agency for Science,Technology and Research, Singapore. (email: [email protected]). .

Z. Gu is with Cixi Institute of Biomedical Engineering, Chinese Academyof Sciences, and Shanghai University, China (email: [email protected]).

H. Fu and D. W. K. Wong are with Ocular Imaging (iMED) Departmentin Institute for Infocomm Research, Agency for Science, Technology and Re-search, Singapore (email: [email protected], [email protected]).

J. Liu is with Cixi Institute of Biomedical Engineering, Chinese Academyof Sciences, China (email: [email protected])

(a) (b)

(c) (d)

Fig. 1. Retinal images: (a)-(b) from eyes with cataract and (c)-(d) from eyeswithout cataract

optic cup segmentation or cup to disc ratio assessment [6]–[9], retinal vessel detection [10], [11], glaucoma detection[12], [13], diabetic retinopathy detection [14], [15], age-relatedmacular degeneration detection [16]–[19], and pathologicalmyopia detection [20]. Image quality is an essential factorfor proper development and validation of the algorithms [21].Very often, low quality images lead to poor performance ofautomatic analysis. Therefore, it is necessary to restore theimage for better analysis. There are many factors that mayaffect the quality of retinal images. In retinal imaging, anillumination light passes through the lens of human eye toreach the retina. Ideally, the light will be reflected back tothe fundus camera by the retina to form the retinal images.However, the human eye is not a perfect optical system andthe light received by the fundus camera is often attenuatedalong the path of the light. This can be serious when the lensof human eye is affected by diseases such as cataracts.

Cataract is caused by clouding of human-lens in the eye.Since a retinal image is captured through the lens of humaneye, a cataract will lead to attenuation and scattering ofthe light travelling through the human-lens, just as a cloudycamera lens reducing the quality of a photograph. The processof retina images by cataractous human-lens is referred to asclouding and the processing to remove the effect is referredto as declouding. Meanwhile, cataract accounts for 33% ofblindness worldwide [22] and its global prevalence in adultsover 50 years of age was about 47.8% in 2002 [23]. The

arX

iv:1

805.

0662

5v2

[cs

.CV

] 2

2 M

ay 2

018

gaojing
附注
摘要 - 视网膜眼底照片已被用于诊断许多眼部疾病,如青光眼,病理性近视,年龄相关性黄斑变性和糖尿病视网膜病变。随着计算机科学的发展,已经开发出计算机辅助诊断技术来自动处理和分析视网膜图像。分析中的挑战之一是视网膜图像的质量经常降低。例如,人类晶状体中的白内障会使视网膜图像衰减,就像阴影相机镜头一样,会降低照片的质量。它经常模糊视网膜图像中的细节,并在视网膜图像处理和分析任务中发布挑战。在本文中,我们将视网膜图像的降解近似为人 - 透镜衰减和散射的组合。然后提出一种新颖的结构保持引导视网膜图像滤波(SGRIF),以基于衰减和散射模型恢复图像。拟议的SGRIF包括全球结构转移步骤和全局边缘保持平滑步骤。我们的研究结果表明,所提出的SGRIF方法能够改善视网膜图像的对比度,通过直方图光度测量,直方图扩散和局部光度的变化来测量。此外,我们进一步探索了SGRIF对后续视网膜图像处理和分析任务的好处。在基于深度学习的视杯分割和基于稀疏学习的杯盘比(CDR)计算的两个应用中,我们的结果表明我们能够通过SGRIF处理的图像实现更准确的视杯分割和CDR测量。
gaojing
附注
图1。视网膜图像:(a) - (b)来自白内障的眼睛和(c) - (d)来自没有白内障
gaojing
附注
视网膜眼底照片已被临床医生广泛用于诊断和监测许多眼部疾病,包括青光眼,病理性近视,年龄相关性黄斑变性和糖尿病性视网膜病变。由于手动评估图像是繁琐,昂贵和主观的,因此开发了计算机辅助诊断方法[1]来自动分析图像以进行视盘分割[2] - [5],
gaojing
附注
视杯分割或CDR比率评估的眼睛[6] - [9],视网膜血管检测[10], [11],青光眼检测[12],[13],糖尿病视网膜病变检测[14],[15],年龄相关性黄斑变性检测[16] - [19],以及病理性近视检测[20]。图像质量是算法正确开发和验证的重要因素[21]。通常,低质量图像会导致自动分析性能不佳。因此,有必要恢复图像以便更好地分析。有许多因素可能影响视网膜图像的质量。在视网膜成像中,照明光通过人眼的透镜到达视网膜。理想地,光将通过视网膜反射回眼底照相机以形成视网膜图像。然而,人眼不是完美的光学系统,并且眼底照相机接收的光通常沿着光路衰减。当人眼的镜片受到诸如白内障的疾病影响时,这可能是严重的。
gaojing
附注
白内障是由眼睛中人类晶状体的混浊引起的。由于通过人眼镜头捕获视网膜图像,白内障将导致穿过人类镜头的光的衰减和散射,就像阴影相机镜头降低照片质量一样。通过白内障人体晶状体进行视网膜图像的过程被称为模糊,并且去除该效果的处理被称为去除。同时,白内障占全世界失明的33%[22],其在全球50岁以上成人患病率在2002年约为47.8%[23]。白内障的高患病率使其成为影响视网膜成像的不可忽视的因素。根据白内障晶状体中浑浊的严重程度和位置,视网膜图像可能在不同严重程度下降级。它经常模糊图像的细节,影响诊断过程并在视网膜图像处理和分析中发布挑战。图1显示了四个视网膜图像样本,其中第一行显示来自患有白内障的眼睛的两个图像,第二行显示来自没有任何白内障的眼睛的两个图像。我们可以看到,白内障会降低视网膜图像的质量并降低图像的动态范围。由于降解是由人类镜头散射的光引起的,我们将散射光称为镜头光。去除由透镜光引入的混浊效果将增加视网膜图像的对比度并校正失真。
Page 2: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 2

high prevalence of cataract makes it an innegligible factorthat affects retinal imaging. Depending on the severity andthe location of the clouding in the cataractous lens, the retinalimages may be degraded at different levels of severity. Itoften obscures the details of the images, affects the diagnosticprocess and posts challenges in retinal image processing andanalysis. Fig. 1 shows four samples of retinal images, wherethe first row shows two images from eyes with cataract andthe second row shows two images from eyes without anycataract. As we can see, the cataract degrades the retinalimages and reduces the dynamic range of the images. Sincethe degradation is caused by light scattered by the human-lens,we call the scattered light as lens-light. Removal of the cloudyeffect introduced by lens-light will increase the contrasts of theretinal image and correct the distortion.

Besides cataracts, many other factors affect the quality ofretinal images as well. Studies have shown that refractiveanomalies such as defocus or astigmatism often blur the image[24]. Optical aberration also blurs the retinal image, eventhough its contribution is typically smaller than that of defocusor astigmatism [25]. Studies have shown there is a largeincrement of ocular aberration with age [26]. Scattering innormal eyes is usually small. However, it is also known that thescattering increases with age. These different factors make thevisual appearance of the retinal images from different subjecteyes different.

In retinal image processing, the low image quality dueto the clouding often results in many challenges to learn agood representation of the images for structure detection orsegmentation, lesion detection and other analysis. In opticcup segmentation, the boundary between the optic cup andneuroretinal rim may have a wide range of gradients and mayhave been affected by cataractous lens or other factors. Inhaemorrhage detection or vessel segmentation, the intensity ofthe vessels may vary largely from one image to another and thevessel boundaries might also be obscured. This motivates us toremove the clouding effect due to the scattering of human-lensand increase the contrast of the retinal images.

In [27], the degradation due to cataracts is modelled as

Ic(p) = αLcrc(p)t(p) + Lc(1− t(p)), (1)

where α is a constant denoting the attenuation of retinalillumination due to the cataracts, c ∈ {r, g, b} denotes thered, green or blue channel of the image, rc(p) denotes theretinal reflectance function, Lc is the illumination of the funduscamera and t(p) describes the portion of the light that doesnot reach the camera.

The above model requires a precataract clear image toestimate α. However, such image is often unavailable and theillumination light may alse be different if the precataract imageis caputured under different conditions. On the other hand,the value of the constant α only affects the final output imagelinearly. One can apply a linear scale 1/α on the output imageif a precataract image is available to determine α. Therefore,we propose to use a slightly simpler model:

Ic(p) = Dc(p)t(p) + Lc(1− t(p)), (2)

where Dc(p) = Lcrc(p) denotes the image captured underideal human-lens. The model in equation (2) is the same asthe dehaze model in computer vision where the attenuation ofthe images by the haze or fog are modeled by air attenuationand scattering [28]. This model can be considered as a specialcase of (1) by letting α = 1, i.e., ignoring the attenuation α.By applying the model on the retinal image, the problem ofremoving the clouding effect due to lens scattering becomesa common dehaze problem in computer vision.

In computer vision, many methods [28]–[33] have beenproposed to solve the dehaze problem in equation (2). In [28],Markov random field was used to enhance the local contrast ofthe restored image, however, it often produces over-saturatedimages. In [29], Fattal proposed a model that accounts for bothsurface shading and scene transmission, but it does not performwell in heavy haze. In [30], a novel dark channel prior wasused. But this assumption is not always valid for retinal imageswhich do not have many shadows or complex structures asnatural scene images. Guided image filtering (GIF) [31] isa promising technology proposed recently for single imagehaze removal. It has a limitation that it often oversmoothes theregions close to flat and does not preserve the fine structurewhich might be important in retinal image processing oranalysis. To overcome the limitation, we propose a methodto preserve the structure in the original images. Motivated byGIF, we propose a structure-preserving guided retinal imagefiltering (SGRIF), which is composed of a global structuretransfer filter and a global edge-preserving smoothing filter.Then we apply SGRIF on the retina images for subsequentanalysis. Different from most of work that ends with imagequality evaluation, we further investigate how the processbenefits subsequent automatic analysis from the processedimages. Two different applications including deep learningbased optic disc segmentation and sparse-learning based cup-to-disc ratio (CDR) computation are conducted to show theadvantage of our method.

Contribution: Our main contributions are summarized asfollows.

1) We propose a structure preserving guided retinal imagefiltering for declouding of the retinal images.

2) Our experiments show that the proposed SGRIF algo-rithm is able to improve the contrast of the image andmaintain the edges for further analysis.

3) Our method benefits subsequent analysis as well. Inapplications of deep learning based optic disc segmen-tation, the pre-processing using the proposed SGRIFimproves the segmentation results. In application ofsparse-learning based CDR computation, we also showthat the proposed method improves the accuracy of CDRcomputation.

The remaining sections of the paper are organized asfollows. Section II introduces our proposed method to removethe clouding effect caused by lens-light. The method includesa step of global structure transferring and a step of globaledge-preserving smoothing. Experimental results are given inSection III to show the effectiveness of the proposed method toimprove the contrast of the retinal images and its application

gaojing
附注
除白内障外,许多其他因素也会影响视网膜图像的质量。研究表明,散焦或散光等屈光异常常使图像模糊[24]。光学像差也使视网膜图像模糊,即使其贡献通常小于散焦或散光[25]。研究表明,随着年龄的增长,眼睛畸变有很大的增加[26]。正常眼睛的散射通常很小。然而,还已知散射随着年龄而增加。这些不同的因素使得来自不同主体眼睛的视网膜图像的视觉外观不同。
gaojing
附注
在视网膜图像处理中,由于混浊导致的低图像质量经常导致许多挑战以学习用于结构检测或分割,病变检测和其他分析的图像的良好表示。在视杯分割中,视杯和神经视网膜边缘之间的边界可能具有宽范围的梯度,并且可能受到白内障晶状体或其他因素的影响。在出血检测或血管分割中,血管的强度可能在很大程度上从一个图像到另一个图像变化,并且血管边界也可能变得模糊。这促使我们消除由于人 - 晶状体散射造成的混浊效应并增加视网膜图像的对比度。
gaojing
附注
在[27]中,白内障引起的退化被建模为
gaojing
附注
其中α是表示由白内障引起的视网膜照射衰减的常数,表示图像的红色,绿色或蓝色通道,表示视网膜反射功能,表示眼底照相机的照明,描述部分没有到达相机的光线。
gaojing
附注
上述模型需要预先显示清晰图像来估计α。然而,如果在不同条件下捕获预先图像,则这种图像通常是不可用的并且照明光也可能是不同的。另一方面,常数α的值仅线性地影响最终输出图像。如果可以使用预表示图像来确定α,则可以在输出图像上应用线性标尺。因此,我们建议使用稍微简单的模型
gaojing
附注
其中表示在理想人体镜头下拍摄的图像。等式(2)中的模型与计算机视觉中的去雾模型相同,其中雾霾或雾的图像衰减通过空气衰减和散射来模拟[28]。该模型可以被认为是(1)的特殊情况,即让,即忽略衰减α。通过在视网膜图像上应用该模型,消除由于透镜散射引起的混浊效应的问题成为计算机视觉中常见的除雾问题。
gaojing
附注
在计算机视觉中,已经提出了许多方法[28] - [33]来解决方程(2)中的去雾问题。在[28]中,马尔可夫随机场用于增强恢复图像的局部对比度,但是,它经常产生过饱和的图像。在[29]中,Fattal提出了一个模型,既考虑了表面阴影和场景传输,但在重雾中表现不佳。在[30]中,使用了一种新颖的暗通道先验。但是,对于没有像自然场景图像那样具有许多阴影或复杂结构的视网膜图像,这种假设并不总是有效的。引导图像滤波(GIF)[31]是最近提出的用于单个图像雾度去除的有前景的技术。它有一个局限性,它经常使接近fl的区域过度平滑,并且不保留在视网膜图像处理或分析中可能很重要的细微结构。为了克服这个限制,我们提出了一种保留原始图像结构的方法。在GIF的推动下,我们提出了一种结构保持引导视网膜图像滤波(SGRIF),它由全局结构传递滤波器和全局边缘保持平滑滤波器组成。然后我们将SGRIF应用于视网膜图像以供后续分析。与大多数以图像质量评估结束的工作不同,我们进一步研究了该过程如何从处理后的图像中获得后续的自动分析。两种不同的应用包括基于深度学习的视盘分割和基于稀疏学习的杯盘比(CDR)计算,以显示我们的方法的优点。
gaojing
高亮
gaojing
附注
1)我们提出了一种保留引导视网膜图像滤波的结构,用于视网膜图像的去除。
gaojing
附注
2)我们的实验表明,所提出的SGRIF算法能够改善图像的对比度并保持边缘以供进一步分析。
gaojing
附注
3)我们的方法也有利于后续分析。在基于深度学习的视盘分割的应用中,使用所提出的SGRIF的预处理改善了分割结果。在基于稀疏学习的CDR计算的应用中,我们还表明所提出的方法提高了CDR计算的准确性。
gaojing
附注
本文的其余部分安排如下。第二节介绍了我们提出的消除透镜光造成的混浊效应的方法。该方法包括全局结构转移步骤和全局边缘保持平滑步骤。实验结果在第III节中给出,以显示所提出的方法改善视网膜图像的对比度的有效性及其在视杯分割和CDR计算中的应用。最后一节提供了结论。
Page 3: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 3

in optic cup segmentation and CDR computation. Conclusionsare provided in the last section.

II. STRUCTURE-PRESERVING GUIDED RETINAL IMAGEFILTERING

In GIF [31], a guidance image G which could be identicalto the input image I is used to guide the process. The outputimage O is computed by a linear transform of the guidanceimage G in a window Wp centered at the pixel p, i.e.,

Oi = apGi + bp,∀i ∈Wp, (3)

where i indicates pixel index.The linear transform coefficients ap, bp are computed by

minimizing the following objective function

E = (Oi − Ii)2 + εa2p, (4)

where ε is a regularization parameter.Its solution is given by:

ap =

1|Wp|

∑i∈Wp

GiIi − µpIpσ2p + ε

, (5)

bp = Ip − apµp, (6)

where Ip denotes the mean of I in Wp, |Wp| is the cardinalityof Wp, µP and σ2

p denote the mean and variance of G in Wp

respectively.Since each pixel i is covered by many overlapping window

Wp, the output of equation (5) from different windows is notidentical. An averaging strategy is used in GIF [31], i.e.,

Oi = apGi + bp, (7)

where ap and bp are the mean values of ap′ and bp′ in Wp:

ap =1

|Wp|∑p′∈Wp

ap′ , (8)

bp =1

|Wp|∑p′∈Wp

bp′ . (9)

It can be seen that a pixel from a high variance area willretain its values while a pixel from a flat area will be smoothedby its nearby pixels. Therefore, the above averaging oftensmoothes away some fine structure in regions close to flat.This is not optimal for retinal image processing and analysis.For example, the boundary between the optic cup and theneuroretinal rim is often weak. The averaging in this regionmight smooth away this boundary and make the optic cupsegmentation more challenging.

To address this problem, we propose a novel structure-preserving guided retinal image filter (SGRIF) to removethe clouding effect caused by light scattered by human-lens.Inspired by the GIF [31], we aim to transfer the structure ofthe guidance image to the input image to preserve the edgesand smooth the transferred image. The proposed SGRIF iscomposed of a global structure transfer filter to transfer thestructure in the retinal image and a global edge-preservingsmoothing filter to smooth the transferred retinal image. In ourmethod, we use a guidance vector field V = (V h, V v). The

inputs of the proposed SGRIF are a retinal image to be filteredand a guidance vector field. A term between the gradients ofthe output image O and the guidance vector field V is definedto penalize the original term in GIF, i.e.,∑

i

‖∇Oi − Vi‖2, (10)

where ∇ denotes the operator to compute the gradients.Combining this with first item in (4), we have the objectivefunction of the proposed global structure transfer filter:

λ∑i

(Oi − Ii)2 + ‖∇Oi − Vi‖2, (11)

where λ is a parameter controlling the trade-off between thetwo terms. Omitting the subscript, the above cost function canbe rewritten as:

λ(O − I)T (O − I) + (DxO − V h)T (DxO − V h)

+ (DyO − V v)T (DyO − V v), (12)

where Dx and Dy denote discrete differentiation operators.The output O is then obtained by solving the following linearequation:

(λA+DTxDx +DT

y Dy)O = λI +DTx V

h +DTy V

v, (13)

where A is an identity matrix.We solve (13) using the fast separating method in [34].

The output image O∗ based on (13) sometimes needs to besmoothed. Fig. 2 shows an example. As we can see, therecovered image O∗ has some visual artifacts if it is notsmoothed. To achieve this, the output image is decomposedinto two layers via an edge-preserving smoothing filter. Thefilter [35]–[37] is formulated as

minφ

∑i

[(φi −O∗i )2 + γ((∂φi∂x )2

|V hi |θ + ε+

(∂φi∂y )2

|V vi |θ + ε)], (14)

where γ, θ, ε are empirically set as 2048, 13/8 and 1/64 inall experiments in this paper. We determine the thresholds bysearching from a reasonable range based on physical meaningand experience. For example γ is set to a large value to makesure that the first term in equation 14 will not dominate theresults. Then we conducted tests using some γ values anddetermine its value. Our results show that the small change ofthe parameters do not affect much the results.

We can rewrite (14) as

(φ−O∗)T (φ−O∗)+γ(φTDTxBxDxφ+φTDT

y ByDyφ), (15)

where Bx = diag{ 1|V hi |θ+ε

}, By = diag{ 1|V vi |θ+ε

}.Setting the derivative of (15) to be zero, the vector φ, that

minimizing the above cost function, is computed by solvingthe following equation:

(A+ γ(DTxBxDx +DT

y ByDy))φ = O∗. (16)

Similar to that in (13), the problem in (16) can be solvedby the fast separate method in [34] as well.

To apply the above models to retinal images, we needto estimate the lens-light Lc, c ∈ {r, g, b}. In this paper,we estimate Lr, Lg , and Lb using the method in [38]. The

gaojing
附注
保持结构的导向视网膜图像
Page 4: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 4

(a) (b) (c)

Fig. 2. Effect of the global edge-preserving smoothing filter. (a) Original image (b) Output image O∗ without edge-preserving smoothing filter (c) Outputimage after edge-preserving smoothing filter

proposed method uses the idea of minimal color channeland simplified dark channel. The simplified dark channel isdecomposed into a base layer and a detail layer via theproposed method, and the former is then used to determinethe transmission map. The simplified dark channels of thenormalized degraded and ideal images are computed as Ic/Lcand Dc/Lc. Define Imin(p) and Dmin(p) as

Imin(p) = min{Ir(p)Lr

,Ig(p)

Lg,Ib(p)

Lb}, (17)

Dmin(p) = min{Dr(p)

Lr,Dg(p)

Lg,Db(p)

Lb}. (18)

It is noted that we do not consider the difference amongthe RGB channels in this paper, though some earlier work[39] shows that blue channel may have more noise. Since thetransmission map t is independent to the color channels, wehave

Imin(p) = (1− t(p)) + Dmin(p)t(p). (19)

Let Wκ(p) be a κ×κ window centered at pixel p. The sim-plified dark channels of the normalized images are computedas:

JDd (p) = minp′∈Wκ(p)

{Dmin(p′)}, (20)

J Id (p) = minp′∈Wκ(p)

{Imin(p′)}. (21)

Since t(p) is usually consistent within W (p), we have

J Id (p) = (1− t(p)) + JDd (p)t(p). (22)

The guidance vector field V = (V h, V v) is computed as

V h(m,n) = Imin(m,n+ 1)− Imin(m,n), (23)

V v(m,n) = Imin(m+ 1, n)− Imin(m,n). (24)

An example of the gradient vector field image computed fromFig. 2(a) is shown in Fig. 3. Combining the gradient vectorfield with equation (11), we obtain the output O∗.

(a) (b)

Fig. 3. An example of the gradient vector field image. (a) magnitude of V h,(b) magnitude of V v

We further smooth the first item φ(p) = 1− t(p) using theedge preserving smoothing filter in (14) and obtain φ∗(p).

The transmission map t(p) is then computed as:

t∗(p) = 1− φ∗(p). (25)

The underlying image is computed as:

Dc(p) =Ic(p)− Lct∗(p)

+ Lc. (26)

In our model, we ignore the attenuation α. In the caseswhere the attenuation α cannot be ignored, we can still solvethe problem in the similar way except that we need to estimateα using a precataract image and compute the final outputimage as 1

αDc(p). In this paper, we simply restore the imagebased on equation (26).

III. EXPERIMENTAL RESULTS

A. Data Set

We conduct experiments using the 650 images from theORIGA data set [40][41]. In the 650 images, there are 203images from eyes with cataracts and 447 images from eyeswithout any cataract. We mainly apply SGRIF on the regionaround the optic disc and evaluate how it affects subsequent

Page 5: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 5

analysis on optic disc. In this paper, we use the disc from theoriginal image and the disc boundaries are kept the same forall three images. This is to prevent the error propagation fromoptic disc segmentation to optic cup segmentation.

B. Evaluation Metrics

To evaluate the performance of the proposed method, wefirst compute how it affects the contrast of the optic disc image.Two evaluation metrics, namely the histogram flatness measure(HFM) and the histogram spread (HS) have been proposed toevaluate the contrast of images previously [42]:

HFM =(∏ni=1 xi)

1n

1n

∑ni=1 xi

, (27)

where xi is the histogram count for the ith histogram bin andn is the total number of histogram bins.

HS =(3rdquartile− 1stquartile) of histogram

(maximum-minimum) of the pixel value range. (28)

We also compute the mean variability of the local luminosity(VLL) [43] throughout the optic disc. Given an image I , wedivide it into N ×N blocks Bi,j , i, j = 1, · · · , N with equalsizes. We first compute the mean of block Bi,j as µ(i, j).

VLL is computed as

V LL =1

N

√√√√ 1

I2

N∑i=1

N∑j=1

(µ(i, j)− I)2, (29)

where I stands for the mean intensity of the entire image.For all the above metrics, a high value indicates a better

result.

C. Results

Table I summarizes the results, from which we can see thatthe proposed method improves the HFM, HS and VLL by5.9%, 4.3% and 134.8%, respectively compared with originalimages. GIF improves VLL, but it does not increase HFM andHS. This is because GIF oversmoothes some region close toflat and reduces the dynamic range of the histograms. Fig. 4shows results from five sample images of optic disc. As wecan see, the proposed SGRIF enhances the contrast betweenthe optic cup and the neuroretinal rim while the improvementby GIF is less clear. Visually, it is difficult to tell if GIF hasoversmoothed some regions but we will show the differencesfrom subsequent analysis below.

D. Application

In order to show the benefits of the declouding, we fur-ther conduct experiments to examine how the decloudingprocessing affects the analysing tasks. Two slightly differentapplications are used as examples: 1) deep learning based opticcup segmentation; 2) sparse learning based cup to disc ratiomeasurement.

1) Deep Learning based Optic Cup Segmentation: In thefirst example, we examine how the decloud affects the opticcup segmentation. Since deep learning [44] has shown tobe promising in segmentation, we use deep learning as thebaseline approach and the U-Net architecture [45] is adoptedin this paper. U-Net is a fully convolution neural network forthe biomedical image segmentation. In our implementation, weuse a simplified U-Net as the original U-net requires muchmore number of parameters to be trained. Our network isillustrated in Fig. 5. Similar to original U-Net, our networkcontains an encoder and a decoder. For each layer of theencoder, it just adopts a convolution layer with stride 2,replacing the original two convolution layers and one poolinglayer. For each layer of the decoder, it takes two outputsas its input: i) the output of last layer in the encoder; ii)the corresponding layer in the encoder with the same size.Then two middle-layer feature maps are concatenated andtransferred to the next deconvolution layer as its input. Finally,our simplified U-Net outputs a prediction grayscale map whichranges from 0 to 255. We calculate the mean square errorbetween prediction map and ground truth as loss function. Insimple U-Net, we use a mini-batch Stochastic Gradient Decent(SGD) algorithm to optimize the loss function, specifically,Adagrad based SGD [46]. The adopted learning rate is 0.001and batch size is 32 under the Tensorflow framework based onUbuntu 16.04 system. We use a momentum of 0.9. All imagesare resized to 384×384 for training and testing. The obtainedprobability map is resized back to original size to obtain thesegmentation results.

In our experiments, the 650 images have been dividedrandomly into set A of 325 training images and set B of325 testing images. In order to evaluate the performance ofthe proposed method for images from eyes with and withoutcataract, the 325 images in set B is divided into a subset of 113images with cataracts and 212 images without any cataract.We then compare the output cup segmentation results in thethree different scenarios: 1) Original: the original training andtesting images are used for training and testing the U-Netmodel; 2) GIF: all the training and testing images are firstfiltered by GIF before U-Net model training and testing 3)SGRIF: all the training and testing images are first filtered bySGRIF before U-Net model training and testing.

To compare the segmentation results, we compute the com-monly used overlapping error E as the evaluation metric.

E = 1− Area(S ∩G)

Area(S ∪G), (30)

where S and G denote the segmented and the manual “groundtruth” optic cup respectively. Table II summarizes the resultsin the three different scenarios. As we can see, the proposedSGRIF reduces the overlapping error by 7.9% relatively from25.3% without filtering image to 23.3% with SGRIF. Thestatistical t-test indicates that the reduction is significant withp < 0.001. However GIF reduces the accuracy in optic cupsegmentation. To visualize the segmentation results, we alsodraw the boundaries of segmented optic cup by the deeplearning models. Fig. 6 shows five examples where the firsttwo are from eyes without cataract and the last three are

Page 6: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 6

Fig. 4. Retinal images of optic disc: first row the original disc images; second row the images processed by GIF and last row the images processed byproposed filter.

TABLE IPERFORMANCE BY VARIOUS METHODS.

Histogram Flatness Measure Histogram Spread Variability of Local LuminosityAll Cataract No Cataract All Cataract No Cataract All Cataract No Cataract

Original 0.5786 0.5295 0.6009 0.2856 0.2692 0.2930 0.0517 0.0363 0.0537GIF [31] 0.4779 0.4324 0.4986 0.2505 0.2329 0.2584 0.1014 0.0999 0.1021

Proposed 0.6129 0.5817 0.6271 0.2980 0.2806 0.3059 0.1214 0.1076 0.1277

Fig. 5. Simple U-Net architecture:

from eyes with cataract. From the results, we can see thatSGRIF improves the optic cup segmentation in both imagesfrom eyes with and without cataracts. SGRIF reduces thesegmentation error as it improves the boundary at tempral sideof the disc. GIF is not very helpful as it often smoothes awaythe boundaries in areaes close to flat. This often happens in thetemporal side of the retinal image where the gradient change

TABLE IIOVERLAPPING ERROR IN OPTIC CUP SEGMENTATION

All Cataract No CataractOriginal 25.3 24.4 25.8GIF [31] 25.6 24.8 26.0

Proposed 23.3 22.8 23.5

is mild. The above observation is inline with our discussion inSection II that GIF might smooth away the weak boundariesand make the optic cup segmentation more challenging.

Visually, it is difficult to tell if GIF has oversmoothed someregions from the images directly. In this paper, we show theintensity profiles to highlight the effect. Fig. 7 shows twoexample where the intensity profiles of horitonal lines (inwhite) are plotted. The intensity profiles from original image,the image processed by GIF and the image processed by theproposed SGRIF method are shown in red, blue, and green,respectively. From the second row, we can see that both GIFand SGRIF improve the contrast near the vessels and SGRIFperforms slightly better. From the third row, we can hardly

Page 7: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 7

(a) (b) (c) (d) (e)

Fig. 6. Effect of filtering on optic cup segmentation: (a) original disc (b) manual ground truth cup boundary (c) segmentation based on original disc image(d) segmentation based on GIF processed disc image (e) segmentation based on proposed SGRIF processed disc image

tell the location of optic cup boudnary from the profiles in redand the blue while we observe a more obvious gradient changein the area pointed by the arrow in the profile in green. Thesmall difference in the profiles can be critical for subsequentanalysis tasks as supported by the optic cup boundary dectionresults.

2) Sparse Learning based CDR Computation: In the sec-ond test, we explore how the decloud affects a direct verticalcup to disc ratio (CDR) assessment using sparse learning. Pre-viously, we have proposed a sparse dissimilarity-constrained

coding (SDC) [8] algorithm to reconstruct each testing discimage y with a set of n training or reference disc imagesX = [x1, x2, · · · , xn with known CDRs r = [r1, r2, · · · , rn].The method is essentially a sparse learning based approach.SDC aims to find a solution w to approximate y by Xw basedon the following objective function:

arg minw‖y −Xw‖2 + λ1 · ‖d ◦ w‖2 + λ2 · ‖w‖1, (31)

where d = [d1, d2, · · · , dn] denote the similarity cost betweeny and X , ◦ denotes dot product. The CDR of testing image y

Page 8: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 8

50 100 150 200 250

Pixels

100

150

200

Inte

nsity

OriginalGIFProposed

Vessel

Optic Cup Boundary

50 100 150 200 250 300

Pixels

100

150

200

250

Inte

nsity

OriginalGIFProposed

Vessel Optic Cup Boundary

Fig. 7. Intensity profiles across the dash lines of the first two samples in Fig. 6: first two rows show the intensity images and the profiles of the first sampleand last two rows show those for the second sample. The intensity profiles from the original image, the image processed by GIF and the image processedby the proposed method are shown in red dash, blue dotted and green solid lines, respectively. From left to right are the original image, image processed byGIF and image processed by the proposed method.

Page 9: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 9

TABLE IIIACCURACY OF CDR COMPUTATION

All Cataract No CataractOriginal 0.0657 0.0626 0.0674GIF [31] 0.0665 0.0625 0.0686

Proposed 0.0639 0.0603 0.0658Inter-Observer Error 0.0767 0.0761 0.0771

is then estimated as

r =1

1TwwrTw, (32)

where 1 is a vector of 1s with length n. More details of theSDC method are kept the same as that in [8] and can be foundin the paper.

To justify the benefit of SGRIF for CDR computation, weconduct CDR computation based on (31). Similar to that inoptic cup segmentation, we conduct tests in three scenarios:1) original reference and testing images 2) all reference andtesting images are processed by GIF; 3) all reference andtesting images are processed by SGRIF. Table III shows theCDR error between automatically computed CDRs and themanually measured CDRs in the testing images. The CDRerrors for retinal images of eyes with and without cataract arecomputed separately. From the results, we obtain a relativereduction of 3.7% for retinal images of eyes with cataractand a relative reduction of 2.4% for retinal images of eyeswithout cataract. It is worth to note that our method alsoimproves the vertical CDR estimation even though it mainlyimproves the edges at the temporal side of the optic disc. Thisis not surprising as the temporal side is an important partin the reconstruction. In additional, we have also invited anophthalmologist to manually measure a second set of CDRs[8] and the inter-observer error is given in Table III as well.

E. Performance on other regions

Besides the region around the optic disc, we have alsoapplied our method to other regions including those withdifferent lesions. Fig. 8 shows some results. From the results,we can see that both GIF and SGRIF improve the contrast ofthe original images. Howerver, we do not have a quantitativemeasurement on how GIF or SGRIF improves subsequentanalysis in lesion area.

IV. CONCLUSIONS

Image quality is an important fact that affecting the per-formance of retinal image analysis. In this paper, we proposestructure-preserving guided retinal image filtering to removeartifacts caused by a cloudy human-lens. Our results showthat the proposed method is able to improve the contrastwithin the retinal images, which is reflected by histogramflatness measurement, histogram spread and variability of localluminosity. We further validate the algorithm with subsequentoptic cup segmentation using deep learning and cup-to-discratio measurement using sparse learning. Both the experimentsin optic cup segmentation and cup-to-disc ratio computationshow that the proposed method is beneficial for the subsequent

tasks. In this paper, we mainly apply the algorithm in theregion around the disc for optic disc analysis. We do not applythat for optic disc segmentation as the optic disc boundary isusually much stronger and will not be smoothed away in GIF.Therefore, our method does not improve the results. Since ourmethod improves the CDR measurement, it could potentiallybe applied for glaucoma detection. Our results also suggestthat the method improve the blood vessel qualitatively. Infuture work, we will also evaluate how the processing affectother analysis such as vessel segmentation and lesion detectionquantitatively.

REFERENCES

[1] M. D. Abramoff, M. K. Garvin, and M. Sonka, “Retinal imaging andimage analysis,” IEEE Trans. Med. Imag., vol. 3, pp. 169–208, 2010.

[2] H. Li and O. Chutatape, “Boundary detection of optic disc by a modifiedasm method,” Pattern Recognition, vol. 36, pp. 2093–2104, 2003.

[3] A. Aquino, M. Gegundez-Arias, and D. Marin, “Detecting the optic discboundary in digital fundus images using morphological, edge detection,and feature extraction techniques,” IEEE Trans. Med. Imag., vol. 29,pp. 1860–1869, 2010.

[4] C. Muramatsu, T. Nakagawa, A. Sawada, Y. Hatanaka, T. Hara, T. Ya-mamoto, and H. Fujita, “Automated segmentation of optic disc regionon retinal fundus photographs: Comparison of contour modeling andpixel classification methods,” Computer Methods and Programs inBiomedicine, vol. 101, pp. 23–32, 2011.

[5] J. Cheng, J. Liu, D. W. K. Wong, F. Yin, C. Cheung, M. Baskaran,T. Aung, and T. Y. Wong, “Automatic optic disc segmentation withperipapillary atrophy elimination,” Int. Conf. of IEEE Eng. in Med. andBio. Soc., pp. 6624–6627, 2011.

[6] M. D. Abramoff, W. L. M. Alward, E. C. Greenlee, L. Shuba, C. Y. Kim,J. H. Fingert, and Y. H. Kwon, “Automated segmentation of theoptic discfrom stereo color photographs using physiologically plausible features,”Invest. Ophthalmol. Vis. Sci., vol. 48, pp. 1665–1673, 2007.

[7] J. Cheng, J. Liu, Y. Xu, F. Yin, D. W. K. Wong, N. M. Tan, D. Tao,C. Y. Cheng, T. Aung, and T. Y. Wong, “Superpixel classification basedoptic disc and optic cup segmentation for glaucoma screening,” IEEETransactions on Medical Imaging, vol. 32, no. 6, pp. 1019–1032, 2013.

[8] J. Cheng, F. Yin, D. W. K. Wong, D. Tao, and J. Liu, “Sparsedissimilarity-constrained coding for glaucoma screening,” IEEE Trans.on Biom. Eng., vol. 62, no. 5, pp. 1395–1403, 2015.

[9] H. Fu, J. Cheng, Y. Xu, D. W. K. Wong, J. Liu, and X. Cao, “Jointoptic disc and cup segmentation based on multi-label deep network andpolar transformation,” IEEE Transactions on Medical Imaging, vol. PP,no. 99, pp. 1–1, 2018.

[10] J. Staal, M. D. Abramoff, M. Niemeijer, M. A. Viergever, and B. vanGinneken, “Ridge based vessel segmentation in color images of theretina,” IEEE Trans. on Med. Imaging, vol. 23, pp. 501–509, 2004.

[11] Y. Zhao, Y. Zhao, Y. Zheng, Y. Liu, Y. Zhao, L. Luo, S. Yang, T. Na,Y. Wang, and J. Liu, “Automatic 2d/3d vessel enhancement in multiplemodality images using a weighted symmetry filter,” IEEE Transactionson Medical Imaging, vol. PP, no. 99, pp. 1–1, 2017.

[12] A. Li, J. Cheng, D. W. K. Wong, and J. Liu, “Integrating holistic andlocal deep features for glaucoma classification,” IEEE Engineering inMedicine and Biology Society (EMBC), pp. 1328–1331, 2016.

[13] M. Mishra, M. K. Nath, and S. Dandapat, “Glaucoma detectionfrom color fundus images,” International Journal of Computer &Communication Technology, vol. 2, pp. 7–10, 2011.

[14] M. D. Abramoff, M. Niemeijer, and S. R. Russell, “Automated detectionof diabetic retinopathy: barriers to translation into clinical practice,”Expert Review of Medical Devices, vol. 7(2), pp. 287296, 2010.

[15] V. Gulshan, L. Peng, M. Coram M., and et al, “Development and vali-dation of a deep learning algorithm for detection of diabetic retinopathyin retinal fundus photographs,” JAMA, vol. 316, no. 22, pp. 2402–2410,2016.

[16] C. Kse, U. evik, and O. Genaliolu, “Automatic segmentation of age-related macular degeneration in retinal fundus images,” Computers inBiology and Medicine, vol. 38, no. 5, pp. 611 – 619, 2008.

[17] Z. Liang, D. W. K. Wong, J. Liu, K. L. Chan, and T. Y. Wong, “Towardsautomatic detection of age-related macular degeneration in retinal fundusimages,” International Conference of the IEEE Engineering in Medicineand Biology, pp. 4100–4103, 2010.

Page 10: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 10

(a) (b) (c)

Fig. 8. Performance of the method in region with lesions: (a) Original (b) image processed by GIF (c) image processed by SGRIF (best viewed in softcopy)

Page 11: Jun Cheng*, Zhengguo Li, Zaiwang Gu, Huazhu Fu, Damon …static.tongtianta.site/paper_pdf/db842e22-0753-11e9-bab7-00163e08bb86.pdfIEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. X, NO. X, 2018 11

[18] A. Gven, “Automatic detection of age-related macular degenerationpathologies in retinal fundus images,” Computer Methods in Biome-chanics and Biomedical Engineering, vol. 16, no. 4, pp. 425–434, 2013.

[19] C. Agurto, E. S. Barriga, V. Murray, S. Nemeth, R. Crammer, W. Bau-man, G. Zamora, M. S. Pattichis, and P. Soliz, “Automatic detectionof diabetic retinopathy and age-related macular degeneration in digitalfundus images,” Investigative Ophthalmology Visual Science, vol. 52,no. 8, pp. 5862–5871, 2011.

[20] J. Liu, D.W.K. Wong, J. H. Lim, N. M. Tan, Z. Zhang, H. Li, F. Yin,B. H. Lee, S. M. Saw, L. Tong, and T. Y. Wong, “Detection ofpathological myopia by pamela with texture-based features through ansvm approach,” Journal of Healthcare Engineering, vol. 1, pp. 1–11,2010.

[21] E. Trucco, A. Ruggeri, T. Karnowski, L. Giancardo, E. Chaum, J. P. Hub-schman, B. Al-Diri B, C. Y. Cheung, D. Wong, M. Abramoff, G. Lim,D. Kumar, P. Burlina, N. M. Bressler, H. F. Jelinek, F. Meriaudeau,G. Quellec, T. Macgillivray, and B. Dhillon, “Validating retinal fundusimage analysis algorithms: issues and a proposal,” Invest OphthalmolVis Sci., vol. 54(5), pp. 3546–3559, 2013.

[22] D. Pascolini and S. P. Mariotti, “Global estimates of visual impairment:2010,” British Journal of Ophthalmology, vol. 96, no. 5, pp. 614–618,2012.

[23] S. Resnikoff, D. Pascolini, D. Etya’ale, I. Kocur, R. Pararajasegaram,G. P. Pokharel, and S. Mariotti, “Global data on visual impairment inthe year 2002,” Bulletin of the World Health Organization, vol. 82, pp.844–51, 12 2004.

[24] S. Marcos, “Image quality of the human eye,” International Ophthal-mology Clinics, vol. 43(2), pp. 43–62, 2003.

[25] J. Liang and D. R. Williams, “Aberrations and retinal image qualityof the normal human eye,” J. Opt. Soc. Am. A, vol. 14, no. 11, pp.2873–2883, Nov 1997.

[26] P. Artal, M. Ferro, I. Miranda, and R. Navarro, “Effects of aging inretinal image quality,” J. Opt. Soc. Am. A., vol. 10, no. 7, pp. 1656–1662, 1993.

[27] E. Peli and T. Peli, “Restoration of retinal images obtained throughcataracts,” IEEE Transactions on Medical Imaging, vol. 8, no. 4, pp.401–406, 1989.

[28] R. T. Tan, “Visibility in bad weather from a single image,” IEEEConference on Computer Vision and Pattern Recognition, pp. 1–8, 2008.

[29] R. Fattal, “Single image dehazing,” ACM Trans. Graph., vol. 27, no. 3,pp. 1–9, 2008.

[30] K. He, J. Sun, and X. Tang, “Single image haze removal usingdark channel prior,” IEEE Trans. On Pattern Analysis and MachineIntelligence, vol. 33, no. 12, pp. 2341–2353, 2011.

[31] K. He, J. Sun, and X. Tang, “Guided image filtering,” IEEE Trans.On Pattern Analysis and Machine Intelligence, vol. 35, no. 6, pp. 1397–1409, 2013.

[32] Z. Li, J. Zheng, Z. Zhu, W. Yao, and S. Wu, “Weighted guided imagefiltering,” IEEE Transactions on Image Processing, vol. 24, no. 1, pp.120–129, 2015.

[33] Z. Li and J. Zheng, “Edge-preserving decomposition-based single imagehaze removal,” IEEE Transactions on Image Processing, vol. 24, no.12, pp. 5432–5441, 2015.

[34] D. Min, S. Choi, J. Lu, B. Ham, K. Sohn, and M. Do, “Fast globalimage smoothing based on weighted least squares,” IEEE trans. onimage processing, vol. 23, pp. 5638–5653, 10 2014.

[35] Z. G. Li, J. H. Zheng, and S. Rahardja, “Detail-enhanced exposurefusion,” IEEE Transactions on Image Processing, vol. 21, no. 11, pp.4672–4676, 2012.

[36] Z. Farbman, R. Fattal, D. Lischinski, and R. Szeliski, “Edge-preservingdecompositions for multi-scale tone and detail manipulation,” ACMTrans. Graph., vol. 27, no. 3, pp. 67:1–67:10, Aug. 2008.

[37] Z. Li and J. Zheng, “Single image de-hazing using globally guidedimage filtering,” IEEE Transactions on Image Processing, vol. 27, no.1, pp. 442–450, Jan 2018.

[38] J.-H. Kim, W.-D. Jang, J.-Y. Sim, and C.-S. Kim, “Optimized contrastenhancement for real-time image and video dehazing,” Journal of VisualCommunication and Image Representation, vol. 24, no. 3, pp. 410 – 425,2013.

[39] A. G. Marrugo, M. S. Millan, M. Sorel, and F. Sroubek, “Retinal imagerestoration by means of blind deconvolution,” Journal of BiomedicalOptics, vol. 16(11), pp. 116016, 2011.

[40] Z. Zhang, F. Yin, J. Liu, W. K. Wong, N. M. Tan, B. H. Lee, J. Cheng,and T. Y. Wong, “Origa-light: An online retinal fundus image databasefor glaucoma analysis and research,” Int. Conf. of IEEE Eng. in Med.and Bio. Soc., pp. 3065–3068, 2010.

[41] J. Cheng, Z. Zhang, D. Tao, D. W. K. Wong, J. Liu, M. Baskaran,T. Aung, and T. Y. Wong, “Similarity regularized sparse group lasso forcup to disc ratio computation,” Biomed. Opt. Express, vol. 8, no. 8, pp.3763–3777, Aug 2017.

[42] A. K. Tripathi, S. Mukhopadhyay, and A. K. Dhara, “Performancemetrics for image contrast,” Int. Conf. on Image Information Processing,pp. 1–4, 2011.

[43] M. Foracchia, E. Grisan, and A. Ruggeri, “Luminosity and contrastnormalization in retinal images,” Medical Image Analysis, vol. 9, no. 3,pp. 179–190, 2005.

[44] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol.521(7553), pp. 436–444, 2015.

[45] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networksfor biomedical image segmentation,” Medical Image Computing andComputer-Assisted Intervention Part III, pp. 234–241, 2015.

[46] J. Duch, E. Hazan, and Y. Singer, “Adaptive subgradient methods foronline learning and stochastic optimization,” J. Mach. Learn. Res., vol.12, pp. 2121–2159, 2011.