recurrent residual convolutional neural network based on u...

12
AbstractDeep learning (DL) based semantic segmentation methods have been providing state-of-the-art performance in the last few years. More specifically, these techniques have been successfully applied to medical image classification, segmentation, and detection tasks. One deep learning technique, U-Net, has become one of the most popular for these applications. In this paper, we propose a Recurrent Convolutional Neural Network (RCNN) based on U-Net as well as a Recurrent Residual Convolutional Neural Network (RRCNN) based on U-Net models, which are named RU-Net and R2U-Net respectively. The proposed models utilize the power of U-Net, Residual Network, as well as RCNN. There are several advantages of these proposed architectures for segmentation tasks. First, a residual unit helps when training deep architecture. Second, feature accumulation with recurrent residual convolutional layers ensures better feature representation for segmentation tasks. Third, it allows us to design better U-Net architecture with same number of network parameters with better performance for medical image segmentation. The proposed models are tested on three benchmark datasets such as blood vessel segmentation in retina images, skin cancer segmentation, and lung lesion segmentation. The experimental results show superior performance on segmentation tasks compared to equivalent models including U- Net and residual U-Net (ResU-Net). Index TermsMedical imaging, Semantic segmentation, Convolutional Neural Networks, U-Net, Residual U-Net, RU-Net, and R2U-Net. I. INTRODUCTION OWADAYS DL provides state-of-the-art performance for image classification [1], segmentation [2], detection and tracking [3], and captioning [4]. Since 2012, several Deep Convolutional Neural Network (DCNN) models have been proposed such as AlexNet [1], VGG [5], GoogleNet [6], Residual Net [7], DenseNet [8], and CapsuleNet [9][65]. A DL based approach (CNN in particular) provides state-of-the-art performance for classification and segmentation tasks for several reasons: first, activation functions resolve training problems in DL approaches. Second, dropout helps regularize the networks. Third, several efficient optimization techniques Md Zahangir Alom 1* , Chris Yakopcic 1 , Tarek M. Taha 1 , and Vijayan K. Asari 1 are with the University of Dayton, 300 College Park, Dayton, OH, 45469, USA. (e-mail: {alomm1, cyakopcic1, ttaha1, vasari1}@udayton.edu). are available for training CNN models [1]. However, in most cases, models are explored and evaluated using classification tasks on very large-scale datasets like ImageNet [1], where the outputs of the classification tasks are single label or probability values. Alternatively, small architecturally variant models are used for semantic image segmentation tasks. For example, a fully-connected convolutional neural network (FCN) also provides state-of-the-art results for image segmentation tasks in computer vision [2]. Another variant of FCN was also proposed which is called SegNet [10]. Fig. 1. Medical image segmentation: retina blood vessel segmentation in the left, skin cancer lesion segmentation, and lung segmentation in the right. Due to the great success of DCNNs in the field of computer vision, different variants of this approach are applied in different modalities of medical imaging including segmentation, classification, detection, registration, and medical information processing. The medical imaging comes from different imaging techniques such as Computer Tomography (CT), ultrasound, X-ray, and Magnetic Resonance Imaging (MRI). The goal of Computer-Aided Diagnosis (CAD) is to obtain a faster and better diagnosis to ensure better treatment of a large number of people at the same time. Additionally, efficient automatic processing without human involvement to reduce human error and also reduces overall time and cost. Due to the slow process and tedious nature of Mahmudul Hasan 2 , is with Comcast Labs, Washington, DC, USA. (e-mail: [email protected]). Recurrent Residual Convolutional Neural Network based on U-Net (R2U-Net) for Medical Image Segmentation Md Zahangir Alom 1* , Student Member, IEEE, Mahmudul Hasan 2 , Chris Yakopcic 1 , Member, IEEE, Tarek M. Taha 1 , Member, IEEE, and Vijayan K. Asari 1 , Senior Member, IEEE N

Upload: others

Post on 11-Jul-2020

9 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

Abstract—Deep learning (DL) based semantic segmentation

methods have been providing state-of-the-art performance in the

last few years. More specifically, these techniques have been

successfully applied to medical image classification, segmentation,

and detection tasks. One deep learning technique, U-Net, has

become one of the most popular for these applications. In this

paper, we propose a Recurrent Convolutional Neural Network

(RCNN) based on U-Net as well as a Recurrent Residual

Convolutional Neural Network (RRCNN) based on U-Net models,

which are named RU-Net and R2U-Net respectively. The proposed

models utilize the power of U-Net, Residual Network, as well as

RCNN. There are several advantages of these proposed

architectures for segmentation tasks. First, a residual unit helps

when training deep architecture. Second, feature accumulation

with recurrent residual convolutional layers ensures better feature

representation for segmentation tasks. Third, it allows us to design

better U-Net architecture with same number of network

parameters with better performance for medical image

segmentation. The proposed models are tested on three

benchmark datasets such as blood vessel segmentation in retina

images, skin cancer segmentation, and lung lesion segmentation.

The experimental results show superior performance on

segmentation tasks compared to equivalent models including U-

Net and residual U-Net (ResU-Net).

Index Terms—Medical imaging, Semantic segmentation,

Convolutional Neural Networks, U-Net, Residual U-Net, RU-Net,

and R2U-Net.

I. INTRODUCTION

OWADAYS DL provides state-of-the-art performance for

image classification [1], segmentation [2], detection and

tracking [3], and captioning [4]. Since 2012, several Deep

Convolutional Neural Network (DCNN) models have been

proposed such as AlexNet [1], VGG [5], GoogleNet [6],

Residual Net [7], DenseNet [8], and CapsuleNet [9][65]. A DL

based approach (CNN in particular) provides state-of-the-art

performance for classification and segmentation tasks for

several reasons: first, activation functions resolve training

problems in DL approaches. Second, dropout helps regularize

the networks. Third, several efficient optimization techniques

Md Zahangir Alom1*, Chris Yakopcic1, Tarek M. Taha1, and Vijayan K.

Asari1 are with the University of Dayton, 300 College Park, Dayton, OH,

45469, USA. (e-mail: {alomm1, cyakopcic1, ttaha1, vasari1}@udayton.edu).

are available for training CNN models [1]. However, in most

cases, models are explored and evaluated using classification

tasks on very large-scale datasets like ImageNet [1], where the

outputs of the classification tasks are single label or probability

values. Alternatively, small architecturally variant models are

used for semantic image segmentation tasks. For example, a

fully-connected convolutional neural network (FCN) also

provides state-of-the-art results for image segmentation tasks in

computer vision [2]. Another variant of FCN was also proposed

which is called SegNet [10].

Fig. 1. Medical image segmentation: retina blood vessel segmentation in the

left, skin cancer lesion segmentation, and lung segmentation in the right.

Due to the great success of DCNNs in the field of computer

vision, different variants of this approach are applied in

different modalities of medical imaging including

segmentation, classification, detection, registration, and

medical information processing. The medical imaging comes

from different imaging techniques such as Computer

Tomography (CT), ultrasound, X-ray, and Magnetic Resonance

Imaging (MRI). The goal of Computer-Aided Diagnosis (CAD)

is to obtain a faster and better diagnosis to ensure better

treatment of a large number of people at the same time.

Additionally, efficient automatic processing without human

involvement to reduce human error and also reduces overall

time and cost. Due to the slow process and tedious nature of

Mahmudul Hasan2, is with Comcast Labs, Washington, DC, USA. (e-mail: [email protected]).

Recurrent Residual Convolutional Neural

Network based on U-Net (R2U-Net) for

Medical Image Segmentation

Md Zahangir Alom1*, Student Member, IEEE, Mahmudul Hasan2, Chris Yakopcic1, Member, IEEE,

Tarek M. Taha1, Member, IEEE, and Vijayan K. Asari1, Senior Member, IEEE

N

gaojing
附注
摘要 - 基于深度学习(DL)的语义分割方法在过去几年中一直在提供最先进的性能。更具体地,这些技术已成功应用于医学图像分类,分割和检测任务。一种深度学习技术U-Net已成为这些应用程序中最受欢迎的技术之一。在本文中,我们提出了一种基于U-Net的递归卷积神经网络(RCNN)以及基于U-Net模型的递归残余卷积神经网络(RRCNN),分别命名为RU-Net和R2U-Net。所提出的模型利用了U-Net,残余网络以及RCNN的强大功能。这些提出的架构有几个优点用于分段任务。首先,剩余单元有助于培训深层架构。其次,具有重复残余卷积层的特征累积确保了用于分割任务的更好的特征表示。第三,它允许我们设计具有相同数量的网络参数的更好的U-Net架构,具有更好的医学图像分割性能。所提出的模型在三个基准数据集上进行测试,例如视网膜图像中的血管分割,皮肤癌分割和肺病变分割。与包括UNet和残余U-Net(ResU-Net)在内的等效模型相比,实验结果表明在分割任务方面具有优越的性能。 索引术语 - 医学成像,语义分割,卷积神经网络,U-Net,残余U-Net,RU-Net和R2U-Net。
gaojing
附注
基于U-Net (R2U-Net)循环残差卷积神经网络的医学图像分割 https://cloud.tencent.com/developer/article/1143570 https://arxiv.org/
gaojing
附注
一,导言 DL为图像分类[1],分割[2],检测和跟踪[3]以及字幕[4]提供了最先进的性能。自2012年以来,已经提出了几种深度卷积神经网络(DCNN)模型,如AlexNet [1],VGG [5],GoogleNet [6],残余网[7],DenseNet [8]和CapsuleNet [9] [65]。 ]。基于DL的方法(特别是CNN)为分类和分段任务提供了最先进的性能,原因如下:首先,激活功能解决了DL方法中的训练问题。其次,dropout有助于规范网络。第三,有几种有效的优化技术可用于训练CNN模型[1]。但是,在大多数情况下,使用像ImageNet [1]这样的超大规模数据集上的分类任务来探索和评估模型,其中分类任务的输出是单个标签或概率值。或者,小的体系结构变体模型用于语义图像分割任务。例如,完全连接的卷积神经网络(FCN)还为计算机视觉中的图像分割任务提供了最先进的结果[2]。还提出了FCN的另一种变体,称为SegNet [10]。
gaojing
附注
10。 图1.医学图像分割:左侧视网膜血管分割,皮肤癌病变分割,右侧肺部分割。
gaojing
附注
由于DCNN在计算机视觉领域的巨大成功,该方法的不同变体被应用于医学成像的不同模态,包括分割,分类,检测,配准和医学信息处理。医学成像来自不同的成像技术,例如计算机断层扫描(CT),超声,X射线和磁共振成像(MRI)。计算机辅助诊断(CAD)的目标是获得更快更好的诊断,以确保同时更好地治疗大量人群。此外,高效的自动处理无需人工参与,可减少人为错误,还可减少总体时间和成本。由于手动分割方法的过程缓慢且繁琐,所以对计算机算法有很大的需求,这些算法可以在没有人工交互的情况下快速准确地进行分割。然而,医学图像分割存在一些局限性,包括数据稀缺性和类不平衡。大多数时候,由于多种原因,无法获得大量标签(通常为数千个)用于培训[11]。标记数据集需要该领域的专家,这是昂贵的,并且需要大量的努力和时间。有时,应用不同的数据转换或增强技术(数据白化,旋转,平移和缩放)来增加可用标记样本的数量[12,13和14]。此外,基于补丁的方法用于解决类不平衡问题。在这项工作中,我们已经评估了基于补丁和整个基于图像的方法的建议方法。但是,要从基于补丁的方法切换到与整个图像一起使用的基于像素的方法,我们必须意识到类不平衡问题。在语义分割的情况下,为图像背景分配标签,并为前景区域分配目标类。因此,解决了类不平衡问题而没有任何麻烦。在[13,14]中引入了两种先进的技术,包括交叉熵损失和骰子相似性,以便有效地训练分类和分割任务。
gaojing
高亮
gaojing
高亮
gaojing
高亮
gaojing
高亮
gaojing
高亮
Page 2: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

manual segmentation approaches, there is a significant demand

for computer algorithms that can do segmentation quickly and

accurately without human interaction. However, there are some

limitations of medical image segmentation including data

scarcity and class imbalance. Most of the time the large number

of labels (often in the thousands) for training is not available for

several reasons [11]. Labeling the dataset requires an expert in

this field which is expensive, and it requires a lot of effort and

time. Sometimes, different data transformation or augmentation

techniques (data whitening, rotation, translation, and scaling)

are applied for increasing the number of labeled samples

available [12, 13, and 14]. In addition, patch based approaches

are used for solving class imbalance problems. In this work, we

have evaluated the proposed approaches on both patch-based

and entire image-based approaches. However, to switch from

the patch-based approach to the pixel-based approach that

works with the entire image, we must be aware of the class

imbalance problem. In the case of semantic segmentation, the

image backgrounds are assigned a label and the foreground

regions are assigned a target class. Therefore, the class

imbalance problem is resolved without any trouble. Two

advanced techniques including cross-entropy loss and dice

similarity are introduced for efficient training of classification

and segmentation tasks in [13, 14].

Furthermore, in medical image processing, global

localization and context modulation is very often applied for

localization tasks. Each pixel is assigned a class label with a

desired boundary that is related to the contour of the target

lesion in identification tasks. To define these target lesion

boundaries, we must emphasize the related pixels. Landmark

detection in medical imaging [15, 16] is one example of this.

There were several traditional machine learning and image

processing techniques available for medical image

segmentation tasks before the DL revolution, including

amplitude segmentation based on histogram features [17], the

region based segmentation method [18], and the graph-cut

approach [19]. However, semantic segmentation approaches

that utilize DL have become very popular in recent years in the

field of medical image segmentation, lesion detection, and

localization [20]. In addition, DL based approaches are known

as universal learning approaches, where a single model can be

utilized efficiently in different modalities of medical imaging

such as MRI, CT, and X-ray.

According to a recent survey, DL approaches are applied to

almost all modalities of medical imagining [20, 21].

Furthermore, the highest number of papers have been published

on segmentation tasks in different modalities of medical

imaging [20, 21]. A DCNN based brain tumor segmentation and

detection method was proposed in [22].

From an architectural point of view, the CNN model for

classification tasks requires an encoding unit and provides class

probability as an output. In classification tasks, we have

performed convolution operations with activation functions

followed by sub-sampling layers which reduces the

dimensionality of the feature maps. As the input samples

traverse through the layers of the network, the number of

feature maps increases but the dimensionality of the feature

maps decreases. This is shown in the first part of the model (in

green) in Fig. 2. Since, the number of feature maps increase in

the deeper layers, the number of network parameters increases

respectively. Eventually, the Softmax operations are applied at

the end of the network to compute the probability of the target

classes.

As opposed to classification tasks, the architecture of

segmentation tasks requires both convolutional encoding and

decoding units. The encoding unit is used to encode input

images into a larger number of maps with lower dimensionality.

The decoding unit is used to perform up-convolution (de-

convolution) operations to produce segmentation maps with the

same dimensionality as the original input image. Therefore, the

architecture for segmentation tasks generally requires almost

double the number of network parameters when compared to

the architecture of the classification tasks. Thus, it is important

to design efficient DCNN architectures for segmentation tasks

which can ensure better performance with less number of

network parameters.

This research demonstrates two modified and improved

segmentation models, one using recurrent convolution

networks, and another using recurrent residual convolutional

networks. To accomplish our goals, the proposed models are

Fig. 2. U-Net architecture consisted with convolutional encoding and decoding units that take image as input and produce the segmentation feature maps with

respective pixel classes.

gaojing
高亮
gaojing
附注
此外,在医学图像处理中,全局定位和上下文调制经常应用于本地化任务。为每个像素分配一个类标签,该标签具有与识别任务中的目标病变的轮廓相关的期望边界。要定义这些目标病变边界,我们必须强调相关像素。医学成像中的地标检测[15,16]就是其中的一个例子。在DL革命之前,有几种传统的机器学习和图像处理技术可用于医学图像分割任务,包括基于直方图特征的振幅分割[17],基于区域的分割方法[18]和图形切割方法[19]。 。然而,近年来在医学图像分割,病变检测和定位领域中,利用DL的语义分割方法变得非常流行[20]。另外,基于DL的方法被称为通用学习方法,其中单个模型可以在诸如MRI,CT和X射线的医学成像的不同模态中有效地利用。
gaojing
附注
根据最近的一项调查,DL方法几乎适用于所有医学想象模式[20,21]。 此外,最多的论文已经发表在不同医学成像模式的分割任务上[20,21]。 在[22]中提出了一种基于DCNN的脑肿瘤分割和检测方法。
gaojing
附注
从架构的角度来看,用于分类任务的CNN模型需要编码单元并提供类概率作为输出。 在分类任务中,我们使用激活函数执行卷积运算,然后执行子采样层,这降低了特征映射的维度。 当输入样本遍历网络层时,特征映射的数量增加但特征映射的维度减小。 这在图2的模型的第一部分(绿色)中示出。因为,在更深层中特征图的数量增加,网络参数的数量分别增加。 最终,Softmax操作应用于网络的末端以计算目标类的概率。
gaojing
附注
图2.U-Net架构由卷积编码和解码单元组成,卷积编码和解码单元将图像作为输入并产生具有各自像素类的分割特征图。
gaojing
附注
与分类任务相反,分割任务的架构需要卷积编码和解码单元。 编码单元用于将输入图像编码成具有较低维度的大量地图。 解码单元用于执行上卷积(去卷积)操作以产生具有与原始输入图像相同的维度的分割图。 因此,与分类任务的体系结构相比,分割任务的体系结构通常需要几乎两倍的网络参数。 因此,为分割任务设计有效的DCNN架构是很重要的,这可以用较少的网络参数确保更好的性能。
gaojing
附注
这项研究演示了两种改进和改进的分割模型,一种使用递归卷积网络,另一种使用循环残余卷积网络。 为了实现我们的目标,所提出的模型在不同的医学想象模式下进行评估,如图1所示。这项工作的贡献可归纳如下:
Page 3: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

evaluated on different modalities of medical imagining as

shown in Fig. 1. The contributions of this work can be

summarized as follows:

1) Two new models RU-Net and R2U-Net are introduced for

medical image segmentation.

2) The experiments are conducted on three different

modalities of medical imaging including retina blood vessel

segmentation, skin cancer segmentation, and lung

segmentation.

3) Performance evaluation of the proposed models is

conducted for the patch-based method for retina blood vessel

segmentation tasks and the end-to-end image-based approach

for skin lesion and lung segmentation tasks.

4) Comparison against recently proposed state-of-the-art

methods that shows superior performance against equivalent

models with same number of network parameters.

The paper is organized as follows: Section II discusses related

work. The architectures of the proposed RU-Net and R2U-Net

models are presented in Section III. Section IV, explains the

datasets, experiments, and results. The conclusion and future

direction are discussed in Section V.

II. RELATED WORK

Semantic segmentation is an active research area where

DCNNs are used to classify each pixel in the image

individually, which is fueled by different challenging datasets

in the fields of computer vision and medical imaging [23, 24,

and 25]. Before the deep learning revolution, the traditional

machine learning approach mostly relied on hand engineered

features that were used for classifying pixels independently. In

the last few years, a lot of models have been proposed that have

proved that deeper networks are better for recognition and

segmentation tasks [5]. However, training very deep models is

difficult due to the vanishing gradient problem, which is

resolved by implementing modern activation functions such as

Rectified Linear Units (ReLU) or Exponential Linear Units

(ELU) [5,6]. Another solution to this problem is proposed by

He et al., a deep residual model that overcomes the problem

utilizing an identity mapping to facilitate the training process

[26].

In addition, CNNs based segmentation methods based on

FCN provide superior performance for natural image

segmentation [2]. One of the image patch-based architectures is

called Random architecture, which is very computationally

intensive and contains around 134.5M network parameters.

The main drawback of this approach is that a large number of

pixel overlap and the same convolutions are performed many

times. The performance of FCN has improved with recurrent

neural networks (RNN), which are fine-tuned on very large

datasets [27]. Semantic image segmentation with DeepLab is

one of the state-of-the-art performing methods [28]. SegNet

consists of two parts, one is the encoding network which is a

13-layer VGG16 network [5], and the corresponding decoding

network uses pixel-wise classification layers. The main

contribution of this paper is the way in which the decoder up-

samples its lower resolution input feature maps [10]. Later, an

improved version of SegNet, which is called Bayesian SegNet

was proposed in 2015 [29]. Most of these architectures are

explored using computer vision applications. However, there

are some deep learning models that have been proposed

specifically for the medical image segmentation, as they

consider data insufficiency and class imbalance problems.

One of the very first and most popular approaches for

semantic medical image segmentation is called “U-Net” [12].

A diagram of the basic U-Net model is shown in Fig. 2.

According to the structure, the network consists of two main

parts: the convolutional encoding and decoding units. The basic

convolution operations are performed followed by ReLU

activation in both parts of the network. For down sampling in

the encoding unit, 2×2 max-pooling operations are performed.

In the decoding phase, the convolution transpose (representing

up-convolution, or de-convolution) operations are performed to

up-sample the feature maps. The very first version of U-Net was

used to crop and copy feature maps from the encoding unit to

the decoding unit. The U-Net model provides several

advantages for segmentation tasks: first, this model allows for

the use of global location and context at the same time. Second,

it works with very few training samples and provides better

performance for segmentation tasks [12]. Third, an end-to-end

pipeline process the entire image in the forward pass and

directly produces segmentation maps. This ensures that U-Net

preserves the full context of the input images, which is a major

advantage when compared to patch-based segmentation

approaches [12, 14].

Fig. 3. RU-Net architecture with convolutional encoding and decoding units using recurrent convolutional layers (RCL) based U-Net architecture. The residual

units are used with RCL for R2U-Net architecture.

gaojing
附注
1)引入了两种新型号RU-Net和R2U-Net用于医学图像分割。 2)实验在三种不同的医学成像模式上进行,包括视网膜血管分割,皮肤癌分割和肺部分割。 3)对于用于视网膜血管分割任务的基于贴片的方法和用于皮肤病变和肺部分割任务的端对端基于图像的方法,进行所提出的模型的性能评估。 4)与最近提出的最先进的方法的比较,这些方法对具有相同数量的网络参数的等效模型表现出优越的性能。 本文的结构如下:第二节讨论了相关工作。 拟议的RU-Net和R2U-Net模型的架构在第III节中介绍。 第四部分解释了数据集,实验和结果。 第五节讨论了结论和未来方向。
Page 4: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

However, U-Net is not only limited to the applications in the

domain of medical imaging, nowadays this model is massively

applied for computer vision tasks as well [30, 31]. Meanwhile,

different variants of U-Net models have been proposed,

including a very simple variant of U-Net for CNN-based

segmentation of Medical Imaging data [32]. In this model, two

modifications are made to the original design of U-Net: first, a

combination of multiple segmentation maps and forward

feature maps are summed (element-wise) from one part of the

network to the other. The feature maps are taken from different

layers of encoding and decoding units and finally summation

(element-wise) is performed outside of the encoding and

decoding units. The authors report promising performance

improvement during training with better convergence

compared to U-Net, but no benefit was observed when using a

summation of features during the testing phase [32]. However,

this concept proved that feature summation impacts the

performance of a network. The importance of skipped

connections for biomedical image segmentation tasks have

been empirically evaluated with U-Net and residual networks

[33]. A deep contour-aware network called Deep Contour-

Aware Networks (DCAN) was proposed in 2016, which can

extract multi-level contextual features using a hierarchical

architecture for accurate gland segmentation of histology

images and shows very good performance for segmentation

[34]. Furthermore, Nabla-Net: a deep dig-like convolutional

architecture was proposed for segmentation in 2017 [35].

Other deep learning approaches have been proposed based

on U-Net for 3D medical image segmentation tasks as well. The

3D-Unet architecture for volumetric segmentation learns from

sparsely annotated volumetric images [13]. A powerful end-to-

end 3D medical image segmentation system based on

volumetric images called V-net has been proposed, which

consists of a FCN with residual connections [14]. This paper

also introduces a dice loss layer [14]. Furthermore, a 3D deeply

supervised approach for automated segmentation of volumetric

medical images was presented in [36]. High-Res3DNet was

proposed using residual networks for 3D segmentation tasks in

2016 [37]. In 2017, a CNN based brain tumor segmentation

approach was proposed using a 3D-CNN model with a fully

connected CRF [38]. Pancreas segmentation was proposed in

[39], and Voxresnet was proposed in 2016 where a deep voxel

wise residual network is used for brain segmentation. This

architecture utilizes residual networks and summation of

feature maps from different layers [40].

Alternatively, we have proposed two models for semantic

segmentation based on the architecture of U-Net in this paper.

The proposed Recurrent Convolutional Neural Networks

(RCNN) model based on U-Net is named RU-Net, which is

shown in Fig. 3. Additionally, we have proposed a residual

RCNN based U-Net model which is called R2U-Net. The

following section provides the architectural details of both

models.

III. RU-NET AND R2U-NET ARCHITECTURES

Inspired by the deep residual model [7], RCNN [41], and U-

Net [12], we propose two models for segmentation tasks which

are named RU-Net and R2U-Net. These two approaches utilize

the strengths of all three recently developed deep learning

models. RCNN and its variants have already shown superior

performance on object recognition tasks using different

benchmarks [42, 43]. The recurrent residual convolutional

operations can be demonstrated mathematically according to

the improved-residual networks in [43]. The operations of the

Recurrent Convolutional Layers (RCL) are performed with

respect to the discrete time steps that are expressed according

to the RCNN [41]. Let’s consider the 𝑥𝑙 input sample in the 𝑙𝑡ℎ

layer of the residual RCNN (RRCNN) block and a pixel located

at (𝑖, 𝑗) in an input sample on the kth feature map in the RCL.

Additionally, let’s assume the output of the network 𝑂𝑖𝑗𝑘𝑙 (𝑡) is

at the time step t. The output can be expressed as follows as:

𝑂𝑖𝑗𝑘𝑙 (𝑡) = (𝑤𝑘

𝑓)

𝑇∗ 𝑥𝑙

𝑓(𝑖,𝑗)(𝑡) + (𝑤𝑘

𝑟)𝑇 ∗ 𝑥𝑙𝑟(𝑖,𝑗)

(𝑡 − 1) + 𝑏𝑘 (1)

Here 𝑥𝑙𝑓(𝑖,𝑗)

(𝑡) and 𝑥𝑙𝑟(𝑖,𝑗)

(𝑡 − 1) are the inputs to the

standard convolution layers and for the 𝑙𝑡ℎ RCL respectively.

The 𝑤𝑘𝑓 and 𝑤𝑘

𝑟 values are the weights of the standard

convolutional layer and the RCL of the kth feature map

respectively, and 𝑏𝑘 is the bias. The outputs of RCL are fed to

the standard ReLU activation function 𝑓 and are expressed:

ℱ(𝑥𝑙 , 𝑤𝑙) = 𝑓(𝑂𝑖𝑗𝑘𝑙 (𝑡)) = max (0, 𝑂𝑖𝑗𝑘

𝑙 (𝑡)) (2)

ℱ(𝑥𝑙 , 𝑤𝑙) represents the outputs from of lth layer of the

RCNN unit. The output of ℱ(𝑥𝑙 , 𝑤𝑙) is used for down-sampling

and up-sampling layers in the convolutional encoding and

decoding units of the RU-Net model respectively. In the case of

R2U-Net, the final outputs of the RCNN unit are passed through

the residual unit that is shown Fig. 4(d). Let’s consider that the

output of the RRCNN-block is 𝑥𝑙+1 and can be calculated as

follows:

𝑥𝑙+1 = 𝑥𝑙 + ℱ(𝑥𝑙 , 𝑤𝑙) (3)

Here, 𝑥𝑙 represents the input samples of the RRCNN-block.

The 𝑥𝑙+1 sample is used the input for the immediate succeeding

sub-sampling or up-sampling layers in the encoding and

decoding convolutional units of R2U-Net. However, the

number of feature maps and the dimensions of the feature maps

for the residual units are the same as in the RRCNN-block

shown in Fig. 4 (d).

Fig. 4. Different variant of convolutional and recurrent convolutional units (a)

Forward convolutional units, (b) Recurrent convolutional block (c) Residual

convolutional unit, and (d) Recurrent Residual convolutional units (RRCU).

The proposed deep learning models are the building blocks

of the stacked convolutional units shown in Fig. 4(b) and (d).

gaojing
附注
III。 RU-NET和R2U-NET架构 受深度残差模型[7],RCNN [41]和U-Net [12]的启发,我们提出了两种分割任务模型,分别命名为RU-Net和R2U-Net。这两种方法利用了最近开发的所有三种深度学习模型的优势。 RCNN及其变体已经在使用不同基准的对象识别任务中表现出优越的性能[42,43]。可以根据[43]中改进的残差网络在数学上证明循环残余卷积运算。循环卷积层(RCL)的操作是针对根据RCNN [41]表达的离散时间步长执行的。让我们考虑残差RCNN(RRCNN)块的lth层中的xl输入样本和位于RCL中第k个特征映射的输入样本中的(i,j)的像素。另外,我们假设网络的输出Oijkl(t)是在时间步t。输出可表示如下:
gaojing
附注
这里xlf(i,j)(t)和xlr(i,j)(t-1)分别是标准卷积层和lthRCL的输入。 wkf和wkr值分别是标准卷积层和第k个特征映射的RCL的权重,bk是偏差。 RCL的输出被馈送到标准ReLU激活函数f并表示:
gaojing
附注
ℱ(xl,wl)表示RCNN单元的第l层的输出。 ℱ(xl,wl)的输出分别用于RU-Net模型的卷积编码和解码单元中的下采样和上采样层。 在R2U-Net的情况下,RCNN单元的最终输出通过图4(d)所示的剩余单元。 让我们考虑一下RRCNN块的输出是xl+ 1,可以按如下方式计算:
gaojing
附注
这里,xl表示RRCNN块的输入样本。 xl+ 1样本用于R2U-Net的编码和解码卷积单元中的紧接在后的子采样或上采样层的输入。 然而,特征图的数量和残差单元的特征图的尺寸与图4(d)所示的RRCNN块中的相同。
gaojing
附注
图4.卷积和递归卷积单位的不同变体(a)前向卷积单位,(b)循环卷积单位(c)残差卷积单位,和(d)递归残差卷积单位(RRCU)。
Page 5: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

There are four different architectures evaluated in this work.

First, U-Net with forward convolution layers and feature

concatenation is applied as an alternative to the crop and copy

method found in the primary version of U-Net [12]. The basic

convolutional unit of this model is shown in Fig. 4(a). Second,

U-Net with forward convolutional layers with residual

connectivity is used, which is often called residual U-net

(ResU-Net) and is shown in Fig. 4(c) [14]. The third

architecture is U-Net with forward recurrent convolutional

layers as shown in Fig. 4(b), which is named RU-Net. Finally,

the last architecture is U-Net with recurrent convolution layers

with residual connectivity as shown in Fig. 4(d), which is

named R2U-Net. The pictorial representation of the unfolded

RCL layers with respect to time-step is shown in Fig 5. Here

t=2 (0 ~ 2), refers to the recurrent convolutional operation that

includes one single convolution layer followed by two sub-

sequential recurrent convolutional layers. In this

implementation, we have applied concatenation to the feature

maps from the encoding unit to the decoding unit for both RU-

Net and R2U-Net models.

Fig. 5. Unfolded recurrent convolutional units for t = 2 (left) and t = 3 (right).

The differences between the proposed models with respect to

the U-Net model are three-fold. This architecture consists of

convolutional encoding and decoding units same as U-Net.

However, the RCLs and RCLs with residual units are used

instead of regular forward convolutional layers in both the

encoding and decoding units. The residual unit with RCLs helps

to develop a more efficient deeper model. Second, the efficient

feature accumulation method is included in the RCL units of

both proposed models. The effectiveness of feature

accumulation from one part of the network to the other is shown

in the CNN-based segmentation approach for medical imaging.

In this model, the element-wise feature summation is performed

outside of the U-Net model [32]. This model only shows the

benefit during the training process in the form of better

convergence. However, our proposed models show benefits for

both training and testing phases due to the feature accumulation

inside the model. The feature accumulation with respect to

different time-steps ensures better and stronger feature

representation. Thus, it helps extract very low-level features

which are essential for segmentation tasks for different

modalities of medical imaging (such as blood vessel

segmentation). Third, we have removed the cropping and

copying unit from the basic U-Net model and use only

concatenation operations, resulting a much-sophisticated

architecture that results in better performance.

Fig. 6. Example images from training dataset: left column from DRIVE dataset, middle column from STARE dataset and right column from CHASE-DB1

dataset. The first row shows the original images, second row shows fields of

view (FOV), and third row shows the target outputs.

There are several advantages of using the proposed

architectures when compared with U-Net. The first is the

efficiency in terms of the number of network parameters. The

proposed RU-Net, and R2U-Net architectures are designed to

have the same number of network parameters when compared

to U-Net and ResU-Net, and RU-Net and R2U-Net show better

performance on segmentation tasks. The recurrent and residual

operations do not increase the number of network parameters.

However, they do have a significant impact on training and

testing performance. This is shown through empirical evidence

with a set of experiments in the following sections [43]. This

approach is also generalizable, as it easily be applied deep

learning models based on SegNet [10], 3D-UNet [13], and V-

Net [14] with improved performance for segmentation tasks.

IV. EXPERIMENTAL SETUP AND RESULTS

To demonstrate the performance of the RU-Net and R2U-Net

models, we have tested them on three different medical imaging

datasets. These include blood vessel segmentations from retina

images (DRIVE, STARE, and CHASE_DB1 shown in Fig. 6),

skin cancer lesion segmentation, and lung segmentation from

2D images. For this implementation, the Keras, and

TensorFlow frameworks are used on a single GPU machine

with 56G of RAM and an NIVIDIA GEFORCE GTX-980 Ti.

A. Database Summary

1) Blood Vessel Segmentation

We have experimented on three different popular datasets for

retina blood vessel segmentation including DRIVE, STARE,

and CHASH_DB1. The DRIVE dataset is consisted of 40 color

gaojing
附注
所提出的深度学习模型是图4(b)和(d)所示的堆叠卷积单元的构建块。在这项工作中评估了四种不同的体系结构。首先,应用具有前向卷积层和特征连接的U-Net作为U-Net主要版本中的裁剪和复制方法的替代方法[12]。该模型的基本卷积单元如图4(a)所示。其次,使用具有残余连通性的前向卷积层的U-Net,其通常称为残余U-net(ResU-Net),如图4(c)所示[14]。第三种结构是具有前向循环卷积层的U-Net,如图4(b)所示,其名为RU-Net。最后,最后一个体系结构是U-Net,带有具有剩余连通性的循环卷积层,如图4(d)所示,其名为R2U-Net。展开的RCL层相对于时间步长的图形表示如图5所示。这里t = 2(0~2),指的是包括一个单个卷积层,然后是两个子序列递归卷积的循环卷积运算。层。在此实现中,我们将连接应用于从RU-Net和R2U-Net模型的编码单元到解码单元的特征映射。
gaojing
附注
图5. t = 2(左)和t = 3(右)的展开的递归卷积单位。
gaojing
附注
所提出的模型与U-Net模型之间的差异是三倍的。该架构由与U-Net相同的卷积编码和解码单元组成。然而,在编码和解码单元中使用具有残余单元的RCL和RCL代替常规前向卷积层。具有RCL的剩余单元有助于开发更有效的更深层模型。其次,有效的特征累积方法包含在两个提出的模型的RCL单元中。在用于医学成像的基于CNN的分割方法中示出了从网络的一部分到另一部分的特征累积的有效性。在该模型中,逐元素特征求和在U-Net模型之外执行[32]。该模型仅以更好的收敛的形式显示培训过程中的益处。然而,由于模型内部的特征积累,我们提出的模型显示了训练和测试阶段的好处。关于不同时间步长的特征累积确保了更好和更强的特征表示。因此,它有助于提取非常低级的特征,这对于不同的医学成像模式(例如血管分割)的分割任务是必不可少的。第三,我们从基本的U-Net模型中删除了裁剪和复制单元,并仅使用连接操作,从而产生了一个非常复杂的架构,从而获得了更好的性能。
gaojing
附注
图6.来自训练数据集的示例图像:来自DRIVE数据集的左列,来自STARE数据集的中间列和来自CHASE-DB1数据集的右列。第一行显示原始图像,第二行显示视野(FOV),第三行显示目标输出。
gaojing
附注
与U-Net相比,使用所提出的架构有几个优点。第一个是网络参数数量方面的效率。与U-Net和ResU-Net相比,建议的RU-Net和R2U-Net架构具有相同数量的网络参数,RU-Net和R2U-Net在分段任务上表现出更好的性能。循环和剩余操作不会增加网络参数的数量。但是,它们确实对培训和测试性能产生了重大影响。这通过经验证据显示,并在以下各节中进行了一系列实验[43]。这种方法也是可以推广的,因为它可以很容易地应用基于SegNet [10],3D-UNet [13]和V-Net [14]的深度学习模型,并提高了分割任务的性能。
gaojing
附注
为了演示RU-Net和R2U-Net模型的性能,我们在三种不同的医学成像数据集上进行了测试。 这些包括来自视网膜图像的血管分割(图6中所示的DRIVE,STARE和CHASE_DB1),皮肤癌病变分割和来自2D图像的肺分割。 对于此实现,Keras和TensorFlow框架在具有56G RAM和NIVIDIA GEFORCE GTX-980 Ti的单GPU机器上使用。
Page 6: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

retinal images in total, in which 20 samples are used for training

and remaining 20 samples are used for testing. The size of each

original image is 565×584 pixels [44]. To develop a square

dataset, the images are cropped to only contain the data from

columns 9 through 574, which then makes each image 565×565

pixels. In this implementation, we considered 190,000

randomly selected patches from 20 of the images in the DRIVE

dataset, where 171,000 patches are used for training, and the

remaining 19,000 patches used for validation. The size of each

patch is 48×48 for all three datasets shown in Fig. 7. The second

dataset, STARE, contains 20 color images, and each image has

a size of 700×605 pixels [45, 46]. Due to the smaller number of

samples, two approaches are applied very often for training and

testing on this dataset. First, training sometimes performed with

randomly selected samples from all 20 images [53].

Fig. 7. Example patches in the left and corresponding outputs of patches are

shown in the right.

Fig. 8. Experimental outputs for DRIVE dataset using R2UNet: first row shows input image in gray scale, second row show ground truth, and third row shows

the experimental outputs.

Another approach is the “leave-one-out” method, in which

each image is tested, and training is conducted on the remaining

19 samples [47]. Therefore, there is no overlap between training

and testing samples. In this implementation, we used the “leave-

one-out” approach for STARE dataset. The CHASH_DB1

dataset contains 28 color retina images and the size of each

image is 999×960 pixels [48]. The images in this dataset were

collected from both left and right eyes of 14 school children.

The dataset is divided into two sets where samples are selected

randomly. A 20-sample set is used for training and the

remaining 8 samples are used for testing.

As the dimensionality of the input data larger than the entire

DRIVE dataset, we have considered 250,000 patches in total

from 20 images for both STARE and CHASE_DB1. In this case

225,000 patches are used for training and the remaining 25,000

patches are used for validation. Since the binary FOV (which

is shown in second row in Fig. 6) is not available for the STARE

and CHASE_DB1 datasets, we generated FOV masks using a

similar technique to the one described in [47]. One advantage

of the patch-based approach is that the patches give the network

access to local information about the pixels, which has impact

on overall prediction. Furthermore, it ensures that the classes of

the input data are balanced. The input patches are randomly

sampled over an entire image, which also includes the outside

region of the FOV.

2) Skin Cancer Segmentation

This dataset is taken from the Kaggle competition on skin

lesion segmentation that occurred in 2017 [49]. This dataset

contains 2000 samples in total. It consists of 1250 training

samples, 150 validation samples, and 600 testing samples. The

original size of each sample was 700×900, which was rescaled

to 256×256 for this implementation. The training samples

include the original images, as well as corresponding target

binary images containing cancer or non-cancer lesions. The

target pixels are represented with a value of either 255 or 0 for

the pixels outside of the target lesion.

3) Lung Segmentation

The Lung Nodule Analysis (LUNA) competition at the

Kaggle Data Science Bowl in 2017 was held to find lung lesions

in 2D and 3D CT images. The provided dataset consisted of 534

2D samples with respective label images for lung segmentation

[50]. For this study, 70% of the images are used for training and

the remaining 30% are used for testing. The original image size

was 512×512, however, we resized the images to 256×256

pixels in this implementation.

B. Quantitative Analysis Approaches

For quantitative analysis of the experimental results, several

performance metrics are considered, including accuracy (AC),

sensitivity (SE), specificity (SP), F1-score, Dice coefficient

(DC), and Jaccard similarity (JS). To do this we also use the

variables True Positive (TP), True Negative (TN), False

Positive (FP), and False Negative (FN). The overall accuracy is

calculated using Eq. (4), and sensitivity is calculated using Eq.

(5).

𝐴𝐶 = 𝑇𝑃+𝑇𝑁

𝑇𝑃+𝑇𝑁+𝐹𝑃+𝐹𝑁 (4)

𝑆𝐸 = 𝑇𝑃

𝑇𝑃+𝐹𝑁 (5)

Furthermore, specificity is calculated using the following Eq.

(6).

𝑆𝑃 = 𝑇𝑁

𝑇𝑁+𝐹𝑃 (6)

The DC is expressed as in Eq. (7) according to [51]. Here GT

refers to the ground truth and SR refers the segmentation result.

gaojing
附注
定量分析方法
gaojing
附注
对于实验结果的定量分析,考虑了若干性能指标,包括准确度(AC),灵敏度(SE),特异性(SP),F1得分,骰子系数(DC)和Jaccard相似性(JS)。 为此,我们还使用变量True Positive(TP),True Negative(TN),False Positive(FP)和False Negative(FN)。 使用方程式计算总体准确度。 (4),使用方程(5)计算灵敏度
gaojing
附注
DC表示为等式1。 (7)根据[51]。 这里GT指的是地面实况,SR指的是分割结果。
gaojing
附注
图8.使用R2UNet的DRIVE数据集的实验输出:第一行显示灰度的输入图像,第二行显示地面实况,第三行显示实验输出。
gaojing
附注
图7.左边的示例补丁和补丁的相应输出显示在右侧。
Page 7: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

𝐷𝐶 = 2 |𝐺𝑇∩𝑆𝑅|

|𝐺𝑇|+|𝑆𝑅| (7)

The JS is represented using Eq. (8) as in [52].

𝐽𝑆 = |𝐺𝑇∩𝑆𝑅|

|𝐺𝑇∪𝑆𝑅| (8)

However, the area under curve (AUC) and the receiver

operating characteristics (ROC) curve are common evaluation

measures for medical image segmentation tasks. In this

experiment, we utilized both analytical methods to evaluate the

performance of the proposed approaches considering the

mentioned criterions against existing state-of-the-art

techniques.

Fig. 9. Training accuracy of the proposed models of RU-Net, and R2U-Net against ResU-Net and U-Net.

C. Results

1) Retina Blood Vessel Segmentation Using the DRIVE

Dataset

The precise segmentation results achieved with the proposed

R2U-Net model are shown in Fig. 8. Figs. 9 and 10 show the

training and validation accuracy when using the DRIVE

dataset. These figures show that the proposed R2U-Net and

RU-Net models provide better performance during both the

training and validation phase when compared to U-Net and

ResU-Net.

Fig. 10. Validation accuracy of the proposed models against ResU-Net and U-

Net.

2) Retina blood vessel segmentation on the STARE dataset

The experimental outputs of R2U-Net when using the

STARE dataset are shown in Fig. 11. The training and

validation accuracy for the STARE dataset is shown in Figs. 12

and 13 respectively.

R2U-Net shows a better performance than all other models

during training. In addition, the validation accuracy in Fig. 13

demonstrates that the RU-Net and R2U-Net models provide

better validation accuracy when compared to the equivalent U-

Net and ResU-Net models. Thus, the performance demonstrates

the effectiveness of the proposed approaches for segmentation

tasks.

Fig. 11. Experimental outputs of STARE dataset using R2UNet: first row shows

input image after performing normalization, second row show ground truth, and

third row shows the experimental outputs.

Fig. 12. Training accuracy in STARE dataset for R2U-Net, RU-Net, ResU-Net, and U-Net.

Fig. 13. Validation accuracy in STARE dataset for R2U-Net, RU-Net, ResU-Net, and U-Net.

3) CHASE_DB1

For qualitative analysis, the example outputs of R2U-Net are

shown in Fig. 14. For quantitative analysis, the results are given

gaojing
附注
然而,曲线下面积(AUC)和接收器操作特征(ROC)曲线是医学图像分割任务的常用评估测量。 在这个实验中,我们利用两种分析方法来评估所提出的方法的性能,考虑所提到的针对现有最先进技术的标准。
gaojing
附注
图9.提议的RU-Net模型和R2U-Net对ResU-Net和U-Net的训练准确性。
gaojing
附注
1)使用DRIVE数据集进行视网膜血管分割 用所提出的R2U-Net模型实现的精确分割结果如图8所示。 图9和10显示了使用DRIVE数据集时的训练和验证准确性。 这些数据表明,与U-Net和ResU-Net相比,提议的R2U-Net和RU-Net模型在培训和验证阶段提供了更好的性能。
gaojing
附注
图10.针对ResU-Net和U-Net所提出的模型来验证准确性。
gaojing
附注
2)STARE数据集上的视网膜血管分割 使用STARE数据集时R2U-Net的实验输出如图11所示.STARE数据集的训练和验证准确度如图11和12所示。 分别为12和13。 R2U-Net在训练期间表现出比所有其他模型更好的性能。 此外,图13中的验证准确性表明,与等效的U-Net和ResU-Net模型相比,RU-Net和R2U-Net模型提供了更好的验证准确性。 因此,性能证明了所提出的分割任务方法的有效性。
gaojing
附注
图11.使用R2UNet的STARE数据集的实验输出:第一行显示执行归一化后的输入图像,第二行显示地面实况,第三行显示实验输出。
gaojing
附注
图12. R2U-Net,RU-Net,ResU-Net和U-Net的STARE数据集的训练准确性。
gaojing
附注
图13.R2U-Net,RU-Net,ResU-Net和U-Net的STARE数据集中的验证准确性。
Page 8: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

in Table I. From the table, it can be concluded that in all cases,

the proposed RU-Net and R2U-Net models show better

performance in terms of AUC and accuracy. The ROC for the

highest AUCs for the R2U-Net model on each of the three retina

blood vessel segmentation datasets is shown in Fig. 15.

Fig. 14. Qualitative analysis for CHASE_DB1 dataset. The segmentation

outputs of 8 testing samples using R2U-Net. First row shows the input images, second row is ground truth, and third row shows the segmentation outputs using

R2U-Net.

4) Skin Cancer Lesion Segmentation

In this implementation, this dataset is preprocessed with

mean subtraction and normalized according to the standard

deviation. We used the ADAM optimization technique with a

learning rate of 2×10-4 and binary cross entropy loss. In

addition, we also calculated MSE error during the training and

validation phase. In this case 10% of the samples are used for

validation during training with a batch size of 32 and 150

epochs.

The training accuracy of the proposed models R2U-Net and

RU-Net was compared with that of ResU-Net and U-Net for an

end-to-end image based segmentation approach. The result is

shown in Fig. 16. The validation accuracy is shown in Fig. 17.

In both cases, the proposed models show better performance

when compared with the equivalent U-Net and ResU-Net

models. This clearly demonstrates the robustness of the

proposed models in end-to-end image-based segmentation

tasks.

Fig. 15. AUC for retina blood vessel segmentation for the best performance

achieved with R2U-Net.

TABLE I. EXPERIMENTAL RESULTS OF PROPOSED APPROACHES FOR RETINA BLOOD VESSEL SEGMENTATION AND COMPARISON AGAINST OTHER

TRADITIONAL AND DEEP LEARNING-BASED APPROACHES. Dataset Methods Year F1-score SE SP AC AUC

DRIVE Chen [53] 2014 - o.7252 0.9798 0.9474 0.9648

Azzopardi [54] 2015 - 0.7655 0.9704 0.9442 0.9614

Roychowdhury[55] 2016 - 0.7250 0.9830 0.9520 0.9620

Liskowsk [56] 2016 - 0.7763 0.9768 0.9495 0.9720

Qiaoliang Li [57] 2016 - 0.7569 0.9816 0.9527 0.9738

U-Net 2018 0.8142 0.7537 0.9820 0.9531 0.9755

Residual U-Net 2018 0.8149 0.7726 0.9820 0.9553 0.9779

Recurrent U-Net 2018 0.8155 0.7751 0.9816 0.9556 0.9782

R2U-Net 2018 0.8171 0.7792 0.9813 0.9556 0.9784

STARE Marin et al. [58] 2011 - 0.6940 0.9770 0.9520 0.9820

Fraz [59] 2012 - 0.7548 0.9763 0.9534 0.9768

Roychowdhury[55] 2016 - 0.7720 0.9730 0.9510 0.9690

Liskowsk [56] 2016 - 0.7867 0.9754 0.9566 0.9785

Qiaoliang Li [57] 2016 - 0.7726 0.9844 0.9628 0.9879

U-Net 2018 0.8373 0.8270 0.9842 0.9690 0.9898

Residual U-Net 2018 0.8388 0.8203 0.9856 0.9700 0.9904

Recurrent U-Net 2018 0.8396 0.8108 0.9871 0.9706 0.9909

R2U-Net 2018 0.8475 0.8298 0.9862 0.9712 0.9914

CHASE_DB1 Fraz [59] 2012 - 0.7224 0.9711 0.9469 0.9712

Fraz [60] 2014 - - - 0.9524 0.9760

Azzopardi [54] 2015 - 0.7655 0.9704 0.9442 0.9614

Roychowdhury[55] 2016 - 0.7201 0.9824 0.9530 0.9532

Qiaoliang Li [57] 2016 - 0.7507 0.9793 0.9581 0.9793

U-Net 2018 0.7783 0.8288 0.9701 0.9578 0.9772

Residual U-Net 2018 0.7800 0.7726 0.9820 0.9553 0.9779

Recurrent U-Net 2018 0.7810 0.7459 0.9836 0.9622 0.9803

R2U-Net 2018 0.7928 0.7756 0.9820 0.9634 0.9815

gaojing
附注
3)CHASE_DB1 对于定性分析,R2U-Net的示例输出如图14所示。对于定量分析,结果在表I中给出。从表中可以得出结论,在所有情况下,建议的RU-Net和R2U -Net模型在AUC和准确度方面表现出更好的性能。 在图15中显示了三个视网膜血管分割数据集中的每一个上的R2U-Net模型的最高AUC的ROC。
gaojing
附注
图14. CHASE_DB1数据集的定性分析。 使用R2U-Net对8个测试样本进行分段输出。 第一行显示输入图像,第二行显示专家分割,第三行显示使用R2U-Net的分割输出。
gaojing
附注
4)皮肤癌病变分割 在该实现中,该数据集使用平均减法进行预处理并根据标准偏差进行归一化。我们使用ADAM优化技术,学习率为2×10-4,二元交叉熵损失。此外,我们还在训练和验证阶段计算了MSE误差。在这种情况下,10%的样品在培训期间用于验证,批量大小为32和150个时期。 将所提出的模型R2U-Net和RU-Net的训练精度与ResU-Net和U-Net的训练精度进行比较,以实现端到端的基于图像的分割方法。结果如图16所示。验证精度如图17所示。在这两种情况下,与等效的U-Net和ResU-Net模型相比,所提出的模型表现出更好的性能。这清楚地证明了所提出的模型在端到端基于图像的分割任务中的稳健性。
gaojing
附注
图15.用于视网膜血管分割的AUC,用于使用R2U-Net实现的最佳性能。
gaojing
高亮
Page 9: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

Fig. 16. Training accuracy for skin lesion segmentation.

The quantitative results of this experiment were compared

against existing methods as shown in Table II. Some of the

example outputs from the testing phase are shown in Fig. 18.

The first column shows the input images, the second column

shows the ground truth, the network outputs are shown in the

third column, and the fourth column demonstrates the final

outputs after performing post processing with a threshold of 0.5.

Figure 18 shows promising segmentation results.

Fig. 17. Validation accuracy for skin lesion segmentation.

In most cases, the target lesions are segmented accurately

with almost the same shape of ground truth. However, if we

observe the second and third rows in Fig. 18, it can be clearly

seen that the input images contain two spots, one is a target

lesion and the other bright spot which is not a target. This result

is obtained even though the non-target lesion is brighter than

the target lesion shown in the third row in Fig. 18. The R2U-

Net model still segments the desired part accurately, which

clearly shows the robustness of the proposed segmentation

method.

We have compared the performance of the proposed

approaches against recently published results with respect to

sensitivity, specificity, accuracy, AUC, and DC. The proposed

R2U-Net model provides a testing accuracy 0.9424 with a

higher AUC, which is 0.9419. The average AUC for skin lesion

segmentation is shown in Fig. 19. In addition, we calculated the

average DC in the testing phase and achieved 0.8616, which is

around 1.26% better than recently proposed alternatives [62].

Furthermore, the JSC and F1 scores are calculated and the R2U-

Net model obtains 0.9421 for JSC and 0.8920 for F1 score for

skin lesion segmentation with t=3. These results are achieved

with a R2U-Net model that only contains about 1.037 million

(M) network parameters. Contrarily, the work presented in [61]

evaluated VGG-16 and Incpetion-V3 models for skin lesion

segmentation, but those networks contained around 138M and

23M network parameters respectively.

Fig. 18. This results demonstrates qualitative assessment of the proposed R2U-Net for skin cancer segmentation task with t=3. First column is the input

sample, second column is ground truth, third column shows the outputs from

TABLE II. EXPERIMENTAL RESULTS OF PROPOSED APPROACHES FOR SKIN CANCER LESION SEGMENTATION AND COMPARISON AGAINST OTHER

EXISTING APPROACHES. JACCARD SIMILARITY SCORE (JSC). Methods Year SE SP JSC F1-score AC AUC DC

Conv. classifier VGG-16 [61] 2017 0.533 - - - 0.6130 0.6420 -

Conv. classifier Inception-v3[61] 2017 0.760 - - - 0.6930 0.7390 -

Melanoma detection [62] 2017 - - - - o.9340 - 0.8490

Skin Lesion Analysis [63] 2017 0.8250 0.9750 - - 0.9340 - -

U-Net (t=2) 2018 0.9479 0.9263 0.9314 0.8682 0.9314 0.9371 0.8476

ResU-Net (t=2) 2018 0.9454 0.9338 0.9367 0.8799 0.9367 0.9396 0.8567

RecU-Net (t=2) 2018 0.9334 0.9395 0.9380 0.8841 0.9380 0.9364 0.8592

R2U-Net (t=2) 2018 0.9496 0.9313 0.9372 0.8823 0.9372 0.9405 0.8608

R2U-Net (t=3) 2018 0.9414 0.9425 0.9421 0.8920 0.9424 0.9419 0.8616

gaojing
附注
图16.皮肤病变分割的训练准确性。
gaojing
附注
将该实验的定量结果与现有方法进行比较,如表II所示。 测试阶段的一些示例输出如图18所示。第一列显示输入图像,第二列显示基本事实,网络输出显示在第三列,第四列显示最终输出 在执行阈值为0.5的后处理之后。 图18显示了有希望的分割结果。
gaojing
附注
图17.皮肤病变分割的验证准确性。
gaojing
附注
在大多数情况下,目标病变被准确地分割,具有几乎相同的地面事实形状。然而,如果我们观察图18中的第二和第三行,可以清楚地看到输入图像包含两个斑点,一个是目标病变,另一个是不是目标的亮点。即使非目标病变比图18中第三行所示的目标病变更亮,也可以获得该结果.R2U-Net模型仍然准确地分割所需部分,这清楚地显示了所提出的分割方法的稳健性。 我们将提出的方法的性能与最近公布的结果进行了比较,包括灵敏度,特异性,准确性,AUC和DC。建议的R2U-Net模型提供0.9424的测试精度,具有更高的AUC,即0.9419。皮肤病变分割的平均AUC如图19所示。此外,我们计算了测试阶段的平均DC并达到了0.8616,比最近提出的替代方案好了约1.26%[62]。此外,计算了JSC和F1分数,R2U-Net模型获得了JSC的0.9421和用于皮肤病变分割的t = 3的F1评分的0.8920。这些结果是通过R2U-Net模型实现的,该模型仅包含大约103.7万(M)个网络参数。相反,[61]中提出的工作评估了皮肤病变分割的VGG-16和Incpetion-V3模型,但这些网络分别包含大约138M和23M网络参数。
gaojing
附注
图18.该结果证明了对于t = 3的皮肤癌分割任务的所提出的R2U-Net的定性评估。第一列是输入样本,第二列是基础事实,第三列显示网络的输出,第四列显示在用0.5执行阈值处理后的最终结果。
gaojing
高亮
gaojing
高亮
Page 10: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

network, and fourth column show the final resulting after performing thresholding with 0.5.

5) Lung Segmentation

Lung segmentation is very important for analyzing lung

related diseases, and can be applied to lung cancer segmentation

and lung pattern classification for identifying other problems.

In this experiment, the ADAM optimizer is used with a learning

rate of 2×10-4. We used binary cross entropy loss, and also

calculated MSE during training and validation. In this case 10%

of the samples were used for validation with a batch size of 16

and 150 epochs 150. Table III shows the summary of how well

the proposed models performed against equivalent U-Net and

ResU-Net models. The experimental results show that the

proposed models outperform the U-Net and ResU-Net models

with same number of network parameters.

Fig. 19. ROC-AUC for skin segmentation four models with t=2 and t=3.

Furthermore, many models struggle to define the class

boundary properly during segmentation tasks [64]. However, if

we observe the experimental outputs shown in Fig. 20, the

outputs in the third column show different hit maps on the

border, which can be used to define the boundary of the lung

region, while the ground truth tends to have a smooth boundary.

In addition, if we observe the input, ground truth, and output

of this proposed approaches in the second row, it can be

observed that the output of the proposed approaches shows

better segmentation with appropriate contour. The ROC with

AUCs are shown Fig. 21. The highest AUC is achieved with the

proposed approach of R2U-Net with t=3.

D. Evaluation

Most of the cases, the networks are evaluated for different

segmentation tasks with following architectures:

164128256512256 128641 that require

4.2M network parameters and 164128256512256

128641, which require about 8.5M network parameters

respectively. However, we also experimented with U-Net,

ResU-Net, RU-Net, and R2U-Net models with following

structure: 116326412864 32161. In this

case we used a time-step of t=3, which refers to one forward

convolution layer followed by three subsequent recurrent

convolutional layers. This network was tested on skin and lung

lesion segmentation. Though the number of network parameters

increase little bit with respect to the time-step in the recurrent

convolution layer, further improved performance can be clearly

seen in the last rows of Table II and III. Furthermore, we have

evaluated both of the proposed models for patch-based

modeling on retina blood vessel segmentation and end-to-end

image-based methods for skin and lung lesion segmentation.

In both cases, the proposed models outperform existing state-

of-the-art methods including ResU-Net and U-Net in terms of

AUC and accuracy on all three datasets. The network

architectures with different numbers of network parameters

with respect to the different time-step are shown in Table IV.

The processing times during the testing phase for the STARE,

CHASE_DB, and DRIVE datasets were 6.42, 8.66, and 2.84

seconds per sample respectively. In addition, skin cancer

segmentation and lung segmentation take 0.22 and 1.145

seconds per sample respectively.

TABLE IV. ARCHITECTURE AND NUMBER OF NETWORK PARAMETERS. t Network architectures Number of parameters

(million)

2 1-> 16->32->64>128->64 –> 32-

>16->1

0.845

3 1-> 16->32->64>128->64 –> 32->16->1

1.037

TABLE III. EXPERIMENTAL OUTPUTS OF PROPOSED MODELS OF RU-NET AND R2U-NET FOR LUNG SEGMENTATION AND COMPARISON

AGAINST RESU-NET AND U-NET MODELS. Methods Year SE SP JSC F1-Score AC AUC

U-Net (t=2) 2018 0.9696 0.9872 0.9858 0.9658 0.9828 0.9784

ResU-Net(t=2) 2018 0.9555 0.9945 0.9850 0.9690 0.9849 0.9750

RU-Net (t=2) 2018 0.9734 0.9866 0.9836 0.9638 0.9836 0.9800

R2U-Net (t=2) 2018 0.9826 0.9918 0.9897 0.9780 0.9897 0.9872

R2U-Net (t=3) 2018 0.9832 0.9944 0.9918 0.9823 0.9918 0.9889

gaojing
附注
5)肺部分割 肺部分割对于分析肺相关疾病非常重要,并且可以应用于肺癌分割和肺部模式分类以识别其他问题。在本实验中,使用ADAM优化器,学习率为2×10-4。我们使用二元交叉熵损失,并且还在训练和验证期间计算了MSE。在这种情况下,10%的样品用于验证,批量大小为16和150个时期150.表III显示了所提出的模型对等效U-Net和ResU-Net模型的表现如何。实验结果表明,所提出的模型优于U-Net和ResU-Net模型。此外,许多模型在分割任务期间难以正确定义类边界[64]。然而,如果我们观察图20中所示的实验输出,则第三列中的输出显示边界上的不同命中图,其可用于定义肺区域的边界,而地面实况趋于平滑。边界。
gaojing
附注
5)肺部分割 此外,如果我们观察第二行中所提出的方法的输入,基础事实和输出,则可以观察到所提出的方法的输出显示具有适当轮廓的更好分割。具有AUC的ROC如图21所示。使用提出的R2U-Net方法获得最高AUC,其中t = 3。
gaojing
附注
D.评估 大多数情况下,网络将根据以下架构评估不同的分段任务:164128256512256128641需要4.2M网络参数和164128256512 256128641,分别需要大约8.5M的网络参数。但是,我们还尝试了以下结构的U-Net,ResU-Net,RU-Net和R2U-Net模型:1163264128→6432161。在这种情况下,我们使用t = 3的时间步长,其指的是一个前向卷积层,后面是三个随后的递归卷积层。该网络在皮肤和肺部病变分割上进行了测试。尽管网络参数的数量相对于循环卷积层中的时间步长略微增加,但是在表II和III的最后几行中可以清楚地看到进一步改善的性能。此外,我们已经评估了所提出的用于视网膜血管分割的基于贴片的建模的模型以及用于皮肤和肺部病变分割的端对端基于图像的方法。 在这两种情况下,所提出的模型在AUC和所有三个数据集的准确性方面都优于现有的最先进的方法,包括ResU-Net和U-Net。表IV中示出了关于不同时间步长具有不同数量的网络参数的网络架构。 STARE,CHASE_DB和DRIVE数据集在测试阶段的处理时间分别为每个样本6.42,8.66和2.84秒。此外,皮肤癌细分和肺部分割每个样本分别需要0.22和1.145秒。
gaojing
附注
图19.用于皮肤分割的ROC-AUC四个模型,其中t = 2且t = 3。
gaojing
附注
需要的参数数量
Page 11: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

Fig. 20. Qualitative assessment of R2U-Net performance on Lung segmentation

dataset: first column input images, second column ground truth, and third column outputs with R2U-Net.

E. Computational time

The computational time for testing per sample is shown in

Table V for blood vessel segmentation for retina images, skin

cancer, and lung segmentation respectively.

TABLE V. COMPUTATIONAL TIME FOR TESTING PHASE. Dataset Time (Sec.)/ sample

Blood vessel

segmentation

DRIVE 6.42

STARE 8.66

CHASE_DB1 2.84

Skin cancer segmentation 0.22

Lung segmentation 1.15

V. CONCLUSION AND FUTURE WORKS

In this paper, we proposed an extension of the U-Net

architecture using Recurrent Convolutional Neural Networks

and Recurrent Residual Convolutional Neural Networks. The

proposed models are called “RU-Net” and “R2U-Net”

respectively. These models were evaluated using three different

applications in the field of medical imaging including retina

blood vessel segmentation, skin cancer lesion segmentation,

and lung segmentation. The experimental results demonstrate

that the proposed RU-Net, and R2U-Net models show better

performance in segmentation tasks with the same number of

network parameters when compared to existing methods

including the U-Net and residual U-Net (or ResU-Net) models

on all three datasets. In addition, results show that these

proposed models not only ensure better performance during the

training but also in testing phase. In future, we would like to

explore the same architecture with a novel feature fusion

strategy from encoding to the decoding units.

Fig. 21. ROC curve for lung segmentation four models with t=2 and t=3.

REFERENCES

[1] Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. "ImageNet

classification with deep convolutional neural networks." Advances in

neural information processing systems. 2012. [2] Long, J., Shelhamer, E., & Darrell, T. (2015). Fully convolutional

networks for semantic segmentation. In Proceedings of the IEEE

Conference on Computer Vision and Pattern Recognition (pp. 3431-3440).

[3] Wang, Naiyan, et al. "Transferring rich feature hierarchies for robust

visual tracking." arXiv preprint arXiv: 1501.04587 (2015). [4] Mao, Junhua, et al. "Deep captioning with multimodal recurrent neural

networks (m-rnn)." arXiv preprint arXiv: 1412.6632 (2014)

[5] Simonyan, Karen, and Andrew Zisserman. "Very deep convolutional networks for large-scale image recognition." arXiv preprint arXiv:

1409.1556 (2014).

[6] Szegedy, Christian, et al. "Going deeper with convolutions." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.

2015.

[7] He, Kaiming, et al. "Deep residual learning for image recognition." Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition. 2016.

[8] Huang, Gao, et al. "Densely connected convolutional networks." arXiv preprint arXiv:1608.06993 (2016).

[9] Sabour, Sara, Nicholas Frosst, and Geoffrey E. Hinton. "Dynamic routing

between capsules." Advances in Neural Information Processing Systems. 2017.

[10] Badrinarayanan, Vijay, Alex Kendall, and Roberto Cipolla. "Segnet: A

deep convolutional encoder-decoder architecture for image segmentation." arXiv preprint arXiv:1511.00561(2015).

[11] Ciresan, Dan, et al. "Deep neural networks segment neuronal membranes

in electron microscopy images." Advances in neural information processing systems. 2012.

[12] Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net:

Convolutional networks for biomedical image segmentation." International Conference on Medical image computing

and computer-assisted intervention. Springer, Cham, 2015.

[13] Çiçek, Özgün, et al. "3D U-Net: learning dense volumetric segmentation from sparse annotation." International Conference on Medical Image

Computing and Computer-Assisted Intervention. Springer International

Publishing, 2016. [14] Milletari, Fausto, Nassir Navab, and Seyed-Ahmad Ahmadi. "V-net:

Fully convolutional neural networks for volumetric medical image

segmentation." 3D Vision (3DV), 2016 Fourth International Conference on. IEEE, 2016.

[15] Yang, Dong, et al. "Automated anatomical landmark detection ondistal

femur surface using convolutional neural network." Biomedical Imaging (ISBI), 2015 IEEE 12th International Symposium on. IEEE, 2015.

[16] Cai, Yunliang, et al. "Multi-modal vertebrae recognition using

transformed deep convolution network." Computerized Medical Imaging and Graphics 51 (2016): 11-19.

[17] Ramesh, N., J-H. Yoo, and I. K. Sethi. "Thresholding based on histogram approximation." IEE Proceedings-Vision, Image and Signal

Processing 142.5 (1995): 271-279.

[18] Sharma, Neeraj, and Amit Kumar Ray. "Computer aided segmentation of medical images based on hybridized approach of edge and region based

techniques." Proceedings of International Conference on Mathematical

gaojing
附注
图20.对肺部分割数据集的R2U-Net性能的定性评估:第一列输入图像,第二列基础事实和R2U-Net的第三列输出。
gaojing
附注
E.计算时间 每个样品的测试计算时间分别显示在表V中,用于视网膜图像,皮肤癌和肺部分割的血管分割。
gaojing
附注
五,结论和未来的工作 在本文中,我们提出了使用递归卷积神经网络和循环残余卷积神经网络的U-Net架构的扩展。所提出的模型分别称为“RU-Net”和“R2U-Net”。这些模型使用医学成像领域中的三种不同应用进行评估,包括视网膜血管分割,皮肤癌病变分割和肺部分割。实验结果表明,与包括U-Net和剩余U-Net(或ResU-Net)在内的现有方法相比,所提出的RU-Net和R2U-Net模型在具有相同网络参数数量的分段任务中表现出更好的性能。所有三个数据集上的模型。此外,结果表明,这些提出的模型不仅确保了培训期间的更好性能,而且还确保了测试阶段的性能。将来,我们希望通过从编码到解码单元的新颖特征融合策略探索相同的架构。
Page 12: Recurrent Residual Convolutional Neural Network based on U ...static.tongtianta.site/paper_pdf/ce8a44d6-37f5-11e... · amplitude segmentation based on histogram features [17], the

Biology', Mathematical Biology Recent Trends by Anamaya Publishers. 2006.

[19] Boykov, Yuri Y., and M-P. Jolly. "Interactive graph cuts for optimal

boundary & region segmentation of objects in ND images." Computer Vision, 2001. ICCV 2001. Proceedings. Eighth IEEE International

Conference on. Vol. 1. IEEE, 2001.

[20] Litjens, Geert, et al. "A survey on deep learning in medical image analysis." arXiv preprint arXiv:1702.05747 (2017).

[21] Greenspan, Hayit, Bram van Ginneken, and Ronald M. Summers. "Guest

editorial deep learning in medical imaging: Overview and future promise of an exciting new technique." IEEE Transactions on Medical

Imaging 35.5 (2016): 1153-1159.

[22] Havaei, Mohammad, et al. "Brain tumor segmentation with deep neural networks." Medical image analysis 35 (2017): 18-31.

[23] G. Brostow, J. Fauqueur, and R. Cipolla, “Semantic object classes in

video: A high-definition ground truth database,” PRL, vol. 30(2), pp. 88– 97, 2009. [23] S. Song, S. P.

[24] Lichtenberg, and J. Xiao, “Sun rgb-d: A rgb-d scene understanding

benchmark suite,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 567–576, 2015.

[25] Kistler, Michael, et al. "The virtual skeleton database: an open access

repository for biomedical research and collaboration." Journal of medical Internet research 15.11 (2013).

[26] He, Kaiming, et al. "Identity mappings in deep residual

networks." European Conference on Computer Vision. Springer International Publishing, 2016.

[27] S. Zheng, S. Jayasumana, B. Romera-Paredes, V. Vineet, Z. Su, D. Du, C. Huang, and P. H. Torr, “Conditional random fields as recurrent neural

networks,” in Proceedings of the IEEE International Conference on

Computer Vision, pp. 1529–1537, 2015. [28] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille,

“Semantic image segmentation with deep convolutional nets and fully

connected CRFs,” in ICLR, 2015. [29] Kendall, Alex, Vijay Badrinarayanan, and Roberto Cipolla. "Bayesian

segnet: Model uncertainty in deep convolutional encoder-decoder

architectures for scene understanding." arXiv preprint arXiv:1511.02680 (2015).

[30] Zhang, Zhengxin, Qingjie Liu, and Yunhong Wang. "Road Extraction by

Deep Residual U-Net." arXiv preprint arXiv:1711.10684 (2017).

[31] Li, Ruirui, et al. "DeepUNet: A Deep Fully Convolutional Network for

Pixel-level Sea-Land Segmentation." arXiv preprint

arXiv:1709.00201 (2017). [32] Kayalibay, Baris, Grady Jensen, and Patrick van der Smagt. "CNN-based

Segmentation of Medical Imaging Data." arXiv preprint

arXiv:1701.03056 (2017). [33] Drozdzal, Michal, et al. "The importance of skip connections in

biomedical image segmentation." International Workshop on Large-

Scale Annotation of Biomedical Data and Expert Label Synthesis. Springer International Publishing, 2016.

[34] Chen, Hao, et al. "Dcan: Deep contour-aware networks for accurate gland

segmentation." Proceedings of the IEEE conference on Computer Vision and Pattern Recognition. 2016.

[35] McKinley, Richard, et al. "Nabla-net: A Deep Dag-Like Convolutional

Architecture for Biomedical Image Segmentation." International Workshop on Brainlesion: Glioma, Multiple Sclerosis, Stroke and

Traumatic Brain Injuries. Springer, Cham, 2016.

[36] Q. Dou, L. Yu, H. Chen, Y. Jin, X. Yang, J. Qin, and P.-A. Heng, “3D

deeply supervised network for automated segmentation of volumetric

medical images,” Medical Image Analysis, vol. 41, pp. 40–54, 2017.

[37] Li, Wenqi, et al. "On the Compactness, Efficiency, and Representation of 3D Convolutional Networks: Brain Parcellation as a Pretext

Task." International Conference on Information Processing in Medical

Imaging. Springer, Cham, 2017. [38] Kamnitsas, Konstantinos, et al. "Efficient multi-scale 3D CNN with fully

connected CRF for accurate brain lesion segmentation." Medical image

analysis 36 (2017): 61-78. [39] Roth, Holger R., et al. "Deeporgan: Multi-level deep convolutional

networks for automated pancreas segmentation." International

Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2015.

[40] Chen, Hao, et al. "Voxresnet: Deep voxelwise residual networks for

volumetric brain segmentation." arXiv preprint arXiv:1608.05895 (2016).

[41] Liang, Ming, and Xiaolin Hu. "Recurrent convolutional neural network for object recognition." Proceedings of the IEEE Conference on

Computer Vision and Pattern Recognition. 2015.

[42] Alom, Md Zahangir, et al. "Inception Recurrent Convolutional Neural Network for Object Recognition." arXiv preprint

arXiv:1704.07709 (2017).

[43] Alom, Md Zahangir, et al. "Improved Inception-Residual Convolutional Neural Network for Object Recognition." arXiv preprint

arXiv:1712.09888 (2017).

[44] Staal, Joes, et al. "Ridge-based vessel segmentation in color images of the retina." IEEE transactions on medical imaging23.4 (2004): 501-509.

[45] Hoover, A. D., Valentina Kouznetsova, and Michael Goldbaum.

"Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response." IEEE Transactions on Medical

imaging 19.3 (2000): 203-210.

[46] Zhao, Yitian, et al. "Automated vessel segmentation using infinite perimeter active contour model with hybrid region information with

application to retinal images." IEEE transactions on medical

imaging 34.9 (2015): 1797-1807. [47] Soares, João VB, et al. "Retinal vessel segmentation using the 2-D Gabor

wavelet and supervised classification." IEEE Transactions on medical

Imaging 25.9 (2006): 1214-1222. [48] Fraz, Muhammad Moazam, et al. "Blood vessel segmentation

methodologies in retinal images–a survey." Computer methods and

programs in biomedicine 108.1 (2012): 407-433. [49] https://challenge2017.isic-archive.com

[50] https://www.kaggle.com/kmader/finding-lungs-in-ct-data/data. [51] Dice, Lee R. "Measures of the amount of ecologic association between

species." Ecology 26.3 (1945): 297-302.

[52] Jaccard, Paul. "The distribution of the flora in the alpine zone." New phytologist 11.2 (1912): 37-50.

[53] Cheng, Erkang, et al. "Discriminative vessel segmentation in retinal

images by fusing context-aware hybrid features." Machine vision and applications 25.7 (2014): 1779-1792.

[54] Azzopardi, George, et al. "Trainable COSFIRE filters for vessel

delineation with application to retinal images." Medical image analysis 19.1 (2015): 46-57.

[55] Roychowdhury, Sohini, Dara D. Koozekanani, and Keshab K. Parhi.

Blood vessel segmentation of fundus images by major vessel extraction

and subimage classification." IEEE journal of biomedical and health

informatics 19.3 (2015): 1118-1128.

[56] Liskowski, Paweł, and Krzysztof Krawiec. "Segmenting Retinal Blood Vessels With Deep Neural Networks." IEEE transactions on medical

imaging 35.11 (2016): 2369-2380.

[57] Li, Qiaoliang, et al. "A cross-modality learning approach for vessel segmentation in retinal images." IEEE transactions on medical

imaging 35.1 (2016): 109-118.

[58] Marín, Diego, et al. "A new supervised method for blood vessel segmentation in retinal images by using gray- level and moment

invariants-based features." IEEE Transactions on medical imaging 30.1

(2011): 146-158. [59] Fraz, Muhammad Moazam, et al. "An ensemble classification-based

approach applied to retinal blood vessel segmentation." IEEE

Transactions on Biomedical Engineering 59.9 (2012): 2538-2548. [60] Fraz, Muhammad Moazam, et al. "Delineation of blood vessels in

pediatric retinal images using decision trees-based ensemble

classification." International journal of computer assisted radiology and

surgery 9.5 (2014): 795-811.

[61] Burdick, Jack, et al. "Rethinking Skin Lesion Segmentation in a

Convolutional Classifier." Journal of digital imaging (2017): 1-6. [62] Codella, Noel CF, et al. "Skin lesion analysis toward melanoma detection:

A challenge at the 2017 international symposium on biomedical imaging

(isbi), hosted by the international skin imaging collaboration (isic)." arXiv preprint arXiv:1710.05006 (2017).

[63] Li, Yuexiang, and Linlin Shen. "Skin Lesion Analysis Towards

Melanoma Detection Using Deep Learning Network." arXiv preprint arXiv:1703.00577 (2017).

[64] Hsu, Roy Chaoming, et al. "Contour extraction in medical images using

initial boundary pixel selection and segmental contour following." Multidimensional Systems and Signal Processing 23.4

(2012): 469-498.

[65] Alom, Md Zahangir, et al. "The History Began from AlexNet: A Comprehensive Survey on Deep Learning Approaches." arXiv preprint

arXiv:1803.01164 (2018).

gaojing
高亮