Next Article in Journal
Development of a Method for Detection of SARS-CoV-2 Nucleocapsid Antibodies on Dried Blood Spot by DELFIA Immunoassay
Previous Article in Journal
A Mutation in CACNA1S Is Associated with Multiple Supernumerary Cusps and Root Maldevelopment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DBE-Net: Dual Boundary-Guided Attention Exploration Network for Polyp Segmentation

1
School of Integrated Circuits, Anhui University, Hefei 230601, China
2
Anhui Engineering Laboratory of Agro-Ecological Big Data, Hefei 230601, China
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(5), 896; https://doi.org/10.3390/diagnostics13050896
Submission received: 1 February 2023 / Revised: 22 February 2023 / Accepted: 23 February 2023 / Published: 27 February 2023

Abstract

:
Automatic segmentation of polyps during colonoscopy can help doctors accurately find the polyp area and remove abnormal tissues in time to reduce the possibility of polyps transforming into cancer. However, the current polyp segmentation research still has the following problems: blurry polyp boundaries, multi-scale adaptability of polyps, and close resemblances between polyps and nearby normal tissues. To tackle these issues, this paper proposes a dual boundary-guided attention exploration network (DBE-Net) for polyp segmentation. Firstly, we propose a dual boundary-guided attention exploration module to solve the boundary-blurring problem. This module uses a coarse-to-fine strategy to progressively approximate the real polyp boundary. Secondly, a multi-scale context aggregation enhancement module is introduced to accommodate the multi-scale variation of polyps. Finally, we propose a low-level detail enhancement module, which can extract more low-level details and promote the performance of the overall network. Extensive experiments on five polyp segmentation benchmark datasets show that our method achieves superior performance and stronger generalization ability than state-of-the-art methods. Especially for CVC-ColonDB and ETIS, two challenging datasets among the five datasets, our method achieves excellent results of 82.4% and 80.6% in terms of mDice (mean dice similarity coefficient) and improves by 5.1% and 5.9% compared to the state-of-the-art methods.

1. Introduction

Colorectal cancer (CRC) is a common digestive tract malignant tumor, which seriously endangers human health, and its incidence ranks third among all cancers [1]. As one of the most important precursors of colorectal cancer, polyps can easily transform into malignant tumors if not treated in time. A colonoscopy is an effective method to detect colon lesions, which can provide doctors with precise positioning information so that doctors can remove them in time before they become cancerous. However, colonoscopy occasionally overlooks cancer-causing polyps due to the inexperience of physicians. Therefore, automatic and accurate segmentation of polyps in colonoscopy images is of great significance for the clinical prevention of colorectal cancer.
At present, the existing methods of polyp segmentation can be divided into two categories, one is traditional methods, and the other is based on deep learning. Traditional polyp segmentation methods usually rely on handcrafted feature-based [2,3,4,5,6,7] methods to identify polyps, such as texture analysis, color distribution, geometric features, and intensity distribution. Though traditional methods have made considerable progress, these handcrafted feature-based methods still have low polyp segmentation accuracy and poor generalization ability, which cannot meet the requirements of clinical practice.
In recent years, with the continuous development of deep learning, many deep learning-based polyp segmentation methods [8,9,10,11,12,13,14,15,16,17,18,19] have been proven to outperform traditional handcrafted feature-based methods. With the proposal of the fully convolutional network FCN, the problem of image segmentation is introduced to the pixel level. Brandao et al. [20] and Vazquez et al. [21] used FCN to segment polyps from colonoscopy images, which achieved better segmentation accuracy than traditional methods. Subsequently, Ronneberger et al. [22] proposed a U-shaped network for biomedical image segmentation, namely U-Net, which is the most widely used network in medical image segmentation. Currently, U-Net++ [23], ResUnet++ [24], and other variant models based on the U-Net network architecture are applied to polyp segmentation, and this encoding–decoding network has become the mainstream network architecture in the medical field.
Recently, attention mechanisms have been increasingly applied to medical segmentation, especially polyp segmentation, which is often used to enhance the blurred boundaries of polyps [8,9,10] or to extract global and local features [25,26,27,28]. PraNet [8] uses an inverse attention mechanism to mine boundary information with a global feature map, which is generated by a parallel decoder that aggregates high-level feature information together. ACSNet [16] designed a spatial attention scheme to pay more attention to uncertain regions, but it did not explicitly model the relationship between ambiguous regions and boundary regions. Guo et al. [29] designed a ternary guidance module. This module suppresses background regions through a ternary guidance mask, thereby mining more polyp regions in uncertain regions without interference from tissues in the background. However, due to its limited ability to explore a wide range of uncertain regions, it does not fully utilize boundary information, and it is difficult to detect fine boundaries. Lai et al. [30] proposed a boundary-guided attention module that can guide the network to adaptively learn multi-scale boundary features for accurate polyp segmentation. However, it cannot explore more detailed information in uncertain regions, resulting in polyp segment segmentation and false positive segmentation. Inspired by the above, we propose a dual bounding-guided attention exploration module DBE, which can not only explore more uncertain regions in a wider range but also explore more detailed information with boundary masks.
For the multi-scale feature adaptation of polyps, there are some related studies dealing with the variation of polyp size and shape. MCNet [31] proposes a multi-scale context-guided deep network for automatic lesion segmentation with endoscopy images of the gastrointestinal tract, where both global and local contexts are captured as guidance for model training. MSNet [15] proposed a multi-scale subtraction network, which uses a subtraction unit (SU) to generate differential features between adjacent layers of the network and equips perceptual units at different levels with different perceptual fields in a feature pyramid manner, thus obtaining richer multi-scale difference information. Sun et al. [32] applied a dilated convolution at the end of the encoder to expand the receptive field to improve the performance of polyps, but it is difficult to capture more contextual information with only a dilated convolution. In view of the above, in our network, we propose a multi-scale context aggregation enhancement module CAM, which first aggregates the context information of different scales, and then adopts the channel feature pyramid to obtain the most extensive features from different receptive regions through multiple dilated convolutions with different dilation rates.
The transformer was originally proposed for the field of natural language processing (NLP). Vaswani et al. [33] first proposed a transformer for machine translation tasks. With the great success of transformers in the NLP field, Dosovitskiy et al. [34] proposed a vision transformer (ViT), which extended the transformer model architecture to the field of computer vision and was the first pure transformer for image classification. ViT can well replace the convolution operation and can still achieve good results in image classification tasks without relying on convolution. Chen et al. proposed a segmentation network TransUnet, that combines CNN and Transformer. TransUnet not only encodes image features as sequences through the transformer to enhance global context but also makes good use of low-level CNN features through a U-shaped hybrid structure design and sets a new record in CT multi-organ segmentation task. Zhang et al. [25] adopted a two-branch parallel architecture TransFuse, which runs a CNN-based encoder and a Transformer-based segmentation network in a parallel manner. Then, the multi-level features of the two branches are efficiently fused by using the proposed fusion module. Wang et al. [26] proposed a simple CNN-free backbone pyramid vision transformer (PVT) for dense prediction tasks. Compared with the design of ViT for image classification, PVT introduces the pyramid structure into the transformer so that it can be applied to various downstream dense prediction tasks. Compared with CNN, PVT performs better in detection and segmentation tasks. This paper adopts PVT as our encoder and designs a new polyp segmentation framework based on it.
In summary, the current polyp segmentation methods based on deep learning have greatly improved the segmentation accuracy and generalization ability compared with traditional methods, but these methods still have shortcomings when facing the challenge of polyp segmentation. In Figure 1, we provide several representative contrast images as an illustration. The first is the problem of blurred polyp boundaries, as shown in Figure 1a. When the colonoscope moves in the intestine, it will cause problems such as motion blur and reflections, which cause blurred boundaries of polyp images and increase the difficulty of polyp segmentation. The second is the multi-scale adaptability of polyps, as shown in Figure 1a–c. Due to the variable size and shape of polyp tissue, the current polyp segmentation methods still have certain limitations in the ability to extract multi-scale features. The third is the close resemblances between polyps and nearby normal tissues, as shown in Figure 1b,c. Polyps have low contrast to the background and high similarity in texture and color to surrounding tissue, which makes accurate identification of polyp areas difficult.
To solve the above problems, this paper proposes a novel and efficient dual boundary-guided attention exploration network (DBE-Net) for accurate polyp segmentation. This network uses the cooperative exploration of uncertainty regions and border regions as the primary driving force for polyp segmentation. First, to deal with the problem of blurred polyp boundaries, we design a dual-boundary guided attention exploration module (DBE), which gradually approximates real polyp boundaries through a coarse-to-fine strategy. DBE is composed of a ternary mask submodule (TM) and a boundary mask submodule (BM). First, we use TM to suppress the background area, balance the foreground area and enhance the uncertainty area, which can explore more polyp areas in a wider uncertainty area and reduce the interference of normal tissues in the background. Second, we utilize BM to guide the network to adaptively learn the perceptual features of the boundary region to achieve re-enhancement of the polyp boundary. Meanwhile, to address the multi-scale feature adaptation problem, we propose a multi-scale context aggregation augmentation module (CAM). This module first aggregates the features of two adjacent layers to obtain the feature information of different levels. Then, the feature maps are convolved with dilated convolutions with different dilation rates through the channel feature pyramid to accommodate the variation in polyp size and shape. In addition, low-level features usually contain rich, detailed information such as texture, color, and edge information. To address the close resemblances between polyps and nearby normal tissues, we propose a low-level detail enhancement module (LDE). This module can extract more low-level details and provide them with more high-level features for fusion after up-sampling operation to promote the overall performance of the network. Meanwhile, this module uses the combination of channel attention and spatial attention to capture polyp details in different dimensions.
Our main contributions can be highlighted as follows:
  • We propose a new dual boundary-guided attention exploration network termed DBE-Net. Different from existing CNN-based methods, this network adapts the pyramid vision transformer as the encoder to obtain more robust features and is combined with the partial decoder PD as the base framework.
  • We propose a dual boundary-guided attention exploration module, termed DBE. This module adopts a coarse-to-fine strategy to gradually approximate the real polyp boundary from coarse uncertain regions to fine boundary regions through ternary masks and boundary masks according to the prediction results of previous scales to obtain a clearer polyp segmentation boundary.
  • We propose a multi-scale context aggregation augmentation module termed CAM. To address the multi-scale feature adaptation problem, this module adopts multiple dilated convolutions with different dilation rates and obtains the most extensive features from different receptive regions to adapt to the multi-scale changes of polyps.
  • To solve the problem of high similarity between polyps and surrounding tissues, we propose a low-level detail enhancement module, termed LDE. This module can extract more low-level details to improve the overall network performance and achieve more accurate polyp segmentation.

2. Materials and Methods

2.1. Proposed Network Structure

The purpose of polyp segmentation is to identify polyp regions from normal tissues in colonoscopy images and to segment polyps robustly. However, the polyp boundaries are irregular, even somewhat blurred, and disturbed by normal tissues in the background, which brings great challenges to improving polyp segmentation performance. Therefore, we propose a dual boundary-guided attention exploration network DBE-Net, which will be used for accurate polyp segmentation.
As shown in Figure 2, the proposed network mainly consists of four parts, including a pyramid vision transformer (PVT) encoder, dual boundary-guided attention exploration module (DBE), multi-scale context aggregation enhancement module (MCA), and low-level detail enhancement module (LDE). In the encoder stage, to obtain more powerful backbone features and provide more foreground information for subsequent decoding stages, PVTv2 is adopted as our encoder. Compared with traditional CNN methods, which pay more attention to local information, PVTv2 exhibits stronger global information extraction ability and better robustness to input disturbances. In the decoder stage, considering that the transformer is good at providing global information but lacks detailed local information, it makes it difficult to solve the multi-scale feature adaptation problem, including size and shape changes. To cope with this problem, MAC can obtain a wider range of features from different receptive regions by aggregating and enhancing the features of each stage to adapt to the multi-scale changes of polyps, thereby obtaining more abundant local and global features. Then, a partial decoder (PD) obtains preliminary rough segmentation results by aggregating features from the outputs of the MAC modules at different stages. However, the preliminary results obtained by the PD module can only obtain significant foreground regions, and it is difficult to judge accurate polyp boundaries for fuzzy boundaries. Therefore, the DBE adopts a coarse-to-fine strategy to segment polyps. According to the preliminary prediction results obtained by PD, the real polyp boundaries are gradually approximated from rough uncertain regions to fine boundary regions by the ternary mask sub-module and the boundary mask sub-module. Low-level features usually contain rich detail information, while the detail fusion sub-module (DF) in LDE can extract more low-level detail information and provide it to high-level features through up-sampling. At the same time, the detail extraction sub-module (DE) can capture the detailed appearance information to promote the overall performance of the network.
In our network model, we use PVTv2 as the backbone network to extract four multi-scale pyramid features Xi from the input image I, where i ∈ {1, 2, 3, 4}. Then, every two adjacent features of these four features are sent to MCA for multi-scale context aggregation enhancement, leading to three output features M2, M3, and M4. Next, we feed the output features M2, M3, and M4 to PD to fuse, resulting in preliminary rough segmentation results D5. At the same time, the coarse segmentation result D5 from PD and low-level features X1 are fed to LDE for detail enhancement to obtain Li, where I ∈ {1, 2, 3, 4}. Then the PD features D5 are fed into the DBE, which is combined with the detail enhancement features L4 from the LDE output to obtain a clearer boundary segmentation through two stages of boundary exploration from coarse to fine. Meanwhile, the output features D4 and T4 of the two stages serve as side supervision. Then, the enhanced features D4 are up-sampled and fed to the next DBE. There they are again combined with the detail enhancement features L3 to further obtain a clearer boundary segmentation and output two-stage side supervision T3 and D3. This process is repeated in detail enhancement features L2 and L1 while obtaining four side supervisions, T2, D2, T1, and D1. Finally, through four stepwise boundary explorations, the final predicted output D1 is obtained.
Our overall process is as follows:
X i = T ( I ) ( i = 1 , 2 , 3 , 4 )
M i = M ( X i , X i 1 ) ( i = 2 , 3 , 4 )
D 5 = P ( M 2 , M 3 , M 4 )
L i = L ( D 5 , X 1 , X 2 , X 3 , X 4 ) ( i = 1 , 2 , 3 , 4 )
D 1 = D ( L 1 , D ( L 2 , D ( L 3 , D ( L 4 , D 5 ) ) ) )
Here: (1) I represents the input image; (2) T, M, P, L, and D represent the functions of PVTv2, MAC, PD, LDE, and DBE, respectively.

2.2. Dual Boundary-Guided Attention Exploration Module

When the colonoscope moves in the intestine, it may produce problems such as motion blur and reflection, which causes blurred boundaries in the polyp image. To pay more attention to the pixels of the blurred boundary and achieve a clearer polyp segmentation boundary, we design a dual boundary-guided attention exploration module DBE, which adopts a coarse-to-fine gradual approximation strategy. First, according to the prediction results of the previous scale and detail enhancement features, more ground-truth polyp regions are obtained from rough uncertain regions through a ternary mask sub-module. Then, the boundary mask sub-module explores the finer polyp boundaries from the fine boundary regions. Finally, multiple DBE operations are used to gradually approach the true polyp boundary. Figure 3 shows the detailed structure of the DBE. This module consists of two sub-modules, namely the ternary mask sub-module TM and the boundary mask sub-module BM.
For the TM sub-module, it divides the feature map into three regions, namely foreground, background, and uncertainty regions. The foreground represents polyp regions with high response, the background represents non-polyp regions with low response, and uncertainty regions refer to fuzzy regions with intermediate responses. The uncertain area includes foreground and background objects and is a mixture of the two. The purpose of the TM sub-module is to mine as many foreground objects as possible from the uncertain region and exclude background objects. Specifically, the ternary mask assigns different weight values to different regions. To emphasize the uncertain area, the pixels in the uncertain area are set to the highest weight of 1. Pixels in the foreground area are set to 0 to balance the highly responsive areas. To suppress the interference of the background area, the pixels in the background area are set to the lowest weight of −1. The specific process is as follows:
As shown in the gray rounded rectangle box in Figure 3, the low-resolution output feature map Di+1 by the previous stage is up-sampled to obtain Du.
D u = U ( D i + 1 )
where U represents the up-sampling operation.
Then, a ternary mask Dt (Equation (7)) is generated based on two thresholds, αl and αh. In Equation (7), the value of i-th pixel of Du is greater than αh, which means it belongs to the foreground region and sets the weight to 0. If it is less than αl, it belongs to the background area and sets the weight to −1. When it is between αl and αh, it belongs to the uncertainty region and sets the weight to 1.
D t i = { 1 , i f   D u i α l 1 , i f   α l D u i α h 0 , i f   D u i α h
Next, the generated ternary mask Dt is element-wise multiplied by the features Li from the current stage to obtain weighted feature maps Lt. Then, Lt is 3 × 3 convoluted and combined with Du through element-wise addition to generating the output feature maps Ti of the TM sub-module. In particular, the feature maps Ti is used both as side supervision and as the input of the BM sub-module.
T i = f i ( L i D t ) + D u
where represents element-wise multiplication. fi denotes 3 × 3 convolution.
For the BM sub-module, it can extract richer and more detailed boundary field information, which can help to explore more accurate polyp edges. To achieve the goal of enhancing boundary features, we extract boundary information from the output feature maps Ti of the previous DM sub-module through morphological operations. As shown in the cyan rounded rectangular box in Figure 3, the output feature maps Ti of the previous sub-module DM is converted into a binary mask Ds with a threshold of 0.5. If the probability of a pixel belonging to a polyp is greater than 0.5, the pixel is classified as a foreground polyp pixel; otherwise, it is classified as a background pixel. Then, we use morphological operations on the binary mask Ds:
D m = D i l a t e ( D s , E ) E r o d e ( D s , E )
where Dilate(Ds, E) denotes the morphological dilation operation on the binary mask Ds using multi-structure elements E to generate a dilated mask Dd. Erode(Ds, E) denotes the morphological erosion operation on the binary mask Ds using multi-structure elements E to generate an erosion mask De. The morphological dilation operation dilates the polyp border outward, thereby expanding the foreground area. The morphological erosion operation shrinks the polyp boundary, thereby reducing the foreground area.
Similar to [36], we also consider multi-structure elements. Let (m, n) be the center of the binary mask Ds, and a structural element in the (2N + 1) × (2N + 1) square window can be represented as:
E i = { M ( m + m 0 , n + n 0 ) , θ i = i × α | N m 0 , n 0 N }
where i = 0, 1, …, 4N − 1, α = 180°/4N, and θi is the direction angle of the structure element.
In this paper, for the two DBE modules at higher levels, we choose N = 2, and then in a 5 × 5 square window, the direction angles of all structure elements are 0°, 22.5°, 45°, 67.5°, 90°, 112.5°, 135°, and 157.5°. For the two DBE modules at the lower level, we choose N = 3, and then in the 7 × 7 square window, the direction angles of all structure elements are 0°, 15°, 30°, 45°, 60°, 75°, 90°, 105°, 120°, 135°, 150°, and 165°. They contain most of the directions in which the lines extend in the image. In morphological dilation, each structural element Ei in different directions was used to dilate the image, and the averaged result is taken as the dilated mask. Morphological erosion operation is similar. Then, we perform an element-wise subtraction operation on Dd and De to generate the boundary mask Dm.
Next, the generated ternary mask Dm is element-wise multiplied by the input features Li from the previous TM sub-module to obtain boundary-enhanced feature maps Lm. Then, Lm is 3 × 3 convoluted and combined with Ti through element-wise addition to generating the output feature maps Di of the DM sub-module. Meanwhile, Di also serves as the final output of the DBE module. In particular, the feature maps Di is used both as side supervision and as the input of the next DBE module.
D i = f i ( L i D m ) + T i

2.3. Multi-Scale Context Aggregation Enhancement Module

Due to the variable size and shape of polyp tissue, current polyp segmentation methods still have certain limitations in their ability to extract multi-scale features. To address the multi-scale feature adaptation problem of polyps, we introduce a multi-scale context aggregation augmentation module MCA, which can obtain more detailed local and global feature information, as shown in Figure 4. This module complements the location and spatial information through the interaction of adjacent high- and low-layer feature information. At the same time, this module uses the channel feature pyramid (CFP) [37] module to obtain the most extensive multi-scale feature information from different receptive regions to adapt to the multi-scale variation of polyps.
Specifically, we use 1 × 1 convolution on two feature maps from different layers to reduce feature channels to 32 to reduce computational resources. The features from lower layers are down-sampled and connected with the features from higher layers as the input of CFP. The CFP module has four branches, and the dilate rates of different branches are set to (1, 2, 4, 8) to obtain more contextual information. Each branch contains three groups of convolution blocks. Each group of convolution blocks decomposes the standard convolution into 3 × 1 convolution and then performs 1 × 3 convolution. Then, with this factorized convolution [37], 33% of the parameters can be saved with the same number of filters. Finally, the features of the four branches are concatenated together as the output of the MCA module.

2.4. Low-Level Detail Enhancement Module

The appearance of polyps is usually very similar to the surrounding normal tissue; however, low-level feature maps usually contain more detailed information, such as texture, borders, and color. Therefore, we design a low-level detail enhancement module LDE to extract richer detailed information. As shown in Figure 5, the LDE module consists of two parts, namely the detail fusion sub-module DF and the detail extraction sub-module DE.
First, the DF sub-module reduces the interference of background information by the preliminary segmentation result D5 from the PD module. Then, more detailed information is provided on high-level features for fusion through down-sampling to improve the performance of the overall network and improve the accuracy of polyp segmentation. As shown in the grey rounded rectangle box in Figure 5, the features D5 from the PD module are transformed into a binary mask M with a threshold α. If the probability of belonging to a polyp pixel is less than α, it is classified as a background pixel with a weight of 0. Where α is set to 0.2.
M i = { 0 , i f   X 1 i α 1 , i f   X 1 i α
Next, the generated mask M is element-wise multiplied by the low-level features X1 to obtain feature maps X 1 with reduced background noise. Then, X 1 is up-sampled and concatenated with X2, X3, and X4, respectively, to generate detail enhancement feature maps L2, L3, and L4.
X 1 = X 1 M
For the DE sub-module, we mainly adopt the tandem use of channel attention operation Ac( · ) and spatial attention operation As( · ) to capture detailed information on polyps from different dimensions.
L 1 = A s ( A c ( X 1 ) )
The channel attention module formula is as follows:
A c ( F ) = σ ( M L P ( P a v g ( F ) ) + M L P ( P m a x ( F ) ) ) F
where Pavg and Pmax represent the max pooling and average pooling functions, respectively. For the input feature map F, the Pavg and Pmax functions are first applied at each spatial location, resulting in two C × 1 × 1 feature maps. Then each is fed into a shared MLP containing two fully connected layers. Finally, the channel attention map is obtained by adding pixels and passing through an activation function.
The spatial attention module formula is as follows:
A s ( F ) = σ ( f 7 × 7 ( C o n c a t ( P a v g ( F ) , P m a x ( F ) ) ) ) F
where the feature maps output by the channel attention operation is taken as the input of the spatial attention operation. First, the Pavg and Pmax functions are applied at each channel, resulting in two C × 1 × 1 feature maps. Then, they are connected by channel and fed into a 7 × 7 convolutional layer. Finally, the spatial attention map is obtained through the activation function.

2.5. Loss Function

The DBE-Net is trained via a deep supervision manner. There are a total of 9 aspects of supervision across all stages of the network (Figure 1), including Di (i = 1, 2, 3, 4, 5) and Ti (i = 1, 2, 3, 4, 5). In the experiments, we use the weighted binary cross entropy (Lwbce) loss and the weighted intersection over union (Lwiou) loss to supervise each Di and Ti. These two losses restrict the prediction map from the global object perspective and local pixel perspective, respectively. The total loss can be expressed as:
L = L m a i n + L a u x
where Lmain and Laux represent the main loss and auxiliary loss, respectively. The main loss Lmain is to calculate the loss between the feature map Di and the ground truth G, which can be expressed as follows:
L m a i n = i = 1 5 ( L w b c e ( D i , G ) + L w i o u ( D i , G ) )
The auxiliary loss Laux is to calculate the loss between the feature map Ti and the ground truth G, which can be expressed as follows:
L a u x = i = 1 4 ( L w b c e ( T i , G ) + L w i o u ( T i , G ) )

2.6. Datasets

To evaluate the segmentation performance of our proposed method, we conduct experiments on five polyp segmentation data, including Kvasir-SEG [38], CVC-ClinicDB [39], CVC-ColonDB [35], ETIS [21] and CVC-T [5]. The description of these datasets is as follows.
(1)
Kvasir-SEG. The dataset consists of 1000 polyp images and their corresponding ground truth polyp masks annotated by endoscopy experts. The resolution of the images contained in Kvasir-SEG varies from 332 × 487 to 1920 × 1072 pixels.
(2)
CVC-ClinicDB. The dataset, also known as CVC-612, contains 612 open-access images from 25 colonoscopy videos with a resolution of 384 × 288.
(3)
CVC-ColonDB. The dataset consists of 380 images from 15 colonoscopy videos with an image resolution of 574 × 500.
(4)
ETIS. The dataset contains 196 images captured from 34 colonoscopy videos with an image resolution of 1225 × 996.
(5)
CVC-T. This dataset is a subset of EndoScene and contains 60 images from 44 colonoscopy sequences of 36 patients with an image resolution of 574 × 500.
For a fair comparison, we use the same training and test dataset splits as [8,17,18], as shown in Table 1. A total of 1450 training images are used in the training set, of which 900 images are from Kvasir-SEG, and 550 images are from CVC-ClinicDB. A total of 798 test images are used in the test set, of which 100 images are from Kvasir-SEG, and 62 images are from CVC-ClinicDB. Meanwhile, the other three datasets are used for testing, including 380, 196, and 62 images in CVC-ColonBD, ETIS, and CVC-T, respectively.

3. Results

In this section, we employ three kinds of experiments to validate the performance of DBENet, including qualitative and quantitative experiments and an ablation study. In the experiments, we compare our method with six different methods, including state-of-the-art methods, including UNet [22], UNet++ [23], SFA [40], PraNet [8], SANet [17] and CaraNet [18]. Meanwhile, we analyze and compare three evaluation metrics on five different polyp segmentation datasets. For a fair comparison, the segmentation maps for these methods are published by the authors or directly generated from the original code provided by the authors.

3.1. Training Settings and Evaluation Metrics

Our proposed model is implemented in Pytorch with NVIDIA RTX3090 for GPU acceleration. Before feeding into the network, we resize the input image to 352 × 352. Instead of data augmentation, we use a multi-scale training strategy {0.75, 1.0, 1.25}. We start the training with a batch size of 16 and adopt the Adam optimization method to optimize the model parameters. The initial learning rate is set to 0.0001, and the maximum epoch is set to 100. Thresholds αl and αh are set to 0.3 and 0.7, respectively.
In this paper, the mean dice similarity coefficient (mDice), the mean intersection over union (mIoU), and the mean absolute error (MAE) were used to quantitatively evaluate the segmentation performance of polyps. For these evaluation metrics, higher values of mDice and mIoU indicate better segmentation results, while lower values of MAE indicate better segmentation results.
The Dice coefficient is an ensemble similarity measure and a standard measure for comparing pixel-level classification results between predicted masks and ground truth values. The mDice represents the average of the dice similarity coefficients between all predicted images and the ground truth, and the formula is as follows.
m D i c e = 1 n + 1 i = 1 n 2 T P F P + 2 T P + F N
where TP, FP, FN, and n denote true positive, false positive, and false negative predictions and the number of test images, respectively.
The IoU calculates the overlap ratio between the predicted mask and its corresponding ground truth value. The mIoU represents the average of the IoU values between all predicted images and the ground truth, and the formula is as follows.
m I o U = 1 n + 1 i = 1 n T P F P + T P + F N
MAE is used to evaluate pixel-level accuracy. We evaluate pixel-level accuracy by computing the mean absolute error between the model’s predicted and true values, and the formula is as follows.
M A E = 1 n i = 1 n | p ^ i p i |
where p ^ i and p i denote the prediction and the corresponding ground truth for the i-th pixel out of the total n pixels.

3.2. Qualitative Analysis

Figure 6 shows a visual comparison between our model DBE-Net and different competing methods. Apparently, our method can segment polyp regions more accurately and outperform several challenging aspects compared to other competing methods. First, DBE-Net can obtain clearer polyp boundaries and reduce the interference of complex backgrounds such as intestinal folds and dim areas. For example, intestinal folds were mistakenly missed by other methods in Figure 6e, and polyps in dim regions were missed or misjudged by other methods in Figure 6a. In both cases, DBE-Net can accurately detect polyps, which may be attributed to the role of the DBE module. This module gradually enhances the fuzzy edge information through a coarse-to-fine strategy to achieve the purpose of approximating the real polyp boundary. Second, DBE-Net can effectively deal with the multi-scale changes of polyps and identify large polyps and small polyps of different shapes, especially to avoid misjudgment of small polyps. As shown in Figure 6a–c, only our method can accurately detect small polyps and reduce false positives. Third, DBE-Net can also better solve the problem of high similarity between polyps and surrounding tissues. As shown in Figure 6d,e, only our method can segment the polyp region more completely.

3.3. Quantitative Experiments

Table 2 and Table 3 present the quantitative results of the statistical comparison of DBE-Net with six different competing methods on five different datasets. For ease of viewing, the best results for each evaluation metric are highlighted in bold. Considering that the training set is selected from CVC-ClinicDB and Kvasir-SEG, the learning ability of the proposed method is first analyzed quantitatively. As shown in Table 2, our proposed method is optimal on various metrics on both datasets. The mDice of our method achieves excellent results of 93.9% and 92.2% in CVC-ClinicDB and Kvasir-SEG.
To verify the generalization abilities of the model, we test it on three unseen datasets (CVC-ColonDB, ETIS, and CVC-T). As shown in Table 3, the results of the proposed method are still optimal and achieve good generalization performance compared to the six competing methods. Especially, as shown in Figure 7, the generalization ability of our method achieves a significant improvement on the challenging datasets CVC-ColonDB and ETIS. On the CVC-ColonDB dataset, the mDice of our method outperforms CaraNet and SANet by 5.1% and 7.1%, respectively. In terms of the ETIS dataset, the mDice of our method outperforms CaraNet and SANet by 5.9% and 5.6%, respectively. On the CVC-T dataset, the mDice of our method surpasses SANet by 1.5%.

3.4. Ablation Studies

To verify the effectiveness of each component in the model, we conduct ablation experiments on the dual boundary-guided attention exploration module (DBE), multi-scale context aggregation enhancement module (MCA), and low-level detail enhancement module (LDE). The baseline of our backbone network is composed of PVTv2 and PD, and the standard model is composed of “Baseline + MCA + LDE + DBE”. We evaluate the effectiveness of different modules by removing or changing them from the standard model. “w/o MCA”, “w/o LDE,” and “w/o DBE” represent the removal of MCA, LDE, or DBE, respectively, from the standard model.
Ablation studies of DBE: To investigate the effectiveness of the DBE module, we first visualize the output characteristics of the DBE module at different stages. As shown in Figure 8, the redder the area, the more attention the network needs to give. We can observe that it gradually refines the polyp boundary and explores more polyp information as the stage progresses. This shows that the coarse-to-fine strategy of the DBE module can effectively enhance the fuzzy edge information. To quantitatively analyze the effectiveness of the DBE module, we train a version “w/o DBE” that lacks the DBE module. We removed the DBE module altogether and replaced it with an element-wise addition operation. The experimental results are shown in Table 4. Compared to the standard model, the performance of the model without DBE dropped sharply on the five datasets. As shown in the dumbbell diagram in Figure 9, on CVC-ClinicDB, Kvasir-SEG, CVC-ColonDB, ETIS, and CVC-T, compared to the performance of the standard model DBENet, the absence of DBE reduces mDice by 0.8%, 0.6%, 1.7%, 1.6%, and 2.1%, respectively.
Ablation studies of MCA: Similarly, to quantitatively analyze the effectiveness of the MCA module, we train a version “w/o MCA” that lacks the MCA module. We remove all MCA modules and directly fuse the context information of the three high layers (i.e., X2, X3, and X4). The results in Table 4 show that compared with the standard model, the model lacking MCA has a 1.0% and 0.5% drop in mDice on the CVC-ColonDB and ETIS datasets, respectively. In particular, the CVC-ColonDB and ETIS datasets are challenging unseen datasets, which suggests that MCA is helpful for multi-scale generalization.
Ablation studies of LDE: To demonstrate the power of LDE, we train a version “w/o LDE” that lacks the LDE module. As shown in Table 4, removing the LDE results in a slight drop in performance on all five datasets compared to the standard model. On the five datasets, CVC-ClinicDB, Kvasir-SEG, CVC-ColonDB, ETIS, and CVC-T, removing LDE can reduce mDice by 0.5%, 0.4%, 0.6%, 0.1%, and 1.0%, respectively. Therefore, the LDE module plays a role in improving the overall segmentation ability of the model.

4. Discussion

Colonoscopy is an effective way to detect colon lesions. Automatic segmentation of polyp regions can help doctors improve the accuracy of diagnosis. The current polyp segmentation research based on deep learning still has the following problems: blurry polyp boundaries, multi-scale adaptability of polyps, and close resemblances between polyps and nearby normal tissues. For the fuzzy polyp boundary problem, we use a coarse-to-fine strategy to gradually enhance the true polyp boundary. As can be seen from Figure 8, as the stage goes on, the polyp boundary becomes clearer, and the background interference is more suppressed, which shows that our strategy can effectively enhance the blurred edge information. For the multi-scale adaptability of polyps, the proposed MCA module first supplements position and spatial information by fusing adjacent high- and low-layer feature information and then uses dilated convolutions with different dilation rates to obtain the most extensive multi-scale feature information from different receptive areas to accommodate multi-scale changes in polyps. As shown in Figure 6, our model can effectively identify polyps of different shapes and sizes, especially for the misjudgment of small polyps. For the similarity between polyps and adjacent normal tissues, the proposed LDE module makes full use of the low-level features that usually contain more detailed information, such as texture, borders, and color. We reduce the interference of background information in the low-level features through the initial segmentation result D5 and then capture the detailed information of polyps from different dimensions through the tandem use of channel attention operation and spatial attention operation. From the results of the ablation experiments in Table 4, it can be concluded that the LDE module promotes the overall performance of the model.
The proposed DBE-Net model achieves superior results compared to state-of-the-art methods on all five public polyp segmentation benchmark datasets. From the qualitative results, as shown in Figure 6, it is obvious that the proposed model can obtain sharper segmentation masks while reducing the complex background interference. For polyps of different shapes and sizes, our model can still identify them effectively. Compared with CaraNet, the proposed model improved the mDice by 0.3% and 0.4% on CVC-ClinicDB and Kvasir-SEG datasets, respectively. Test results on three unseen datasets show that our model exhibits a stronger generalization ability. Especially for CVC-ColonDB and ETIS, two challenging datasets among the five datasets, our method achieves excellent results of 82.4% and 80.6% on mDice and improves by 5.1% and 5.9% compared with CaraNet, respectively. This result is significantly ahead of the current state-of-the-art methods.
In this paper, the presented study still has some limitations. We speculate that more data enhancement processing of the original dataset could further improve the efficiency of the model, including increasing the size of larger datasets, applying more image enhancement techniques, and image post-processing steps. The DBE-Net model allows us to design a very deep and clever neural network architecture to achieve more accurate polyp segmentation. However, DBE-Net should not only be limited to the polyp segmentation domain but also can be extended to other medical image segmentation, which will help us to further promote the generalization ability of the proposed model. Furthermore, the proposed model can also be extended to natural image segmentation and other pixel classification tasks. We used all our experience and knowledge to optimize the model, but there may be further optimizations, which may affect the performance of the method.

5. Conclusions

In this paper, we propose a new polyp segmentation framework, DBE-Net, to help address three major challenges in the field of polyp segmentation: blurry polyp boundaries, multi-scale adaptability of polyps, and close resemblances between polyps and nearby normal tissues. To address the blurred boundary problem, we introduce the dual boundary-guided attention exploration module DBE. This module mainly adopts a coarse-to-fine strategy to gradually approach the real polyp boundary. Meanwhile, the proposed network introduces a multi-scale context aggregation enhancement module MCA and a low-level detail enhancement module LDE to help address the problem of polyp multi-scale adaptability and the similarity of polyps to surrounding tissues. Extensive qualitative and quantitative experiments demonstrate that DBE-Net outperforms current state-of-the-art models. For performance analysis on unseen datasets, it is found that the proposed network has a higher generalization ability than other state-of-the-art methods. Especially for the unseen CVC-ColonDB dataset and ETIS dataset, the mDice of the proposed model reaches 82.4% and 80.6%. This result is significantly ahead of current state-of-the-art models. We can conclude that the proposed DBE-Net model can accurately segment polyps and can be considered a powerful polyp segmentation method. Further, we can develop a stable and clinically applicable system for polyp segmentation from colonoscopy video frames.

Author Contributions

Conceptualization, H.M. and C.X.; methodology, H.M.; software, H.M. and C.N.; validation, J.H., Y.L. and H.M.; data curation, C.L. and J.H.; writing—original draft preparation, H.M.; writing—review and editing, C.X.; visualization, C.N.; supervision, C.X.; project administration, C.X.; funding acquisition, C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (No. 2019YFC0117800).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this research are all benchmark datasets and are publicly available for researchers.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Goding Sauer, A.; Fedewa, S.A.; Butterly, L.F.; Anderson, J.C.; Cercek, A.; Smith, R.A.; Jemal, A. Colorectal cancer statistics, 2020. CA Cancer J. Clin. 2020, 70, 145–164. [Google Scholar] [CrossRef] [Green Version]
  2. Alexandre, L.A.; Nobre, N.; Casteleiro, J. Color and position versus texture features for endoscopic polyp detection. In Proceedings of the 2008 International Conference on BioMedical Engineering and Informatics, Sanya, China, 27–30 May 2008; pp. 38–42. [Google Scholar]
  3. Iwahori, Y.; Shinohara, T.; Hattori, A.; Woodham, R.J.; Fukui, S.; Bhuyan, M.K.; Kasugai, K. Automatic Polyp Detection in Endoscope Images Using a Hessian Filter. In Proceedings of the MVA; IEEE: New York, NY, USA, 2013; pp. 21–24. [Google Scholar]
  4. Mamonov, A.V.; Figueiredo, I.N.; Figueiredo, P.N.; Tsai, Y.-H.R. Automated polyp detection in colon capsule endoscopy. IEEE Trans. Med. Imaging 2014, 33, 1488–1502. [Google Scholar] [CrossRef] [Green Version]
  5. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef]
  6. Fiori, M.; Musé, P.; Sapiro, G. A complete system for candidate polyps detection in virtual colonoscopy. Int. J. Pattern. Recognit. Artif. Intell. 2014, 28, 1460014. [Google Scholar] [CrossRef] [Green Version]
  7. Nawarathna, R.; Oh, J.; Muthukudage, J.; Tavanapong, W.; Wong, J.; De Groen, P.C.; Tang, S.J. Abnormal image detection in endoscopy videos using a filter bank and local binary patterns. Neurocomputing 2014, 144, 70–91. [Google Scholar] [CrossRef] [Green Version]
  8. Fan, D.-P.; Ji, G.-P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 263–273. [Google Scholar]
  9. Kim, T.; Lee, H.; Kim, D. UACANet: Uncertainty Augmented Context Attention for Polyp Segmentation. In Proceedings of the 29th ACM International Conference on Multimedia, Chendu, China, 20–24 October 2021; pp. 2167–2175. [Google Scholar]
  10. Guo, X.; Yang, C.; Liu, Y.; Yuan, Y. Learn to Threshold: ThresholdNet With Confidence-Guided Manifold Mixup for Polyp Segmentation. IEEE Trans. Med. Imaging 2021, 40, 1134–1146. [Google Scholar] [CrossRef]
  11. Wu, H.; Zhao, Z.; Zhong, J.; Wang, W.; Wen, Z.; Qin, J. PolypSeg+: A Lightweight Context-Aware Network for Real-Time Polyp Segmentation. IEEE Trans. Cybern 2022, 1–12. [Google Scholar] [CrossRef]
  12. Ta, N.; Chen, H.; Lyu, Y.; Wu, T. BLE-Net: Boundary learning and enhancement network for polyp segmentation. Multimed. Syst. 2022. [Google Scholar] [CrossRef]
  13. Alam, S.; Tomar, N.K.; Thakur, A.; Jha, D.; Rauniyar, A. Automatic polyp segmentation using u-net-resnet50. ar**v 2020. [Google Scholar] [CrossRef]
  14. Nguyen, N.-Q.; Lee, S.-W. Robust Boundary Segmentation in Medical Images Using a Consecutive Deep Encoder-Decoder Network. IEEE Access 2019, 7, 33795–33808. [Google Scholar] [CrossRef]
  15. Zhao, X.; Zhang, L.; Lu, H. Automatic polyp segmentation via multi-scale subtraction network. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 120–130. [Google Scholar]
  16. Zhang, R.; Li, G.; Li, Z.; Cui, S.; Qian, D.; Yu, Y. Adaptive Context Selection for Polyp Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 253–262. [Google Scholar]
  17. Wei, J.; Hu, Y.; Zhang, R.; Li, Z.; Zhou, S.K.; Cui, S. Shallow attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 699–708. [Google Scholar]
  18. Lou, A.; Guan, S.; Loew, M. Caranet: Context axial reverse attention network for segmentation of small medical objects. ar**v 2021. [Google Scholar] [CrossRef]
  19. Guo, X.; Chen, Z.; Liu, J.; Yuan, Y. Non-equivalent images and pixels: Confidence-aware resampling with meta-learning mixup for polyp segmentation. Med. Image Anal. 2022, 78, 102394. [Google Scholar] [CrossRef] [PubMed]
  20. Armato, S.G.; Petrick, N.A.; Brandao, P.; Mazomenos, E.; Ciuti, G.; Caliò, R.; Bianchi, F.; Menciassi, A.; Dario, P.; Koulaouzidis, A.; et al. Fully convolutional neural networks for polyp segmentation in colonoscopy. In Proceedings of the Medical Imaging 2017: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2017. [Google Scholar]
  21. Vazquez, D.; Bernal, J.; Sanchez, F.J.; Fernandez-Esparrach, G.; Lopez, A.M.; Romero, A.; Drozdzal, M.; Courville, A. A Benchmark for Endoluminal Scene Segmentation of Colonoscopy Images. J. Health Eng. 2017, 2017, 4037190. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  23. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. DLMIA ML-CDS 2018 2018. Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11045, pp. 3–11. [Google Scholar] [CrossRef] [Green Version]
  24. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H.D. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  25. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing transformers and cnns for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 14–24. [Google Scholar]
  26. Dong, B.; Wang, W.; Fan, D.-P.; Li, J.; Fu, H.; Shao, L. Polyp-pvt: Polyp segmentation with pyramid vision transformers. ar**v 2021. [Google Scholar] [CrossRef]
  27. Li, S.; Sui, X.; Luo, X.; Xu, X.; Liu, Y.; Goh, R. Medical image segmentation using squeeze-and-expansion transformers. ar**v 2021. [Google Scholar] [CrossRef]
  28. Wang, J.; Huang, Q.; Tang, F.; Meng, J.; Su, J.; Song, S. Stepwise Feature Fusion: Local Guides Global. ar**v 2022. [Google Scholar] [CrossRef]
  29. Guo, Q.; Fang, X.; Wang, L.; Zhang, E. Polyp Segmentation of Colonoscopy Images by Exploring the Uncertain Areas. IEEE Access 2022, 10, 52971–52981. [Google Scholar] [CrossRef]
  30. Lai, H.; Luo, Y.; Zhang, G.; Shen, X.; Li, B.; Lu, J. Toward accurate polyp segmentation with cascade boundary-guided attention. Vis. Comput. 2022. [Google Scholar] [CrossRef]
  31. Wang, S.; Cong, Y.; Zhu, H.; Chen, X.; Qu, L.; Fan, H.; Zhang, Q.; Liu, M. Multi-Scale Context-Guided Deep Network for Automated Lesion Segmentation With Endoscopy Images of Gastrointestinal Tract. IEEE J. Biomed. Health Inf. 2021, 25, 514–525. [Google Scholar] [CrossRef]
  32. Sun, X.; Zhang, P.; Wang, D.; Cao, Y.; Liu, B. Colorectal polyp segmentation by u-net with dilation convolution. In Proceedings of the 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; pp. 851–858. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  34. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. ar**v 2020. [Google Scholar] [CrossRef]
  35. Tajbakhsh, N.; Gurudu, S.R.; Liang, J. Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 2015, 35, 630–644. [Google Scholar] [CrossRef] [PubMed]
  36. Lin, H.; Lu, H.; Zhang, L.J.U. 6th world congress on intelligent control and automation. 2006, 10, 9988–9991. IEEE 2006, 10, 9988–9991. [Google Scholar]
  37. Lou, A.; Loew, M. Cfpnet: Channel-wise feature pyramid for real-time semantic segmentation. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1894–1898. [Google Scholar]
  38. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; Lange, T.d.; Johansen, D.; Johansen, H.D. Kvasir-seg: A segmented polyp dataset. In Proceedings of the International Conference on Multimedia Modeling, Daejeon, Republic of Korea, 5–8 January 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 451–462. [Google Scholar]
  39. Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; Gil, D.; Rodríguez, C.; Vilariño, F.; graphics. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput. Med. Imaging Graph. 2015, 43, 99–111. [Google Scholar] [CrossRef] [PubMed]
  40. Fang, Y.; Chen, C.; Yuan, Y.; Tong, K.-Y. Selective feature aggregation network with area-boundary constraints for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; pp. 302–310. [Google Scholar]
Figure 1. The segmentation examples of our model and CaraNet [18] with different challenge cases: (a) blurred boundaries of polyps, (ac) multi-scale adaptability of polyps, (b,c) close resemblances between polyps and nearby normal tissues. These images are from unseen datasets CVC-ColonDB [35] and ETIS [21], which show that our model has better generalization ability.
Figure 1. The segmentation examples of our model and CaraNet [18] with different challenge cases: (a) blurred boundaries of polyps, (ac) multi-scale adaptability of polyps, (b,c) close resemblances between polyps and nearby normal tissues. These images are from unseen datasets CVC-ColonDB [35] and ETIS [21], which show that our model has better generalization ability.
Diagnostics 13 00896 g001
Figure 2. Overall architecture of the proposed DBE-Net, which consists of a pyramid vision transformer (PVT) encoder, dual boundary-guided attention exploration module (DBE), multi-scale context aggregation enhancement module (MCA), and low-level detail enhancement module (LDE).
Figure 2. Overall architecture of the proposed DBE-Net, which consists of a pyramid vision transformer (PVT) encoder, dual boundary-guided attention exploration module (DBE), multi-scale context aggregation enhancement module (MCA), and low-level detail enhancement module (LDE).
Diagnostics 13 00896 g002
Figure 3. The structure of DBE, which consists of two sub-modules, including the ternary mask sub-module and the boundary mask sub-module. It adopts a coarse-to-fine strategy to gradually approximate the real polyp boundary.
Figure 3. The structure of DBE, which consists of two sub-modules, including the ternary mask sub-module and the boundary mask sub-module. It adopts a coarse-to-fine strategy to gradually approximate the real polyp boundary.
Diagnostics 13 00896 g003
Figure 4. The structure of MCA. It integrates the feature information of adjacent high and low layers and adopts four parallel dilated convolution branches to extract multi-scale features.
Figure 4. The structure of MCA. It integrates the feature information of adjacent high and low layers and adopts four parallel dilated convolution branches to extract multi-scale features.
Diagnostics 13 00896 g004
Figure 5. The structure of LDE, which consists of two sub-modules, including the detail fusion sub-module and the detail extraction sub-module. It is used to extract richer low-level details.
Figure 5. The structure of LDE, which consists of two sub-modules, including the detail fusion sub-module and the detail extraction sub-module. It is used to extract richer low-level details.
Diagnostics 13 00896 g005
Figure 6. Visualization comparison between the proposed method and six state-of-the-art ones.
Figure 6. Visualization comparison between the proposed method and six state-of-the-art ones.
Diagnostics 13 00896 g006
Figure 7. Evaluation of model generalization ability. We provide the mDice results on CVC-ColonDB, ETIS, and CVC-T.
Figure 7. Evaluation of model generalization ability. We provide the mDice results on CVC-ColonDB, ETIS, and CVC-T.
Diagnostics 13 00896 g007
Figure 8. Visualization of output features of the DBE module at different stages. The redder the region, the more attention the network needs to pay.
Figure 8. Visualization of output features of the DBE module at different stages. The redder the region, the more attention the network needs to pay.
Diagnostics 13 00896 g008
Figure 9. Comparison of mDice results between standard model DBE-Net and “wo/DBE” lacking the DBE module.
Figure 9. Comparison of mDice results between standard model DBE-Net and “wo/DBE” lacking the DBE module.
Diagnostics 13 00896 g009
Table 1. The colorectal polyp datasets used in our experiments.
Table 1. The colorectal polyp datasets used in our experiments.
DatasetImage SizeImage NumberNumber of Train SamplesNumber of Test Samples
Kvasir-SEGVariable1000900100
CVC-ClinicDB384 × 28861255062
CVC-ColonDB574 × 5003800380
ETIS1225 × 9661960196
CVC-T574 × 50060060
Table 2. Quantitative comparison of the learning abilities on CVC-ClinicDB and Kvasir-SEG. The best results are shown in bold.
Table 2. Quantitative comparison of the learning abilities on CVC-ClinicDB and Kvasir-SEG. The best results are shown in bold.
MethodCVC-ClinicDBKvasir-SEG
mDicemIouMAEmDicemIouMAE
U-Net0.8230.7550.0190.8180.7460.055
UNet++0.7940.7290.0220.8210.7430.048
SFA0.7000.6070.0420.7230.6110.075
PraNet0.8990.8490.0090.8980.8400.030
SANet0.9160.8590.0120.9040.8470.028
CaraNet0.9360.8870.0070.9180.8650.023
Ours0.9390.8950.0060.9220.8710.022
Table 3. Quantitative comparison of the generalization abilities on CVC-ColonDB, ETIS, and CVC-T. The best results are shown in bold.
Table 3. Quantitative comparison of the generalization abilities on CVC-ColonDB, ETIS, and CVC-T. The best results are shown in bold.
MethodCVC-ColonDBETISCVC-T
mDicemIouMAEmDicemIouMAEmDicemIouMAE
U-Net0.5120.4440.0610.3980.3350.0360.7100.6270.022
UNet++0.4830.4100.0640.4010.3440.0350.7070.6240.018
SFA0.4690.3470.0940.2970.2170.1090.4670.3290.065
PraNet0.7120.6400.0450.6280.5670.0310.8710.7970.010
SANet0.7530.6700.0430.7500.6540.0150.8880.8150.008
CaraNet0.7730.6890.0420.7470.6720.0170.9030.8380.007
Ours0.8240.7430.0270.8060.7250.0150.9030.8370.006
Table 4. Quantitative results of ablation studies of our method under different configurations.
Table 4. Quantitative results of ablation studies of our method under different configurations.
DatasetCVC-ClinicDBKvasir-SEGCVC-ColonDBETISCVC-T
MetricmDicemIoumDicemIoumDicemIoumDicemIoumDicemIou
Baseline0.9290.8790.9130.8580.7980.7160.7920.7090.8910.820
w/o MCA0.9390.8920.9320.8850.8140.7320.8010.7230.9030.836
w/o LDE0.9340.8850.9180.8670.8180.7350.8050.7230.8930.825
w/o DBE0.9310.8830.9160.8650.8070.7260.7900.7040.8820.812
Baseline +
MCA + LDE + DBE
0.9390.8950.9220.8710.8240.7430.8060.7250.9030.837
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, H.; Xu, C.; Nie, C.; Han, J.; Li, Y.; Liu, C. DBE-Net: Dual Boundary-Guided Attention Exploration Network for Polyp Segmentation. Diagnostics 2023, 13, 896. https://doi.org/10.3390/diagnostics13050896

AMA Style

Ma H, Xu C, Nie C, Han J, Li Y, Liu C. DBE-Net: Dual Boundary-Guided Attention Exploration Network for Polyp Segmentation. Diagnostics. 2023; 13(5):896. https://doi.org/10.3390/diagnostics13050896

Chicago/Turabian Style

Ma, Haichao, Chao Xu, Chao Nie, Jubao Han, Yingjie Li, and Chuanxu Liu. 2023. "DBE-Net: Dual Boundary-Guided Attention Exploration Network for Polyp Segmentation" Diagnostics 13, no. 5: 896. https://doi.org/10.3390/diagnostics13050896

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop