U-NET BASED POLYP SEGMENTATION ON KVASIR-SEG DATASET: PERFORMANCE EVALUATION AND COMPARISON WITH STATE-OF-THE-ART METHODS

U-NET BASED POLYP SEGMENTATION ON KVASIR-SEG DATASET: PERFORMANCE EVALUATION AND COMPARISON WITH STATE-OF-THE-ART METHODS

Авторы

  • Mukhriddin Arabboev
  • Shohruh Begmatov
  • Sukhrob Bobojanov

DOI:

https://doi.org/10.5281/zenodo.17874308

Ключевые слова:

U-Net, Polyp Segmentation, Kvasir-SEG, Deep Learning, Medical Image Segmentation, Convolutional Neural Networks.

Аннотация

Medical image segmentation is a fundamental task in computer-assisted diagnosis, enabling accurate
delineation of anatomical structures and pathological regions. In gastrointestinal endoscopy, precise segmentation of
polyps is crucial for early detection and treatment planning. This study presents a U-Net-based approach for polyp
segmentation using the publicly available Kvasir-SEG dataset. The proposed model was trained for 100 epochs with
custom metrics including Dice coefficient, Intersection over Union (IoU), precision, and recall. Experimental results
demonstrate that the model achieved an average Dice score of 0.9449, IoU of 0.9084, precision of 0.9404, and recall of
0.9584, outperforming several recent state-of-the-art methods. Comparative analysis with existing approaches confirms
the effectiveness of the vanilla U-Net architecture when combined with careful preprocessing and hyperparameter tuning.
The findings highlight U-Net's continued relevance in medical image segmentation tasks and suggest directions for future
work, including integration with attention mechanisms and transformer-based architectures.

Биографии авторов

Mukhriddin Arabboev

Tashkent University of Information
Technologies named after Muhammad al-Khwarizmi
Doctor of Philosophy (PhD) in Technical Sciences,
Doctoral (DSc) student

Shohruh Begmatov

Tashkent University of Information
Technologies named after Muhammad al-Khwarizmi
Doctor of Philosophy (PhD) in Technical Sciences,
Doctoral (DSc) student

Sukhrob Bobojanov

Urgench RANCH Technological University,
Associate Professor of the Department of Digital Technologies
Doctor of Philosophy (PhD) in Technical Sciences

Библиографические ссылки

S. Anantharajan, S. Gunasekaran, T. Subramanian, and V. R, “MRI brain tumor detection using deep learning

and machine learning approaches,” Meas. Sensors, vol. 31, no. January 2023, p. 101026, 2024, doi: 10.1016/j.

measen.2024.101026.

M. Arabboev, “BRAIN TUMOR CLASSIFICATION USING TRANSFER LEARNING WITH MOBILENETV2,” Top.

Issues Tech. Sci., vol. 3, no. 5, pp. 51–63, 2025, doi: https://doi.org/10.47390/ts-v3i5y2025N8.

P. Kaur and P. Mahajan, “Detection of brain tumors using a transfer learning-based optimized ResNet152 model in

MR images,” Comput. Biol. Med., vol. 188, no. February, p. 109790, 2025, doi: 10.1016/j.compbiomed.2025.109790.

I. Pacal, M. Alaftekin, and F. D. Zengul, “Enhancing Skin Cancer Diagnosis Using Swin Transformer with Hybrid Shifted

Window-Based Multi-head Self-attention and SwiGLU-Based MLP,” J. Imaging Informatics Med., no. 0123456789,

, doi: 10.1007/s10278-024-01140-8.

M. Arabboev, S. Begmatov, M. Rikhsivoev, S. Bobojanov, G. Tangriberganov, and J. Uraimov, “Deep learningbased

skin cancer classification: a comparative study of pre-trained CNN models,” in 30th International Conference

Information Society and University Studies - IVUS 2025, 2025.

M. Hajiarbabi, “Skin cancer detection using multi-scale deep learning and transfer learning,” J. Med. Artif. Intell., vol. 6,

no. 6, pp. 1–9, 2023, doi: 10.21037/jmai-23-67.

J. B. Thomas, K. V. Shihabudheen, S. M. Sulthan, and A. Al-Jumaily, “Deep Feature Meta-Learners Ensemble Models

for COVID-19 CT Scan Classification,” Electron., vol. 12, no. 3, 2023, doi: 10.3390/electronics12030684.

L. Sun et al., “GACEMV: An ensemble learning framework for constructing COVID-19 diagnosis and prognosis

models,” Biomed. Signal Process. Control, vol. 94, no. April, p. 106305, 2024, doi: 10.1016/j.bspc.2024.106305.

E. Erdem and T. Aydin, “Detection of Pneumonia with a Novel CNN-based Approach,” Sak. Univ. J. Comput. Inf. Sci.,

vol. 4, no. 1, pp. 26–34, 2021, doi: 10.35377/saucis.04.01.787030.

A. K. Mudiyanselage, “Deep Learning-Based Diagnosis of Pneumonia Using Convolutional Neural Networks,” J.

Comput. Mech. Manag., vol. 3, no. 3, pp. 14–21, 2024, doi: 10.57159/gadl.jcmm.3.3.240126.14.

M. Arabboev and S. Begmatov, “Deep learning-based pneumonia detection from chest X-ray images using a

convolutional neural network,” Mod. Innov. Syst. Technol., vol. 5, no. 3, pp. 1018–1026, 2025.

G. Ayana, H. Barki, and S. Choe, “Pathological Insights: Enhanced Vision Transformers for the Early Detection of

Colorectal Cancer,” Cancers (Basel)., vol. 16, no. 7, p. 1441, Apr. 2024, doi: 10.3390/cancers16071441.

A. Santone, M. Cesarelli, and F. Mercaldo, “A Method for Polyp Segmentation Through U-Net Network,” Bioengineering,

vol. 12, no. 3, pp. 1–19, 2025, doi: 10.3390/bioengineering12030236.

D. Jha et al., “Kvasir-SEG: A Segmented Polyp Dataset,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes

Artif. Intell. Lect. Notes Bioinformatics), vol. 11962 LNCS, no. April 2020, pp. 451–462, 2020, doi: 10.1007/978-3-030-

-2_37.

D. Jha et al., “A Comprehensive Study on Colorectal Polyp Segmentation with ResUNet++, Conditional Random Field

and Test-Time Augmentation,” IEEE J. Biomed. Heal. Informatics, vol. 25, no. 6, pp. 2029–2040, 2021, doi: 10.1109/

JBHI.2021.3049304.

G. Yue, W. Han, S. Li, T. Zhou, J. Lv, and T. Wang, “Automated polyp segmentation in colonoscopy images via deep

network with lesion-aware feature selection and refinement,” Biomed. Signal Process. Control, vol. 78, no. November

, p. 103846, 2022, doi: 10.1016/j.bspc.2022.103846.

M. Nouman Noor, M. Nazir, S. A. Khan, I. Ashraf, and O.-Y. Song, “Localization and Classification of Gastrointestinal

Tract Disorders Using Explainable AI from Endoscopic Images,” Appl. Sci., vol. 13, no. 15, p. 9031, Aug. 2023, doi:

3390/app13159031.

R. Nachmani, I. Nidal, D. Robinson, M. Yassin, and D. Abookasis, “Segmentation of polyps based on pyramid vision

transformers and residual block for real-time endoscopy imaging,” J. Pathol. Inform., vol. 14, no. November 2022, p.

, 2023, doi: 10.1016/j.jpi.2023.100197.

T. Shen and X. Li, “Automatic polyp image segmentation and cancer prediction based on deep learning,” Front. Oncol.,

vol. 12, no. January, pp. 1–12, 2023, doi: 10.3389/fonc.2022.1087438.

D. Bhattacharya, D. Eggert, C. Betz, and A. Schlaefer, “Squeeze and multi-context attention for polyp segmentation,”

Int. J. Imaging Syst. Technol., vol. 33, no. 1, pp. 123–142, Jan. 2023, doi: 10.1002/ima.22795.

S. Pan, X. Liu, N. Xie, and Y. Chong, “EG-TransUNet: a transformer-based U-Net with enhanced and guided models

for biomedical image segmentation,” BMC Bioinformatics, vol. 24, no. 1, pp. 1–22, 2023, doi: 10.1186/s12859-023-

-1.

D. He et al., “Dual-guided network for endoscopic image segmentation with region and boundary cues,” Biomed.

Signal Process. Control, vol. 91, no. January, p. 106059, 2024, doi: 10.1016/j.bspc.2024.106059.

M. F. Ahamed, M. R. Islam, M. Nahiduzzaman, M. E. H. Chowdhury, A. Alqahtani, and M. Murugappan, “Automated

Colorectal Polyps Detection from Endoscopic Images using MultiResUNet Framework with Attention Guided

Segmentation,” Human-Centric Intell. Syst., vol. 4, no. 2, pp. 299–315, 2024, doi: 10.1007/s44230-024-00067-1.

W. Dong, B. Du, and Y. Xu, “Shape-intensity-guided U-net for medical image segmentation,” Neurocomputing, vol.

, no. March, 2024, doi: 10.1016/j.neucom.2024.128534.

N. Tirpude, T. Diwan, and M. Dhabu, “Meta-Transformers: A Hybrid Approach for Medical Image Segmentation

with U-Net and Meta-Learning,” IEEE Access, vol. 13, no. February, pp. 64822–64831, 2025, doi: 10.1109/

ACCESS.2025.3559446.

Загрузки

Опубликован

2025-12-01
Loading...