Towards a lightweight 2D U-Net for accurate semantic segmentation of kidney tumors in abdominal CT images

Luca Drole, Isabella Poles, Eleonora D'Arnese, Marco D. Santambrogio

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Accurate segmentation of the kidney anatomy is crucial in the diagnosis and treatment of various kidney diseases. However, 3D U-Net-based Neural Networks entail significant computational requirements, complex architectures, and a fully annotated volumetric dataset. To address these challenges, our study designs and implement a custom image preprocessing workflow that suppresses fat and uninformative structures and compares the performances of 2D U-Net-based Neural Networks for semantic segmentation of kidneys and tumors from abdominal CT images. We found the ResU-Net model to achieve an accuracy of 89.17% for kidney segmentation, outperforming other models, while the Vanilla U-Net during the renal tumor segmentation task, with up to 11.7% higher DSC scores. Moreover, all the investigated methods do not require 3D CNNs, thus reducing computational costs. This comparison could be potentially useful to make a step forward in identifying the most accurate and lightweight technology to aid physicians in diagnosing kidney diseases while improving patient outcomes.
Original languageEnglish
Title of host publicationIEEE EUROCON 2023 - 20th International Conference on Smart Technologies
PublisherInstitute of Electrical and Electronics Engineers
Pages12-17
Number of pages6
ISBN (Electronic)9781665463973
DOIs
Publication statusPublished - 7 Aug 2023
Event20th International Conference on Smart Technologies - Torino, Italy
Duration: 6 Jul 20238 Jul 2023

Conference

Conference20th International Conference on Smart Technologies
Abbreviated titleEUROCON 2023
Country/TerritoryItaly
CityTorino
Period6/07/238/07/23

Fingerprint

Dive into the research topics of 'Towards a lightweight 2D U-Net for accurate semantic segmentation of kidney tumors in abdominal CT images'. Together they form a unique fingerprint.

Cite this