Benchmarking of Deep Architectures for Segmentation of Medical Images

IEEE Trans Med Imaging. 2022 Nov;41(11):3231-3241. doi: 10.1109/TMI.2022.3180435. Epub 2022 Oct 27.

Abstract

In recent years, there were many suggestions regarding modifications of the well-known U-Net architecture in order to improve its performance. The central motivation of this work is to provide a fair comparison of U-Net and its five extensions using identical conditions to disentangle the influence of model architecture, model training, and parameter settings on the performance of a trained model. For this purpose each of these six segmentation architectures is trained on the same nine data sets. The data sets are selected to cover various imaging modalities (X-rays, computed tomography, magnetic resonance imaging), single- and multi-class segmentation problems, and single- and multi-modal inputs. During the training, it is ensured that the data preprocessing, data set split into training, validation, and testing subsets, optimizer, learning rate change strategy, architecture depth, loss function, supervision and inference are exactly the same for all the architectures compared. Performance is evaluated in terms of Dice coefficient, surface Dice coefficient, average surface distance, Hausdorff distance, training, and prediction time. The main contribution of this experimental study is demonstrating that the architecture variants do not improve the quality of inference related to the basic U-Net architecture while resource demand rises.

Publication types

  • Research Support, U.S. Gov't, Non-P.H.S.

MeSH terms

  • Benchmarking
  • Deep Learning*
  • Image Processing, Computer-Assisted / methods
  • Magnetic Resonance Imaging / methods
  • Tomography, X-Ray Computed / methods