Are deep learning classification results obtained on CT scans fair and interpretable?

dc.contributor.authorAshames M.M.A.
dc.contributor.authorDemir A.
dc.contributor.authorGerek O.N.
dc.contributor.authorFidan M.
dc.contributor.authorGulmezoglu M.B.
dc.contributor.authorErgin S.
dc.contributor.authorEdizkan R.
dc.contributor.authorKoc M.
dc.contributor.authorBarkana A.
dc.contributor.authorCalisir C.
dc.date.accessioned2024-07-22T08:01:49Z
dc.date.available2024-07-22T08:01:49Z
dc.date.issued2024
dc.description.abstractFollowing the great success of various deep learning methods in image and object classification, the biomedical image processing society is also overwhelmed with their applications to various automatic diagnosis cases. Unfortunately, most of the deep learning-based classification attempts in the literature solely focus on the aim of extreme accuracy scores, without considering interpretability, or patient-wise separation of training and test data. For example, most lung nodule classification papers using deep learning randomly shuffle data and split it into training, validation, and test sets, causing certain images from the Computed Tomography (CT) scan of a person to be in the training set, while other images of the same person to be in the validation or testing image sets. This can result in reporting misleading accuracy rates and the learning of irrelevant features, ultimately reducing the real-life usability of these models. When the deep neural networks trained on the traditional, unfair data shuffling method are challenged with new patient images, it is observed that the trained models perform poorly. In contrast, deep neural networks trained with strict patient-level separation maintain their accuracy rates even when new patient images are tested. Heat map visualizations of the activations of the deep neural networks trained with strict patient-level separation indicate a higher degree of focus on the relevant nodules. We argue that the research question posed in the title has a positive answer only if the deep neural networks are trained with images of patients that are strictly isolated from the validation and testing patient sets. © The Author(s) 2024.
dc.identifier.DOI-ID10.1007/s13246-024-01419-8
dc.identifier.issn26624729
dc.identifier.urihttp://akademikarsiv.cbu.edu.tr:4000/handle/123456789/11575
dc.language.isoEnglish
dc.publisherSpringer Science and Business Media Deutschland GmbH
dc.rightsAll Open Access; Hybrid Gold Open Access
dc.subjectComputer aided diagnosis
dc.subjectComputerized tomography
dc.subjectDeep neural networks
dc.subjectLearning systems
dc.subjectAccuracy rate
dc.subjectChest computed tomography
dc.subjectClassification results
dc.subjectComputed tomography scan
dc.subjectDNN
dc.subjectInterpretability
dc.subjectInterpretability and reliability
dc.subjectMalignancy classification
dc.subjectPatient images
dc.subjectTraining sets
dc.subjectClassification (of information)
dc.titleAre deep learning classification results obtained on CT scans fair and interpretable?
dc.typeArticle

Files