dc.contributor.author
Krois, Joachim
dc.contributor.author
Garcia Cantu, Anselmo
dc.contributor.author
Chaurasia, Akhilanand
dc.contributor.author
Patil, Ranjitkumar
dc.contributor.author
Chaudhari, Prabhat Kumar
dc.contributor.author
Gaudin, Robert
dc.contributor.author
Gehrung, Sascha
dc.contributor.author
Schwendicke, Falk
dc.date.accessioned
2023-02-28T13:13:40Z
dc.date.available
2023-02-28T13:13:40Z
dc.identifier.uri
https://refubium.fu-berlin.de/handle/fub188/38141
dc.identifier.uri
http://dx.doi.org/10.17169/refubium-37854
dc.description.abstract
We assessed the generalizability of deep learning models and how to improve it. Our exemplary use-case was the detection of apical lesions on panoramic radiographs. We employed two datasets of panoramic radiographs from two centers, one in Germany (Charite, Berlin, n=650) and one in India (KGMU, Lucknow, n=650): First, U-Net type models were trained on images from Charite (n=500) and assessed on test sets from Charite and KGMU (each n=150). Second, the relevance of image characteristics was explored using pixel-value transformations, aligning the image characteristics in the datasets. Third, cross-center training effects on generalizability were evaluated by stepwise replacing Charite with KGMU images. Last, we assessed the impact of the dental status (presence of root-canal fillings or restorations). Models trained only on Charite images showed a (mean +/- SD) F1-score of 54.1 +/- 0.8% on Charite and 32.7 +/- 0.8% on KGMU data (p<0.001/t-test). Alignment of image data characteristics between the centers did not improve generalizability. However, by gradually increasing the fraction of KGMU images in the training set (from 0 to 100%) the F1-score on KGMU images improved (46.1 +/- 0.9%) at a moderate decrease on Charite images (50.9 +/- 0.9%, p<0.01). Model performance was good on KGMU images showing root-canal fillings and/or restorations, but much lower on KGMU images without root-canal fillings and/or restorations. Our deep learning models were not generalizable across centers. Cross-center training improved generalizability. Noteworthy, the dental status, but not image characteristics were relevant. Understanding the reasons behind limits in generalizability helps to mitigate generalizability problems.
en
dc.rights.uri
https://creativecommons.org/licenses/by/4.0/
dc.subject
deep learning models
en
dc.subject
generalizability
en
dc.subject
dental image analysis
en
dc.subject.ddc
600 Technik, Medizin, angewandte Wissenschaften::610 Medizin und Gesundheit::610 Medizin und Gesundheit
dc.title
Generalizability of deep learning models for dental image analysis
dc.type
Wissenschaftlicher Artikel
dcterms.bibliographicCitation.articlenumber
6102
dcterms.bibliographicCitation.doi
10.1038/s41598-021-85454-5
dcterms.bibliographicCitation.journaltitle
Scientific Reports
dcterms.bibliographicCitation.originalpublishername
Springer Nature
dcterms.bibliographicCitation.volume
11
refubium.affiliation
Charité - Universitätsmedizin Berlin
refubium.funding
Springer Nature DEAL
refubium.resourceType.isindependentpub
no
dcterms.accessRights.openaire
open access
dcterms.bibliographicCitation.pmid
33731732
dcterms.isPartOf.eissn
2045-2322