Equitable deep learning for diabetic retinopathy detection using multi-dimensional retinal imaging with fair adaptive scaling: a retrospective study

Abstract

Background: As deep learning becomes increasingly accessible for automated detection of diabetic retinopathy (DR), questions persist regarding its performance equity among diverse identity groups. We aimed to explore the fairness of current deep learning models and further create a more equitable model designed to minimize disparities in performance across groups. Methods: This study used one proprietary and two publicly available datasets, containing two-dimensional (2D) wide-angle color fundus, scanning laser ophthalmoscopy (SLO) fundus, and three-dimensional (3D) Optical Coherence Tomography (OCT) B-Scans, to assess deep learning models for DR detection. We developed a fair adaptive scaling (FAS) module that dynamically adjusts the significance of samples during model training for DR detection, aiming to lessen performance disparities across varied identity groups. FAS was incorporated into both 2D and 3D deep learning models to facilitate the binary classification of DR and non-DR cases. The area under the receiver operating characteristic curve (AUC) was adopted to measure the model performance. Additionally, we devised an equity-scaled AUC metric that evaluates model fairness by balancing overall AUC against disparities among groups. Findings: Using in-house color fundus on the racial attribute, the overall AUC and ES-AUC of EfficientNet after integrating with FAS improved from 0.88 and 0.83 to 0.90 and 0.84 (p < 0.05), where the AUCs for Asians and Whites improved by 0.04 and 0.03, respectively (p < 0.01). On gender, the overall AUC and ES-AUC of EfficientNet after integrating with FAS both improved by 0.01 (p < 0.05). While using in-house SLO fundus on race, the overall AUC and ES-AUC of EfficientNet after integrating FAS improved from 0.80 to 0.83 (p < 0.01), where the AUCs for Asians, Blacks, and Whites improved by 0.02, 0.01 and 0.04, respectively (p < 0.05). On gender, FAS improved EfficientNet's overall AUC and ES-AUC both by 0.02, where the same improvement of 0.02 (p < 0.01) was gained for Females and Males. Using 3D deep learning model DenseNet121 on in-house OCT-B-Scans on race, FAS improved the overall AUC and ES-AUC from 0.875 and 0.81 to 0.884 and 0.82 respectively, where the AUCs for Asians and Blacks improved by 0.03 and 0.02 (p < 0.01). On gender, FAS improved the overall AUC and ES-AUC of DenseNet121 by 0.04 and 0.03, whereas the AUCs for Females and Males improved by 0.05 and 0.04 (p < 0.01), respectively. Interpretation: Existing deep learning models indeed exhibit variable performance across diverse identity groups in DR detection. The FAS proves beneficial in enhancing model equity and boosting DR detection accuracy, particularly for underrepresented groups.

Competing Interest Statement

The authors have declared no competing interest.

Funding Statement

This work was supported by NIH R00 EY028631, NIH R21 EY035298, Research To Prevent Blindness International Research Collaborators Award, and Alcon Young Investigator Grant.

Author Declarations

I confirm all relevant ethical guidelines have been followed, and any necessary IRB and/or ethics committee approvals have been obtained.

Yes

The details of the IRB/oversight body that provided approval or exemption for the research described are given below:

The fundus and OCT data used for developing the equitable deep learning model were from Massachusetts Eye and Ear (MEE) between 2021 and 2023. The institutional review boards (IRB) of MEE approved the creation of the database in this retrospective study. This study complied with the guidelines outlined in the Declaration of Helsinki. In light of the study's retrospective design, the requirement for informed consent was waived.

I confirm that all necessary patient/participant consent has been obtained and the appropriate institutional forms have been archived, and that any patient/participant/sample identifiers included were not known to anyone (e.g., hospital staff, patients or participants themselves) outside the research group so cannot be used to identify individuals.

Yes

I understand that all clinical trials and any other prospective interventional studies must be registered with an ICMJE-approved registry, such as ClinicalTrials.gov. I confirm that any such study reported in the manuscript has been registered and the trial registration ID is provided (note: if posting a prospective study registered retrospectively, please provide a statement in the trial ID field explaining why the study was not registered in advance).

Yes

I have followed all appropriate research reporting guidelines, such as any relevant EQUATOR Network research reporting checklist(s) and other pertinent material, if applicable.

Yes

Data Availability

The Harvard-FairVision30k dataset is available through the public link https://ophai.hms.harvard.edu/datasets/harvard-fairvision30k and was used with approvals. The ODIR-5K dataset is publicly available at https://www.kaggle.com/datasets/andrewmvd/ocular-disease-recognition-odir5k. The in-house data were provided by the Massachusetts Eye and Ear (MEE). The institutional review boards (IRB) of MEE approved the creation of the database in this retrospective study.

留言 (0)

沒有登入
gif