TY - JOUR
T1 - RadImageNet
T2 - An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
AU - Mei, Xueyan
AU - Liu, Zelong
AU - Robson, Philip M.
AU - Marinelli, Brett
AU - Huang, Mingqian
AU - Doshi, Amish
AU - Jacobi, Adam
AU - Cao, Chendi
AU - Link, Katherine E.
AU - Yang, Thomas
AU - Wang, Ying
AU - Greenspan, Hayit
AU - Deyer, Timothy
AU - Fayad, Zahi A.
AU - Yang, Yang
N1 - Publisher Copyright:
© 2022, Radiological Society of North America Inc.. All rights reserved.
PY - 2022/9
Y1 - 2022/9
N2 - Purpose: To demonstrate the value of pretraining with millions of radiologic images compared with ImageNet photographic images on downstream medical applications when using transfer learning. Materials and Methods: This retrospective study included patients who underwent a radiologic study between 2005 and 2020 at an out-patient imaging facility. Key images and associated labels from the studies were retrospectively extracted from the original study inter-pretation. These images were used for RadImageNet model training with random weight initiation. The RadImageNet models were compared with ImageNet models using the area under the receiver operating characteristic curve (AUC) for eight classification tasks and using Dice scores for two segmentation problems. Results: The RadImageNet database consists of 1.35 million annotated medical images in 131 872 patients who underwent CT, MRI, and US for musculoskeletal, neurologic, oncologic, gastrointestinal, endocrine, abdominal, and pulmonary pathologic conditions. For transfer learning tasks on small datasets—thyroid nodules (US), breast masses (US), anterior cruciate ligament injuries (MRI), and meniscal tears (MRI)—the RadImageNet models demonstrated a significant advantage (P < .001) to ImageNet models (9.4%, 4.0%, 4.8%, and 4.5% AUC improvements, respectively). For larger datasets—pneumonia (chest radiography), COVID-19 (CT), SARS-CoV-2 (CT), and intracranial hemorrhage (CT)—the RadImageNet models also illustrated improved AUC (P < .001) by 1.9%, 6.1%, 1.7%, and 0.9%, respectively. Additionally, lesion localizations of the RadImageNet models were improved by 64.6% and 16.4% on thyroid and breast US datasets, respectively. Conclusion: RadImageNet pretrained models demonstrated better interpretability compared with ImageNet models, especially for smaller radiologic datasets.
AB - Purpose: To demonstrate the value of pretraining with millions of radiologic images compared with ImageNet photographic images on downstream medical applications when using transfer learning. Materials and Methods: This retrospective study included patients who underwent a radiologic study between 2005 and 2020 at an out-patient imaging facility. Key images and associated labels from the studies were retrospectively extracted from the original study inter-pretation. These images were used for RadImageNet model training with random weight initiation. The RadImageNet models were compared with ImageNet models using the area under the receiver operating characteristic curve (AUC) for eight classification tasks and using Dice scores for two segmentation problems. Results: The RadImageNet database consists of 1.35 million annotated medical images in 131 872 patients who underwent CT, MRI, and US for musculoskeletal, neurologic, oncologic, gastrointestinal, endocrine, abdominal, and pulmonary pathologic conditions. For transfer learning tasks on small datasets—thyroid nodules (US), breast masses (US), anterior cruciate ligament injuries (MRI), and meniscal tears (MRI)—the RadImageNet models demonstrated a significant advantage (P < .001) to ImageNet models (9.4%, 4.0%, 4.8%, and 4.5% AUC improvements, respectively). For larger datasets—pneumonia (chest radiography), COVID-19 (CT), SARS-CoV-2 (CT), and intracranial hemorrhage (CT)—the RadImageNet models also illustrated improved AUC (P < .001) by 1.9%, 6.1%, 1.7%, and 0.9%, respectively. Additionally, lesion localizations of the RadImageNet models were improved by 64.6% and 16.4% on thyroid and breast US datasets, respectively. Conclusion: RadImageNet pretrained models demonstrated better interpretability compared with ImageNet models, especially for smaller radiologic datasets.
KW - Brain/Brain Stem
KW - CT
KW - Computer Applications–General (Infor-matics)
KW - Evidence-based Medicine
KW - Head/Neck
KW - MR Imaging
KW - Thorax
KW - US
UR - http://www.scopus.com/inward/record.url?scp=85137196972&partnerID=8YFLogxK
U2 - 10.1148/ryai.210315
DO - 10.1148/ryai.210315
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 36204533
AN - SCOPUS:85137196972
SN - 2638-6100
VL - 4
JO - Radiology: Artificial Intelligence
JF - Radiology: Artificial Intelligence
IS - 5
M1 - e210315
ER -