AID-CXR: AI Assisted Detection of Chest X-Rays
Study Details
Study Description
Brief Summary
This study has been added as a sub study to the Simulation Training for Emergency Department Imaging 2 study (ClinicalTrials.gov ID NCT05427838).
The Lunit INSIGHT CXR is a validation study that aims to assess the utility of an Artificial Intelligence-based (AI) chest X-ray (CXR) interpretation tool in assisting the diagnostic accuracy, speed, and confidence of a varied group of healthcare professionals. The study will be conducted using 500 retrospectively collected inpatient and emergency department CXRs from two United Kingdom (UK) hospital trusts. Two fellowship trained thoracic radiologists will independently review all studies to establish the ground truth reference standard. The Lunit INSIGHT CXR tool will be used to analyze each CXR, and its performance will be measured against the expert readers. The study will evaluate the utility of the algorithm in improving reader accuracy and confidence as measured by sensitivity, specificity, positive predictive value, and negative predictive value. The study will measure the performance of the algorithm against ten abnormal findings, including pulmonary nodules/mass, consolidation, pneumothorax, atelectasis, calcification, cardiomegaly, fibrosis, mediastinal widening, pleural effusion, and pneumoperitoneum. The study will involve readers from various clinical professional groups with and without the assistance of Lunit INSIGHT CXR. The study will provide evidence on the impact of AI algorithms in assisting healthcare professionals such as emergency medicine and general medicine physicians who regularly review images in their daily practice.
Condition or Disease | Intervention/Treatment | Phase |
---|---|---|
|
Study Design
Arms and Interventions
Arm | Intervention/Treatment |
---|---|
Readers/Participants Reader Selection: 30 readers will be selected from the following five clinical specialty groups: emergency medicine (ED) adult intensive care (ICU) adult general medicine (AGM) radiographers (Rad) general radiologists Each specialty group consists of 6 members of ranked seniority. For the physicians this consists of: Two 'Juniors' (Foundation Year 1 - Specialty Training 2 years) Two 'Middle Grades' (Registrar from Specialty Training 3 to 6 years) Two Consultants For the radiographers, this consists of: Two 'Junior/Newly qualified radiographers' (up to 18 months experience post qualification) Two 'Mid-experience radiographers' (approx. 3 years' experience) Two 'Reporting radiographers' (5+ years' experience) |
Other: Cases reading
The reading will be done remotely via the Report and Image Quality Control site (www.RAIQC.com), an online platform allowing medical imaging viewing and reporting. Participants can work from any location, but the work must be done from a computer with internet access. For avoidance of doubt, the work cannot be performed from a phone or tablet.
The project is divided into two phases and participants are required to complete both phases. The estimated total involvement in the project is up to 20-24 hours.
Phase 1: Time allowed: 2 weeks
- Review 500 chest X-rays and express a clinical opinion through a structured reporting template (multiple choice, no open text required).
Rest/washout period of 2 weeks.
Phase 2 - Time allowed: 2 weeks
- Review 500 chest X-rays together with an AI report for each case and express your clinical opinion through the same structured reporting template used in Phase A.
|
Ground truthers Two consultant thoracic radiologists. A third senior thoracic radiologist's opinion (>20 years experience) will undertake arbitration. |
Other: Ground truthing
Two consultant thoracic radiologists will independently review the images to establish the 'ground truth' findings on the CXRs, where a consensus is reached this will then be used as the reference standard. In the case of disagreement, a third senior thoracic radiologist's opinion (>20 years experience) will undertake arbitration. A difficulty score will be assigned to each abnormality by the ground truthers using a 4-point Likert scale (1 being easy/obvious to 4 being hard/poorly visualised).
|
Outcome Measures
Primary Outcome Measures
- Performance of AI algorithm: sensitivity [During 4 weeks of reading time]
Evaluation of the Lunit INSIGHT CXR algorithm will be performed comparing it to the reference standard in order to determine sensitivity.
- Performance of AI algorithm: specificity [During 4 weeks of reading time]
Evaluation of the Lunit INSIGHT CXR algorithm will be performed comparing it to the reference standard in order to determine specificity.
- Performance of AI algorithm: Area under the ROC Curve (AU ROC) [During 4 weeks of reading time]
Evaluation of the Lunit INSIGHT CXR algorithm will be performed comparing it to the reference standard. Continuous probability score from the algorithm will be utilized for the ROC analyses, while binary classification results with a predefined operating cut-off will be used for evaluation of sensitivity, specificity, positive predictive value, and negative predictive value.
- Performance of readers with and without AI assistance: Sensitivity [During 4 weeks of reading time]
The study will include two sessions (with and without AI overlay), with all 30 readers reviewing all 500 CXR cases each time separated by a washout period to mitigate recall bias. The cases will be randomised between the two reads and for every reader.
- Performance of readers with and without AI assistance: Specificity [During 4 weeks of reading time]
The study will include two sessions (with and without AI overlay), with all 30 readers reviewing all 500 CXR cases each time separated by a washout period to mitigate recall bias. The cases will be randomised between the two reads and for every reader.
- Performance of readers with and without AI assistance: Area under the ROC Curve (AU ROC) [During 4 weeks of reading time]
The study will include two sessions (with and without AI overlay), with all 30 readers reviewing all 500 CXR cases each time separated by a washout period to mitigate recall bias. The cases will be randomised between the two reads and for every reader.
- Reader speed with vs without AI assistance. [During 4 weeks of reading time]
Mean time taken to review a scan, with vs without AI assistance.
Eligibility Criteria
Criteria
Inclusion Criteria:
- General radiologists/radiographers/physicians who review CXRs as part of their routine clinical practice
Exclusion Criteria:
-
Thoracic radiologists
-
Non-radiology physicians with previous formal postgraduate CXR reporting training.
-
Non-radiology physicians with previous career in radiology, respiratory medicine or thoracic surgery to registrar or consultant level
Contacts and Locations
Locations
Site | City | State | Country | Postal Code | |
---|---|---|---|---|---|
1 | Oxford University Hospitals NHS Foundation Trust | Oxford | Oxfordshire | United Kingdom | OX3 9DU |
Sponsors and Collaborators
- Oxford University Hospitals NHS Trust
Investigators
None specified.Study Documents (Full-Text)
None provided.More Information
Additional Information:
- Richards M. Diagnostics: Recovery and Renewal - Report of the Independent Review of Diagnostic Services for NHS England. NHS England 2022.
- Royal College of Radiologists. Clinical radiology UK workforce census 2019 report. Royal College of Radiologists 2020.
- www.nice.org.uk. (2022). Summary | Artificial intelligence for analysing chest X-ray images | Advice | NICE. [online] Available at:
Publications
- Ahmad HK, Milne MR, Buchlak QD, Ektas N, Sanderson G, Chamtie H, Karunasena S, Chiang J, Holt X, Tang CHM, Seah JCY, Bottrell G, Esmaili N, Brotchie P, Jones C. Machine Learning Augmented Interpretation of Chest X-rays: A Systematic Review. Diagnostics (Basel). 2023 Feb 15;13(4):743. doi: 10.3390/diagnostics13040743.
- Greenhalgh R, Howlett DC, Drinkwater KJ. Royal College of Radiologists national audit evaluating the provision of imaging in the severely injured patient and compliance with national guidelines. Clin Radiol. 2020 Mar;75(3):224-231. doi: 10.1016/j.crad.2019.10.025. Epub 2019 Dec 19.
- Hillis JM, Bizzo BC, Mercaldo S, Chin JK, Newbury-Chaet I, Digumarthy SR, Gilman MD, Muse VV, Bottrell G, Seah JCY, Jones CM, Kalra MK, Dreyer KJ. Evaluation of an Artificial Intelligence Model for Detection of Pneumothorax and Tension Pneumothorax in Chest Radiographs. JAMA Netw Open. 2022 Dec 1;5(12):e2247172. doi: 10.1001/jamanetworkopen.2022.47172.
- Homayounieh F, Digumarthy S, Ebrahimian S, Rueckel J, Hoppe BF, Sabel BO, Conjeti S, Ridder K, Sistermanns M, Wang L, Preuhs A, Ghesu F, Mansoor A, Moghbel M, Botwin A, Singh R, Cartmell S, Patti J, Huemmer C, Fieselmann A, Joerger C, Mirshahzadeh N, Muse V, Kalra M. An Artificial Intelligence-Based Chest X-ray Model on Human Nodule Detection Accuracy From a Multicenter Study. JAMA Netw Open. 2021 Dec 1;4(12):e2141096. doi: 10.1001/jamanetworkopen.2021.41096.
- Jones CM, Buchlak QD, Oakden-Rayner L, Milne M, Seah J, Esmaili N, Hachey B. Chest radiographs and machine learning - Past, present and future. J Med Imaging Radiat Oncol. 2021 Aug;65(5):538-544. doi: 10.1111/1754-9485.13274. Epub 2021 Jun 25.
- Kundu R, Das R, Geem ZW, Han GT, Sarkar R. Pneumonia detection in chest X-ray images using an ensemble of deep learning models. PLoS One. 2021 Sep 7;16(9):e0256630. doi: 10.1371/journal.pone.0256630. eCollection 2021.
- Matsumoto T, Kodera S, Shinohara H, Ieki H, Yamaguchi T, Higashikuni Y, Kiyosue A, Ito K, Ando J, Takimoto E, Akazawa H, Morita H, Komuro I. Diagnosing Heart Failure from Chest X-Ray Images Using Deep Learning. Int Heart J. 2020 Jul 30;61(4):781-786. doi: 10.1536/ihj.19-714. Epub 2020 Jul 18. Erratum In: Int Heart J. 2020;61(5):1088.
- Spiritoso R, Padley S, Singh S. Chest X-ray interpretation in UK intensive care units: A survey 2014. J Intensive Care Soc. 2015 Nov;16(4):339-344. doi: 10.1177/1751143715580141. Epub 2015 May 18.
- van Beek EJR, Ahn JS, Kim MJ, Murchison JT. Validation study of machine-learning chest radiograph software in primary and emergency medicine. Clin Radiol. 2023 Jan;78(1):1-7. doi: 10.1016/j.crad.2022.08.129. Epub 2022 Sep 25.
- Wilson C. X-ray misinterpretation in urgent care: where does it occur, why does it occur, and does it matter? N Z Med J. 2022 Apr 1;135:49-65.
- Wu JT, Wong KCL, Gur Y, Ansari N, Karargyris A, Sharma A, Morris M, Saboury B, Ahmad H, Boyko O, Syed A, Jadhav A, Wang H, Pillai A, Kashyap S, Moradi M, Syeda-Mahmood T. Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents. JAMA Netw Open. 2020 Oct 1;3(10):e2022779. doi: 10.1001/jamanetworkopen.2020.22779.
- Zech JR, Badgeley MA, Liu M, Costa AB, Titano JJ, Oermann EK. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLoS Med. 2018 Nov 6;15(11):e1002683. doi: 10.1371/journal.pmed.1002683. eCollection 2018 Nov.
- 310995 - B