Skip to main content
Advertisement

Main menu

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • Low-Field MRI
    • Alzheimer Disease
    • ASNR Foundation Special Collection
    • Photon-Counting CT
    • View All
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home

User menu

  • Alerts
  • Log in

Search

  • Advanced search
American Journal of Neuroradiology
American Journal of Neuroradiology

American Journal of Neuroradiology

ASHNR American Society of Functional Neuroradiology ASHNR American Society of Pediatric Neuroradiology ASSR
  • Alerts
  • Log in

Advanced Search

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • Low-Field MRI
    • Alzheimer Disease
    • ASNR Foundation Special Collection
    • Photon-Counting CT
    • View All
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home
  • Follow AJNR on Twitter
  • Visit AJNR on Facebook
  • Follow AJNR on Instagram
  • Join AJNR on LinkedIn
  • RSS Feeds

AJNR Awards, New Junior Editors, and more. Read the latest AJNR updates

Article CommentaryWHITE PAPER

Critical Appraisal of Artificial Intelligence–Enabled Imaging Tools Using the Levels of Evidence System

N. Pham, V. Hill, A. Rauschecker, Y. Lui, S. Niogi, C.G. Fillipi, P. Chang, G. Zaharchuk and M. Wintermark
American Journal of Neuroradiology April 2023, DOI: https://doi.org/10.3174/ajnr.A7850
N. Pham
aFrom the Department of Radiology (N.P., G.Z.), Stanford School of Medicine, Palo Alto, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for N. Pham
V. Hill
bDepartment of Radiology (V.H.), Northwestern University Feinberg School of Medicine, Chicago, Illinois
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for V. Hill
A. Rauschecker
cDepartment of Radiology (A.R.), University of California, San Francisco, San Francisco, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for A. Rauschecker
Y. Lui
dDepartment of Radiology (Y.L.), NYU Grossman School of Medicine, New York, New York
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Y. Lui
S. Niogi
eDepartment of Radiology (S.N.), Weill Cornell Medicine, New York, New York
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for S. Niogi
C.G. Fillipi
fDepartment of Radiology (C.G.F.), Tufts University School of Medicine, Boston, Massachusetts
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for C.G. Fillipi
P. Chang
gDepartment of Radiology (P.C.), University of California, Irvine, Irvine, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for P. Chang
G. Zaharchuk
aFrom the Department of Radiology (N.P., G.Z.), Stanford School of Medicine, Palo Alto, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for G. Zaharchuk
M. Wintermark
hDepartment of Neuroradiology (M.W.), The University of Texas MD Anderson Cancer Center, Houston, Texas
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for M. Wintermark
  • Article
  • Figures & Data
  • Info & Metrics
  • Responses
  • References
  • PDF
Loading

Article Figures & Data

Figures

  • Tables
  • FIGURE.
    • Download figure
    • Open in new tab
    • Download powerpoint
    FIGURE.

    Levels of evidence. Proposed 7 levels of evidence for the systematic evaluation of an AI product’s quality and effectiveness in the clinical setting.

Tables

  • Figures
    • View popup
    Table 1:

    Summary of levels of evidence

    Levels of EvidenceElementTypes of Evidence
    Level 1Clinical efficacyOne prospective or randomized clinical trial or meta-analysis
    Level 2Bias and error mitigationAt least 2 independent retrospective studies separate from original institution
    Level 3Reproducibility and generalizabilityAt least 2 retrospective studies with at least 1 from an institution independent of the original institution
    Level 4Technical efficacyTwo retrospective studies from the same institution
    Level 5AData quality and AI model development with external testingOne retrospective study with internal and external data used for final performance reporting
    Level 5BData quality and AI model development with internal testingOne retrospective study with only internal data used for final performance reporting
    Level 6Interoperability and integration into the IT infrastructureAI company can provide a plan including interoperability standards for integration into the existing radiology and hospital digital information systems
    Level 7Legal and regulatory frameworksAI-enabled tool is compliant with current patient data protection, security, privacy, HIPAA, and government regulations
    • Note:—IT indicates information technology.

    • View popup
    Table 2:

    Detailed summary of levels of evidencea

    Levels of EvidenceElementDescriptionTypes of EvidenceSignificance
    Level 1Clinical efficacyClinical efficacy is the assessment of how the AI tool impacts patient care and health care outcomesAI tool has been used in at least 1 prospective study, randomized clinical trial, or meta-analysis demonstrating potential for improved patient care and health care outcomes, including improved mortality, quality of life, morbidity, or reduced health care costAlthough measuring clinical efficacy and added value for an early AI technology is challenging, it remains the single most important feature for clinical success and adoption
    Level 2Bias and error mitigationBiases of different types invariably exist in all data and can lead to AI modeling errors when applied to patients in different clinical settingsAI model can adapt to at least 2 separate institutions different from where it was initially developed (2 independent retrospective studies) AND Peer-reviewed results from those retrospective studies from above are available describing the various populations used to test the AI model, including age ranges, sex, and types of scanners AND AI company has a process to continuously incorporate feedback and improve their model (postdeployment monitoring)Al related errors in clinical practice may be harmful to patients; thus, the AI tool should be tested at multiple different sites, with differing patient demographics, disease prevalence, and imaging vendors to determine operational characteristics, generalizability, and potential pitfalls As clinical and patient care standards are constantly evolving, AI models may need routine surveillance and updates for performance and data drift
    Level 3Reproducibility and generalizabilityAI-enabled tool can be applied to different clinical settings, while demonstrating consistent high-quality resultsAt least 2 retrospective studies showing that the AI tool has performance characteristics similar to or alternative methods in the literature AND At least 1 independent study different from where the AI tool was developed to demonstrate that the AI tool can adapt to at least 1 different institutionA multi-institution approach supports reproducibility and generalizability of the AI tool
    Level 4Technical efficacyTechnical efficacy is the assessment that the AI model correctly performs the task that it was trained to doAI model performance has been shown in 2 retrospective studies to have potential clinical impact compared with similar or alternative methods in the literature AND These retrospective studies can be from the same institutionRecent investigations suggest that less than 40% of AI models have peer-reviewed evidence available on their efficacy6
    Level 5AData quality and AI model development with external testingAI models are prone to overfitting and an external test data set during development should be used to report final performance metricsLevel 5B evidence as described with an external test set used for performance validationInclusion of an external data set during the development phase supports the generalizability of the AI tool
    Level 5BData quality and AI model development with internal testingAI company should provide peer-reviewed information about the characteristics of the data used to develop the AI model AI company can explain how the AI model makes decisions that are relevant to patient careOne retrospective study showing the following: Peer-reviewed results detailing the inclusion/exclusion criteria, source, and type of data used to train, validate, and test the AI model AND Peer-reviewed results describing how the AI model was developed, including use of a standard of reference that is widely accepted for the intended clinical task AND No final external test set was used for final performance reportingData characteristics will influence the suitability and applicability of the AI model to the target patient population of interest Selection of a high-quality standard of reference is important for accurately comparing the peak performance of the AI model to that of current clinical practice
    Level 6Interoperability and integration into the IT infrastructureAI software should integrate seamlessly into the hospital information system, radiology information system, and PACS to be clinically usefulAI company can provide a plan including interoperability standards for integration into the existing radiology and hospital digital information systems AI company can provide on-site demonstration of clinical integration and potential impact on workflow before full deploymentSuccessful clinical implementation of an AI tool requires close collaboration between the AI company and site experts, including radiologists, referring physicians, data scientists, and information technologists Real time demonstration is an important mechanism for identifying potential site-specific problems
    Level 7Legal and regulatory frameworksPatient consent, privacy, and confidentiality laws will vary depending on state, local, and institutional regulationsAI-enabled tool is compliant with current patient data protection, security, privacy, HIPAA, and government regulationsAI companies, health care systems, and radiologists are key gatekeepers of patient autonomy, privacy, and safety
    • ↵a Appropriate reporting of AI model performance will depend on the task; however, examples of relevant statistical measures include ROC, sensitivity, specificity, and positive and negative predictive values, among others.

PreviousNext
Back to top
Advertisement
Print
Download PDF
Email Article

Thank you for your interest in spreading the word on American Journal of Neuroradiology.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Critical Appraisal of Artificial Intelligence–Enabled Imaging Tools Using the Levels of Evidence System
(Your Name) has sent you a message from American Journal of Neuroradiology
(Your Name) thought you would like to see the American Journal of Neuroradiology web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Cite this article
N. Pham, V. Hill, A. Rauschecker, Y. Lui, S. Niogi, C.G. Fillipi, P. Chang, G. Zaharchuk, M. Wintermark
Critical Appraisal of Artificial Intelligence–Enabled Imaging Tools Using the Levels of Evidence System
American Journal of Neuroradiology Apr 2023, DOI: 10.3174/ajnr.A7850

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
0 Responses
Respond to this article
Share
Bookmark this article
Critical Appraisal of Artificial Intelligence–Enabled Imaging Tools Using the Levels of Evidence System
N. Pham, V. Hill, A. Rauschecker, Y. Lui, S. Niogi, C.G. Fillipi, P. Chang, G. Zaharchuk, M. Wintermark
American Journal of Neuroradiology Apr 2023, DOI: 10.3174/ajnr.A7850
del.icio.us logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Purchase

Jump to section

  • Article
    • Abstract
    • ABBREVIATIONS:
    • CONCLUSIONS
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • Responses
  • References
  • PDF

Related Articles

  • No related articles found.
  • PubMed
  • Google Scholar

Cited By...

  • Outcomes of Radiologist Recommendations for Temporal Bone CT to Assess Superior Semicircular Canal Dehiscence on Temporal Bone MRI
  • Empowering Data Sharing in Neuroscience: A Deep Learning Deidentification Method for Pediatric Brain MRIs
  • Automated Detection of Steno-Occlusive Lesion on Time-of-Flight MR Angiography: An Observer Performance Study
  • Postoperative Karnofsky performance status prediction in patients with IDH wild-type glioblastoma: a multimodal approach integrating clinical and deep imaging features
  • A Siamese U-Transformer for change detection on MRI brain for multiple sclerosis, a model development and external validation study
  • A Radiomic "Warning Sign" of Progression on Brain MRI in Individuals with MS
  • Crossref (15)
  • Google Scholar

This article has been cited by the following articles in journals that are participating in Crossref Cited-by Linking.

  • Artificial Intelligence in Thyroidology: A Narrative Review of the Current Applications, Associated Challenges, and Future Directions
    David Toro-Tobon, Ricardo Loor-Torres, Mayra Duran, Jungwei W. Fan, Naykky Singh Ospina, Yonghui Wu, Juan P. Brito
    Thyroid® 2023 33 8
  • Artificial intelligence in medicine and research – the good, the bad, and the ugly
    Victor Grech, Sarah Cuschieri, Abdelazeem A. Eldawlatly
    Saudi Journal of Anaesthesia 2023 17 3
  • A Radiomic “Warning Sign” of Progression on Brain MRI in Individuals with MS
    Brendan S. Kelly, Prateek Mathur, Gerard McGuinness, Henry Dillon, Edward H. Lee, Kristen W. Yeom, Aonghus Lawlor, Ronan P. Killeen
    American Journal of Neuroradiology 2024 45 2
  • Hepatic encephalopathy post-TIPS: Current status and prospects in predictive assessment
    Xiaowei Xu, Yun Yang, Xinru Tan, Ziyang Zhang, Boxiang Wang, Xiaojie Yang, Chujun Weng, Rongwen Yu, Qi Zhao, Shichao Quan
    Computational and Structural Biotechnology Journal 2024 24
  • Automated Detection of Steno-Occlusive Lesion on Time-of-Flight MR Angiography: An Observer Performance Study
    Hunjong Lim, Dongjun Choi, Leonard Sunwoo, Jae Hyeop Jung, Sung Hyun Baik, Se Jin Cho, Jinhee Jang, Tackeun Kim, Kyong Joon Lee
    American Journal of Neuroradiology 2024 45 9
  • Cultivating diagnostic clarity: The importance of reporting artificial intelligence confidence levels in radiologic diagnoses
    Mobina Fathi, Kimia Vakili, Ramtin Hajibeygi, Ashkan Bahrami, Shima Behzad, Armin Tafazolimoghadam, Hadiseh Aghabozorgi, Reza Eshraghi, Vivek Bhatt, Ali Gholamrezanezhad
    Clinical Imaging 2025 117
  • La regulación legal de la inteligencia artificial en la Unión Europea: guía práctica para radiólogos
    Á. Morales Santos, S. Lojo Lendoiro, M. Rovira Cañellas, P. Valdés Solís
    Radiología 2024 66 5
  • Controversies in the Application of AI in Radiology—Is There Medico-Legal Support? Aspects from Romanian Practice
    Ana-Maria Ungureanu, Sergiu-Ciprian Matei, Daniel Malita
    Diagnostics 2025 15 2
  • The legal regulation of artificial intelligence in the European Union: A practical guide for radiologists
    Á. Morales Santos, S. Lojo Lendoiro, M. Rovira Cañellas, P. Valdés Solís
    Radiología (English Edition) 2024 66 5
  • Checklist Approach to Developing and Implementing AI in Clinical Settings: Instrument Development Study
    Ayomide Owoyemi, Joanne Osuchukwu, Megan E Salwei, Andrew Boyd
    JMIRx Med 2025 6

More in this TOC Section

  • Amyloid Imaging Abnormalities in Alzheimer’s
  • The Updated Neuroradiology Milestones: Synapsing from 1.0 to 2.0
Show more WHITE PAPER

Similar Articles

Advertisement

Indexed Content

  • Current Issue
  • Accepted Manuscripts
  • Article Preview
  • Past Issues
  • Editorials
  • Editor's Choice
  • Fellows' Journal Club
  • Letters to the Editor
  • Video Articles

Cases

  • Case Collection
  • Archive - Case of the Week
  • Archive - Case of the Month
  • Archive - Classic Case

More from AJNR

  • Trainee Corner
  • Imaging Protocols
  • MRI Safety Corner
  • Book Reviews

Multimedia

  • AJNR Podcasts
  • AJNR Scantastics

Resources

  • Turnaround Time
  • Submit a Manuscript
  • Submit a Video Article
  • Submit an eLetter to the Editor/Response
  • Manuscript Submission Guidelines
  • Statistical Tips
  • Fast Publishing of Accepted Manuscripts
  • Graphical Abstract Preparation
  • Imaging Protocol Submission
  • Evidence-Based Medicine Level Guide
  • Publishing Checklists
  • Author Policies
  • Become a Reviewer/Academy of Reviewers
  • News and Updates

About Us

  • About AJNR
  • Editorial Board
  • Editorial Board Alumni
  • Alerts
  • Permissions
  • Not an AJNR Subscriber? Join Now
  • Advertise with Us
  • Librarian Resources
  • Feedback
  • Terms and Conditions
  • AJNR Editorial Board Alumni

American Society of Neuroradiology

  • Not an ASNR Member? Join Now

© 2025 by the American Society of Neuroradiology All rights, including for text and data mining, AI training, and similar technologies, are reserved.
Print ISSN: 0195-6108 Online ISSN: 1936-959X

Powered by HighWire