Skip to main content

Main menu

  • Home
  • Our journals
    • Clinical Medicine
    • Future Healthcare Journal
  • Subject collections
  • About the RCP
  • Contact us

Future Healthcare Journal

  • FHJ Home
  • Content
    • Current
    • Ahead of print
    • Archive
  • Author guidance
    • Instructions for authors
    • Submit online
  • About FHJ
    • Scope
    • Editorial board
    • Policies
    • Information for reviewers
    • Advertising

User menu

  • Log in

Search

  • Advanced search
RCP Journals
Home
  • Log in
  • Home
  • Our journals
    • Clinical Medicine
    • Future Healthcare Journal
  • Subject collections
  • About the RCP
  • Contact us
Advanced

Future Healthcare Journal

futurehosp Logo
  • FHJ Home
  • Content
    • Current
    • Ahead of print
    • Archive
  • Author guidance
    • Instructions for authors
    • Submit online
  • About FHJ
    • Scope
    • Editorial board
    • Policies
    • Information for reviewers
    • Advertising

The implementation of artificial intelligence in radiology: a narrative review of patient perspectives

Scott Hemphill, Katherine Jackson, Stephen Bradley and Bobby Bhartia
Download PDF
DOI: https://doi.org/10.7861/fhj.2022-0097
Future Healthc J March 2023
Scott Hemphill
ASir Charles Gairdner Hospital, Perth, Australia
Roles: resident medical officer
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • For correspondence: scotthemphill94@gmail.com
Katherine Jackson
BLeeds Institute of Health Sciences, University of Leeds, Leeds, UK
Roles: general Practitioner and NIHR academic clinical lecturer
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Stephen Bradley
CLeeds Teaching Hospitals NHS Trust, Leeds, UK
Roles: consultant in thoracic and oncological radiology
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Bobby Bhartia
CLeeds Teaching Hospitals NHS Trust, Leeds, UK
Roles: consultant in thoracic and oncological radiology
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info & Metrics
Loading

ABSTRACT

Aim To synthesise research on the view of the public and patients of the use of artificial intelligence (AI) in radiology investigations.

Methods A literature review of narrative synthesis of qualitative and quantitative studies that reported views of the public and patients toward the use of AI in radiology.

Results Only seven studies related to patient and public views were retrieved, suggesting that this is an underexplored area of research. Two broad themes, of confidence in the capabilities of AI, and the accountability and transparency of AI, were identified.

Conclusions Both optimism and concerns were expressed by participants. Transparency in the implementation of AI, scientific validation, clear regulation and accountability were expected. Combined human and AI interpretation of imaging was strongly favoured over AI acting autonomously. The review highlights the limited engagement of the public in the adoption of AI in a radiology setting. Successful implementation of AI in this field will require demonstrating not only adequate accuracy of the technology, but also its acceptance by patients.

KEYWORDS:
  • artificial intelligence
  • radiology
  • patient perspective
  • machine learning

Background

In recent years, there has been increasing interest in the application of artificial intelligence (AI) within radiology. Advances in such technology come at a time where the volume of radiological investigations is increasing, and AI could offer the potential to improve both the speed and accuracy of radiological reporting.1 A search for the terms ‘artificial intelligence’ and ‘radiology’ within PubMed yielded 26 citations from 2016, increasing to 1,539 citations in 2021, an almost 6,000% increase. Well over 100 CE-marked products are available within Europe, although with significant variation in the levels of peer-reviewed evidence to demonstrate their efficacy.2

A key consideration in the implementation of AI is the trust and acceptance of the technology among stakeholders. Clinical professionals and trainees remain uncertain about the effect of the introduction of AI into the field of radiology.1,3–5 The stakeholders who bear the greatest risks from new technologies are patients. Achieving adequate understanding of the attitudes and concerns of patients is crucial to ensure their interests are represented in determining how the technology is used to deliver clinical care. Accordingly, we undertook a narrative review to gather and explore current evidence regarding patient perspectives on utilising AI within medical imaging.

Methods

Search strategy

On 9 June 2022, we conducted an electronic database search for published and unpublished studies of patient perspectives on AI in radiology. The search strategy is outlined in Table 1 and documented in full in supplementary material S1 along with the study protocol in supplementary material S2.

View this table:
  • View inline
  • View popup
Table 1.

Search terms formulated for qualitative studies detailing patient views on artificial intelligence in radiology

Data collection, narrative synthesis and quality assessment

All titles and abstracts were reviewed by one researcher (SH) with a random selection of 20% independently screened by a second researcher (SB). The full text of any potentially eligible article was then obtained and assessed for eligibility. An appraisal of the literature for themes using an inductive process as described by Arai et al6 and the highlighting of data around these themes were conducted separately by SH and BB and then summarised in consensus with SB. A quality assessment of the studies was undertaken by SB and BB using a mixed-methods appraisal tool.7 Narrative synthesis was guided by methods described by Arai et al and findings reported in accordance with Enhancing Transparency in Reporting the Synthesis of Qualitative Research (ENTREQ) criteria.6,8

Results

The search yielded a total of 2,832 studies, of which 116 were selected for full-text review, resulting in seven studies in the final review (Fig 1). These were five cross-sectional survey studies and two semi-structured interview studies. The studies were conducted in the Netherlands (n=3), Germany (n=2), Sweden (n=1) and the UK (n=1) and were all published between 2019 and 2021 (Table 2). Two studies focused on breast cancer screening and included only female participants. None of the studies reported restrictions by ethnicity. Critical appraisal of the studies was performed using the Mixed Methods Appraisal Tool7 (supplementary material S3).

Fig 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Fig 1.

Study selection.

View this table:
  • View inline
  • View popup
Table 2.

Description of studies that satisfied the study eligibility criteria

Strengths and limitations

This is the first review of published studies on patient perspectives on the introduction of AI in radiology. We identified only a few studies that met eligibility criteria. Despite the studies being conducted in different countries and in heterogeneous health systems, the themes explored were reasonably consistent.

All of the studies were conducted in Western Europe, and all but one was performed within a medical setting, which could limit the generalisability of the opinions expressed. Patients' prior knowledge of AI would plausibly affect their perceptions and preparedness to accept the technology in image interpretation. Only two studies11,14 recorded participants' prior understanding of AI. Although the influence of prior knowledge of AI on the attitudes of participants was modest, it is a limitation because the other included studies did not assess it. Analysis of non-responders was not reported in any of the studies, which is a potential sources of bias in the reported results.

Discussion of the key themes

Our results demonstrate the lack of evidence regarding patient views compared with other groups, with only seven eligible studies identified, compared with 43 studies that explored the views of clinical professionals (34 studies) and medical students (nine studies) (supplementary material S4).

Two broad recurring themes were identified; confidence in the capabilities of AI for diagnostic imaging, and views regarding the accountability and transparency of a diagnostic process that might include AI. The demographic features that were associated with views on the adoption of these technologies are presented in Table 3.

View this table:
  • View inline
  • View popup
Table 3.

Summary of patient data addressing the themes of confidence in the capabilities of AI, its accountability and transparency, and demographic features associated with these views

Confidence in the capabilities of AI

Patients recognised the future implementation of AI in the diagnostic process as inevitable.9,10 The potential to add safety, apply state-of-the-art analysis and facilitate faster results (and, thus, reduce anxiety) were identified by patients as desirable benefits.9–11 The highest levels of confidence were consistently expressed for systems utilising a combination of AI and clinical interpretation, sometimes coined ‘AI-supported decision making’.9,10,12 In only one study, there was there greater confidence in a clinician's interpretation without the use of AI.13 In general, a greater level of trust was expressed in the ability of the clinician to correctly interpret imaging compared with AI, with a consistent negative attitude toward AI being a sole replacement for human diagnostic decision making. However, AI was seen as potentially slightly superior to clinicians in incorporating state-of-the-art information to make decisions based on the most current evidence.10,14

Digital literacy, prior experience of AI and educational attainment were factors associated with greater acceptance of the use of AI.9,11–15 Age was evaluated in two studies and found to be not, or only weakly, associated with reduced acceptability of AI.9,13 Acceptability was reduced if there was a preference for human interaction, a history of chronic disease or an inability to work. Patient acceptance of AI was reduced as the severity of the disease being assessed increased.14 When discussing test results, most patients felt that human interaction and empathy were vital parts of this process, particularly in understanding the reliability of results and their clinical impact.10

Accountability and transparency

Clinician opinion is strongly favoured over AI in circumstances in which clinician and AI did not agree,13,14 with patients viewing the role of AI as a decision aid for clinicians.

Accountability for errors if AI were operating autonomously was not explored in depth. There was uncertainty expressed regarding whether the responsibility would belong to the radiologist, the wider clinical team or the AI provider.10,12 This reflects similar uncertainty expressed within the radiology profession and highlights the need for an ethical and legal framework within which AI systems are expected to operate.4,15 Patients expressed a wish for transparency regarding the role of AI in a decision-making process and an expectation of the reasoning behind AI-based decisions.10,11,14

There was uncertainty regarding the responsibility for regulating the efficacy of AI systems.10,12 Generally, patients expected that AI systems would be subject to independent scientific validation and operate within a regulatory or legal framework,10,11,15 with the duty for developing this validation considered to belong to the ‘scientific community’.10 These expectations are at odds with the current state of implementation of AI. A recent review of over 100 radiology CE-marked AI products available for clinical use found that 64 lacked peer-review evidence and approximately half of the evidence produced was co-funded or co-authored by the vendor.2

Although patients placed great faith in human interpretation, the inherent risk of errors in current practice was recognised, as were the potential benefits of AI to assist reaching a correct diagnosis. This is reflected in the preference for a model of human interpretation with AI as a second opinion, with the human as the final arbitrator of the outcome.

Conclusions

There is limited published literature regarding the views of patients on the use of AI. This review indicates that patients see the potential value of this technology, but expect transparency around its role, with high expectations of validation and regulation. Many of the uncertainties around issues of responsibility for decision making and governance are shared by clinicians.

Key uncertainties surround attitudes to the topic of diagnostic error. The self-learning algorithms utilised by AI will mean that the mechanisms for an error might remain unclear even to the manufacturer. Although the limitations of human perception and decision making resulting in error are difficult to accept, they are at least comprehensible. Arguments that the likelihood of an error might have been greater without AI are nuanced and prone to the bias of hindsight. Public opinion regarding the relative responsibilities of the radiologist, manufacturer and wider regulatory framework remain unclear.

There appears to have been only limited exploration of the public's attitude toward important governance issues, such as expectations around transparency and consent, ensuring personal privacy and use of personal data for the development of commercial products.

With increasing demands and limited resources, the justifications to develop AI as a replacement for human interpretation will increase. Yet, a clear majority of patients identified in this review currently favour the implementation of AI as an aid rather than a replacement for decision making, especially when the potential stakes are high. This highlights a separation between the possible roles of AI and what is currently considered acceptable. Developing an understanding of the concerns that underlie these views is a key aspect of addressing them and realising the many potential roles for AI.

Summary points

  • Public and patient views are substantially under-represented in current literature regarding opinions on implementing AI into radiology.

  • A combination of human and AI image interpretation is strongly favoured. AI working independent of human oversight is not widely supported.

  • Patients expect transparency in the implementation of AI, together with independent scientific validation, clear regulation and accountability.

Supplementary material

Additional supplementary material may be found in the online version of this article at www.rcpjournals.org/content/futurehosp

S1 – Search Strategy

S2 – Study Protocol

S3 – Quality assessment

S4 – References for studies containing the views of other stakeholders

  • © Royal College of Physicians 2023. All rights reserved.

References

  1. ↵
    1. Neri E
    , de Souza N, Brady A, et al. What the radiologist should know about artificial intelligence – an ESR white paper. Insights Imaging 2019;10:44.
    OpenUrl
  2. ↵
    1. van Leeuwen KG
    , Schalekamp S, Rutten MJCM, et al. Artificial intelligence in radiology: 100 commercially available products and their scientific evidence. Euro Radiol 2021;31:3797–804.
    OpenUrlPubMed
  3. ↵
    1. Gallix B
    , Chong J. Artificial intelligence in radiology: who's afraid of the big bad wolf? Eur Radiol 2019;29:1637–39.
    OpenUrl
  4. ↵
    1. European Society of Radiology (ESR)
    . Impact of artificial intelligence on radiology: a EuroAIM survey among members of the European Society of Radiology. Insights Imaging 2019;10:105.
    OpenUrlCrossRefPubMed
  5. ↵
    1. Pinto Dos Santos D
    , Giese D, Brodehl S, et al. Medical students' attitude towards artificial intelligence: a multicentre survey. Eur Radiol 2019;29:1640–46.
    OpenUrlCrossRefPubMed
  6. ↵
    1. Arai L
    , Britten N, Popay J, et al. Testing methodological developments in the conduct of narrative synthesis: a demonstration review of research on the implementation of smoke alarm interventions. Evidence Policy 2007;3:361–83.
    OpenUrlCrossRef
  7. ↵
    1. Hong QN
    , Gonzalez-Reyes A, Pluye P. Improving the usefulness of a tool for appraising the quality of qualitative, quantitative and mixed methods studies, the Mixed Methods Appraisal Tool (MMAT). J Eval Clin Pract 2018;24:459–67.
    OpenUrlPubMed
  8. ↵
    1. Tong A
    , Flemming K, McInnes E, et al. Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ. BMC Med Res Methodol 2012;12:181.
    OpenUrlCrossRefPubMed
  9. ↵
    1. Jonmarker O
    , Strand F, Brandberg Y, Lindholm P. The future of breast cancer screening: what do participants in a breast cancer screening program think about automation using artificial intelligence? Acta Radiol Open 2019;8:2058460119880315.
    OpenUrl
  10. ↵
    1. Haan M
    , Ongena YP, Hommes S, et al. A qualitative study to understand patient perspective on the use of artificial intelligence in radiology. J Am Coll Radiol 2019;16:1416–19.
    OpenUrlCrossRefPubMed
  11. ↵
    1. Müller A
    , Mertens SM, Göstemeyer G, et al. Barriers and Enablers for Artificial Intelligence in Dental Diagnostics: A Qualitative Study. J Clin Med 2021;10:1612.
    OpenUrl
  12. ↵
    1. Ongena YP
    , Yakar D, Haan M, et al. Artificial Intelligence in Screening Mammography: A Population Survey of Women's Preferences. J Am Coll Radiol 2021;18:79–86.
    OpenUrlPubMed
  13. ↵
    1. York T
    , Jenney H, Jones G. Clinician and computer: a study on patient perceptions of artificial intelligence in skeletal radiography. BMJ Health Care Inform 2020;27:e100233.
    OpenUrl
  14. ↵
    1. Lennartz S
    , Dratsch T, Zopfs D, et al. Use and control of artificial intelligence in patients across the medical workflow: single-center questionnaire study of patient perspectives. J Med Internet Res 2021;23:e24221.
    OpenUrlPubMed
  15. ↵
    1. Ongena YP
    , Haan M, Yakar D, et al. Patients' views on the implementation of artificial intelligence in radiology: development and validation of a standardized questionnaire. Eur Radiol 2020;30:1033–40.
    OpenUrlCrossRefPubMed
Back to top
Previous articleNext article

Article Tools

Download PDF
Article Alerts
Sign In to Email Alerts with your Email Address
Citation Tools
The implementation of artificial intelligence in radiology: a narrative review of patient perspectives
Scott Hemphill, Katherine Jackson, Stephen Bradley, Bobby Bhartia
Future Healthc J Mar 2023, 10 (1) 63-68; DOI: 10.7861/fhj.2022-0097

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
The implementation of artificial intelligence in radiology: a narrative review of patient perspectives
Scott Hemphill, Katherine Jackson, Stephen Bradley, Bobby Bhartia
Future Healthc J Mar 2023, 10 (1) 63-68; DOI: 10.7861/fhj.2022-0097
Reddit logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • ABSTRACT
    • Background
    • Methods
    • Results
    • Strengths and limitations
    • Discussion of the key themes
    • Confidence in the capabilities of AI
    • Accountability and transparency
    • Conclusions
    • Summary points
    • Supplementary material
    • References
  • Figures & Data
  • Info & Metrics

Related Articles

  • No related articles found.
  • Google Scholar

Cited By...

  • No citing articles found.
  • Google Scholar

More in this TOC Section

  • The impact of interprofessional incivility on medical performance, service and patient care: a systematic review
  • Personalised future prescribing using pharmacogenomics: A resumé of a joint Royal College of Physicians / British Pharmacological Society working party report
Show more Review

Similar Articles

FAQs

  • Difficulty logging in.

There is currently no login required to access the journals. Please go to the home page and simply click on the edition that you wish to read. If you are still unable to access the content you require, please let us know through the 'Contact us' page.

  • Can't find the CME questionnaire.

The read-only self-assessment questionnaire (SAQ) can be found after the CME section in each edition of Clinical Medicine. RCP members and fellows (using their login details for the main RCP website) are able to access the full SAQ with answers and are awarded 2 CPD points upon successful (8/10) completion from:  https://cme.rcplondon.ac.uk

Navigate this Journal

  • Journal Home
  • Current Issue
  • Ahead of Print
  • Archive

Related Links

  • ClinMed - Home
  • FHJ - Home

Other Services

  • Advertising
futurehosp Footer Logo
  • Home
  • Journals
  • Contact us
  • Advertise
HighWire Press, Inc.

Follow Us:

  • Follow HighWire Origins on Twitter
  • Visit HighWire Origins on Facebook

Copyright © 2021 by the Royal College of Physicians