Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Explainable AI in Healthcare: Models, Applications, and Challenges

View through CrossRef
The healthcare industry is changing due to the application of Artificial Intelligence (AI) because it opens the possibility of making predictions in analytics, clinical decision support, medical imaging diagnostics, and custom treatment. Nevertheless, most of the recent AI models, particularly the deep learning models, are classified as black boxes since they are not interpretable. This veil puts the ethics, law and practice of health care in doubt where judgment directly affects patient safety and trust. The concept of explainable artificial intelligence (XAI) has emerged as a significant research field to make AI systems easier to understand and more open, interpretable, and credible by providing explanations of their behaviour at model level. This article is a review of explainable AI in healthcare that covers the types of models, their uses, and challenges. It further describes common techniques, including the post-hoc interpretability techniques (e.g. SHAP, LIME), the inherently interpretable models (e.g. decision trees, rule-based systems), and hybrids. It has also been discussed within the framework of diagnostic imaging, electronic health records (EHRs), drug discovery, and precision medicine. Also, there are still concerns regarding how to deal with trade-offs between accuracy and interpretability, how to measure evaluation metrics in a standardized manner, how to evaluate fairly, and how to translate XAI into clinical practice. The findings of the existing literature indicate that XAI will increase compliance with clinicians and regulators and patient empowerment. But under varying clinical conditions, scalability, the expense and variability of interpretability, continue to be key bottlenecks. The new directions are building the context-specific XAI models, the federated learning support, the alignment of the models with the ethical and legal principles, and GDPR. In this paper, the explainable AI is identified as a key to the responsible use of AI in healthcare. XAI also ensures that AI is used to make healthcare delivery safer, more ethical and more effective by closing the gap between complex models and the way human beings make decisions.
Title: Explainable AI in Healthcare: Models, Applications, and Challenges
Description:
The healthcare industry is changing due to the application of Artificial Intelligence (AI) because it opens the possibility of making predictions in analytics, clinical decision support, medical imaging diagnostics, and custom treatment.
Nevertheless, most of the recent AI models, particularly the deep learning models, are classified as black boxes since they are not interpretable.
This veil puts the ethics, law and practice of health care in doubt where judgment directly affects patient safety and trust.
The concept of explainable artificial intelligence (XAI) has emerged as a significant research field to make AI systems easier to understand and more open, interpretable, and credible by providing explanations of their behaviour at model level.
This article is a review of explainable AI in healthcare that covers the types of models, their uses, and challenges.
It further describes common techniques, including the post-hoc interpretability techniques (e.
g.
SHAP, LIME), the inherently interpretable models (e.
g.
decision trees, rule-based systems), and hybrids.
It has also been discussed within the framework of diagnostic imaging, electronic health records (EHRs), drug discovery, and precision medicine.
Also, there are still concerns regarding how to deal with trade-offs between accuracy and interpretability, how to measure evaluation metrics in a standardized manner, how to evaluate fairly, and how to translate XAI into clinical practice.
The findings of the existing literature indicate that XAI will increase compliance with clinicians and regulators and patient empowerment.
But under varying clinical conditions, scalability, the expense and variability of interpretability, continue to be key bottlenecks.
The new directions are building the context-specific XAI models, the federated learning support, the alignment of the models with the ethical and legal principles, and GDPR.
In this paper, the explainable AI is identified as a key to the responsible use of AI in healthcare.
XAI also ensures that AI is used to make healthcare delivery safer, more ethical and more effective by closing the gap between complex models and the way human beings make decisions.

Related Results

Perceptions of Telemedicine and Rural Healthcare Access in a Developing Country: A Case Study of Bayelsa State, Nigeria
Perceptions of Telemedicine and Rural Healthcare Access in a Developing Country: A Case Study of Bayelsa State, Nigeria
Abstract Introduction Telemedicine is the remote delivery of healthcare services using information and communication technologies and has gained global recognition as a solution to...
Integrating quantum neural networks with machine learning algorithms for optimizing healthcare diagnostics and treatment outcomes
Integrating quantum neural networks with machine learning algorithms for optimizing healthcare diagnostics and treatment outcomes
The rapid advancements in artificial intelligence (AI) and quantum computing have catalyzed an unprecedented shift in the methodologies utilized for healthcare diagnostics and trea...
Assessing Explainable in Artificial Intelligence: A TOPSIS Approach to Decision-Making
Assessing Explainable in Artificial Intelligence: A TOPSIS Approach to Decision-Making
Explainable in Artificial Intelligence (AI) is the ability to comprehend and explain how AI models generate judgments or predictions. The complexity of AI systems, especially machi...
A systematic review on the healthcare system in Jordan: Strengths, weaknesses, and opportunities for improvement
A systematic review on the healthcare system in Jordan: Strengths, weaknesses, and opportunities for improvement
Introduction: This systematic review examines the strengths and weaknesses of Jordan's healthcare system, providing valuable insights for healthcare providers, policymakers, and re...
PERSPECTIVES FOR COMPETITION IN THE HEALTHCARE INDUSTRY
PERSPECTIVES FOR COMPETITION IN THE HEALTHCARE INDUSTRY
A paradox has been established in the modern healthcare industry - consumers can choose between many alternatives but with high uncertainty, while healthcare establishments have nu...
The Hazards of Data Mining in Healthcare
The Hazards of Data Mining in Healthcare
From the mid-1990s, data mining methods have been used to explore and find patterns and relationships in healthcare data. During the 1990s and early 2000's, data mining was a topic...
Revolutionizing multimodal healthcare diagnosis, treatment pathways, and prognostic analytics through quantum neural networks
Revolutionizing multimodal healthcare diagnosis, treatment pathways, and prognostic analytics through quantum neural networks
The advent of quantum computing has introduced significant potential to revolutionize healthcare through quantum neural networks (QNNs), offering unprecedented capabilities in proc...

Back to Top