A review of interpretable ml in healthcare: Taxonomy, applications, challenges, and future directions

Abdullah, T.A.A. and Zahid, M.S.M. and Ali, W. (2021) A review of interpretable ml in healthcare: Taxonomy, applications, challenges, and future directions. Symmetry, 13 (12).

Full text not available from this repository.
Official URL: https://www.scopus.com/inward/record.uri?eid=2-s2....

Abstract

We have witnessed the impact of ML in disease diagnosis, image recognition and classifi-cation, and many more related fields. Healthcare is a sensitive field related to people�s lives in which decisions need to be carefully taken based on solid evidence. However, most ML models are complex, i.e., black-box, meaning they do not provide insights into how the problems are solved or why such decisions are proposed. This lack of interpretability is the main reason why some ML models are not widely used yet in real environments such as healthcare. Therefore, it would be beneficial if ML models could provide explanations allowing physicians to make data-driven decisions that lead to higher quality service. Recently, several efforts have been made in proposing interpretable machine learning models to become more convenient and applicable in real environments. This paper aims to provide a comprehensive survey and symmetry phenomena of IML models and their applications in healthcare. The fundamental characteristics, theoretical underpinnings needed to develop IML, and taxonomy for IML are presented. Several examples of how they are applied in healthcare are investigated to encourage and facilitate the use of IML models in healthcare. Fur-thermore, current limitations, challenges, and future directions that might impact applying ML in healthcare are addressed. © 2021 by the authors. Licensee MDPI, Basel, Switzerland.

Item Type: Article
Impact Factor: cited By 0
Depositing User: Ms Sharifah Fahimah Saiyed Yeop
Date Deposited: 25 Mar 2022 06:41
Last Modified: 25 Mar 2022 06:41
URI: http://scholars.utp.edu.my/id/eprint/30316

Actions (login required)

View Item
View Item