Please use this identifier to cite or link to this item:
|Title:||Explainable artificial intelligence (XAI) for healthcare decision-making||Authors:||Zeng, Zhiwei||Keywords:||Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence||Issue Date:||2022||Publisher:||Nanyang Technological University||Source:||Zeng, Z. (2022). Explainable artificial intelligence (XAI) for healthcare decision-making. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/155849||Abstract:||With the increasing ubiquity of Artificial Intelligence (AI) in daily life and into more critical applications, explainability has been brought under the spotlight due to ethical, social, and legal concerns. Explainability is particularly important in application domains involving high-stakes decision-making, such as healthcare, medicine, law, and defence, as simply having high performance may not satisfy all the desiderata of stakeholders. Stakeholders in these domains often require greater insights into the decision-making process, better support for tracing decision trails and identifying errors, as well as greater trustworthiness from an AI system. An eXplainable AI (XAI) is one that has the ability to explain the outputs or functioning of a model in a clear and understandable manner to a human. In the explanation process of XAI, explainability approaches provide explanatory information to human stakeholders in order to facilitate their understanding of the outputs and functioning of an AI model. The renewed understanding then affects the extent to which the desiderata are satisfied. Over the past decade, a plethora of explainability approaches have been proposed, ranging from ante-hoc methods that seek to build explainability features into a model itself to post-hoc methods that use techniques to improve the explainability of an already developed model. Despite the rapid growth of the XAI field, three key open problems emerge from a review of the existing XAI methods. Firstly, post-hoc methods have been increasingly used for high-stakes decisions, but most of them provide explanatory information that is not faithful or not insightful to what the original model computes. Secondly, most existing methods can only provide associational explanations and lack the structure and reasoning ability to provide contrastive and counterfactual explanations. Thirdly, most existing methods do not provide explanations to suit different levels of understanding. Research problem 1 - Post-hoc explainability methods have quickly gained popularity as they are mostly model-agnostic and highly extensible. However, without elucidating or understanding the model’s inner workings, explanations generated by post-hoc methods may not faithfully represent the underlying relationships of the model and often do not provide enough information to facilitate the understanding of how the output is obtained. In this thesis, I propose an ante-hoc explainability methodology to support the making and explaining of high-stakes decisions in healthcare. It is an argumentative methodology that can infuse explainability into the model and generate explanations that are faithful and insightful to the inner workings and internal representations of the model. It follows a “debate-like” process that shares many similarities with the way humans deliberate and make decisions, thereby giving it unique advantages in generating human-understandable explanations. I instantiate the argumentative methodology in three explainable decision-making approaches for the modelling and reasoning of contexts, qualitative preferences as well as both attack and support relationships. Research problem 2 - Although contrastive and counterfactual explanations have attracted more research interest recently, existing XAI methods mostly generate associational explana- tions. Most existing methods are good at answering “Why X? ” but poor at answering questions that require higher-level inference, such as “Why not-X? ”. This can be attributed to the fact that most of them lack the necessary structures and mechanisms to perform high-level logical reasoning. Building on the structures and reasoning ability of the three proposed explain- able decision-making approaches, I construct different forms of contrastive and counterfactual explanations that can answer questions such as “Why X instead of Y? ” and “Why not-X? ”. Research problem 3 - Different levels of understanding are often required in most practical scenarios. However, most existing XAI methods do not provide explanations that can cater to the different levels of understanding. In this thesis, I propose and study formal properties of explanatory information in order to generate such explanations. I propose notions and computational constructs for generating explanations that can provide different amounts of justification or have different dialectical strengths, as well as focused explanations that only include selective information in the decision-making process. As a case study, I implement a proposed explainable decision-making approach for the diagnostics and prognostics of Alzheimer’s Disease (AD). Empirical evaluation with real-world AD datasets and human evaluations were conducted to assess the performance and explainability of the approach. In terms of decision performance, my approach achieved the highest accuracies and F1-scores for both the diagnosis and prognosis task when compared with six machine learning models. In terms of explainability, the evaluation results with 107 human subjects indicate that the explanations generated by my approach achieved better accessibility and verifiability than existing XAI models. With the results on accuracy and explainability, it may be concluded that the proposed argumentative explainability methodology can better satisfy the desiderata for XAI in the healthcare domain.||URI:||https://hdl.handle.net/10356/155849||DOI:||10.32657/10356/155849||Rights:||This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).||Fulltext Permission:||open||Fulltext Availability:||With Fulltext|
|Appears in Collections:||IGS Theses|
Files in This Item:
|Thesis_final_Zhiwei Zeng_signed with watermark.pdf||2.44 MB||Adobe PDF||View/Open|
Updated on May 25, 2022
Updated on May 25, 2022
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.