Corsar, David
XEQ Scale for Evaluating XAI Experience Quality Grounded in Psychometric Theory
Wijekoon, Anjana, Wiratunga, Nirmalie, Corsar, David, Martin, Kyle, Nkisi-Orji, Ikechukwu, Díaz-Agudo, Belen, Bridge, Derek
Explainable Artificial Intelligence (XAI) aims to improve the transparency of autonomous decision-making through explanations. Recent literature has emphasised users' need for holistic "multi-shot" explanations and the ability to personalise their engagement with XAI systems. We refer to this user-centred interaction as an XAI Experience. Despite advances in creating XAI experiences, evaluating them in a user-centred manner has remained challenging. To address this, we introduce the XAI Experience Quality (XEQ) Scale (pronounced "Seek" Scale), for evaluating the user-centred quality of XAI experiences. Furthermore, XEQ quantifies the quality of experiences across four evaluation dimensions: learning, utility, fulfilment and engagement. These contributions extend the state-of-the-art of XAI evaluation, moving beyond the one-dimensional metrics frequently developed to assess single-shot explanations. In this paper, we present the XEQ scale development and validation process, including content validation with XAI experts as well as discriminant and construct validation through a large-scale pilot study. Out pilot study results offer strong evidence that establishes the XEQ Scale as a comprehensive framework for evaluating user-centred XAI experiences.
Tell me more: Intent Fulfilment Framework for Enhancing User Experiences in Conversational XAI
Wijekoon, Anjana, Corsar, David, Wiratunga, Nirmalie, Martin, Kyle, Salimi, Pedram
The evolution of Explainable Artificial Intelligence (XAI) has emphasised the significance of meeting diverse user needs. The approaches to identifying and addressing these needs must also advance, recognising that explanation experiences are subjective, user-centred processes that interact with users towards a better understanding of AI decision-making. This paper delves into the interrelations in multi-faceted XAI and examines how different types of explanations collaboratively meet users' XAI needs. We introduce the Intent Fulfilment Framework (IFF) for creating explanation experiences. The novelty of this paper lies in recognising the importance of "follow-up" on explanations for obtaining clarity, verification and/or substitution. Moreover, the Explanation Experience Dialogue Model integrates the IFF and "Explanation Followups" to provide users with a conversational interface for exploring their explanation needs, thereby creating explanation experiences. Quantitative and qualitative findings from our comparative user study demonstrate the impact of the IFF in improving user engagement, the utility of the AI system and the overall user experience. Overall, we reinforce the principle that "one explanation does not fit all" to create explanation experiences that guide the complex interaction through conversation.
Towards Feasible Counterfactual Explanations: A Taxonomy Guided Template-based NLG Method
Salimi, Pedram, Wiratunga, Nirmalie, Corsar, David, Wijekoon, Anjana
Counterfactual Explanations (cf-XAI) describe the smallest changes in feature values necessary to change an outcome from one class to another. However, many cf-XAI methods neglect the feasibility of those changes. In this paper, we introduce a novel approach for presenting cf-XAI in natural language (Natural-XAI), giving careful consideration to actionable and comprehensible aspects while remaining cognizant of immutability and ethical concerns. We present three contributions to this endeavor. Firstly, through a user study, we identify two types of themes present in cf-XAI composed by humans: content-related, focusing on how features and their values are included from both the counterfactual and the query perspectives; and structure-related, focusing on the structure and terminology used for describing necessary value changes. Secondly, we introduce a feature actionability taxonomy with four clearly defined categories, to streamline the explanation presentation process. Using insights from the user study and our taxonomy, we created a generalisable template-based natural language generation (NLG) method compatible with existing explainers like DICE, NICE, and DisCERN, to produce counterfactuals that address the aforementioned limitations of existing approaches. Finally, we conducted a second user study to assess the performance of our taxonomy-guided NLG templates on three domains. Our findings show that the taxonomy-guided Natural-XAI approach (n-XAI^T) received higher user ratings across all dimensions, with significantly improved results in the majority of the domains assessed for articulation, acceptability, feasibility, and sensitivity dimensions.
Behaviour Trees for Creating Conversational Explanation Experiences
Wijekoon, Anjana, Corsar, David, Wiratunga, Nirmalie
This paper presented an XAI system specification and an interactive dialogue model to facilitate the creation of Explanation Experiences (EE). Such specifications combine the knowledge of XAI, domain and system experts of a use case to formalise target user groups and their explanation needs and to implement explanation strategies to address those needs. Formalising the XAI system promotes the reuse of existing explainers and known explanation needs that can be refined and evolved over time using user evaluation feedback. The abstract EE dialogue model formalised the interactions between a user and an XAI system. The resulting EE conversational chatbot is personalised to an XAI system at run-time using the knowledge captured in its XAI system specification. This seamless integration is enabled by using Behaviour Trees (BT) to conceptualise both the EE dialogue model and the explanation strategies. In the evaluation, we discussed several desirable properties of using BTs over traditionally used STMs or FSMs. BTs promote the reusability of dialogue components through the hierarchical nature of the design. Sub-trees are modular, i.e. a sub-tree is responsible for a specific behaviour, which can be designed in different levels of granularity to improve human interpretability. The EE dialogue model consists of abstract behaviours needed to capture EE, accordingly, it can be implemented as a conversational, graphical or text-based interface which caters to different domains and users. There is a significant computational cost when using BTs for modelling dialogue, which we mitigate by using memory. Overall, we find that the ability to create robust conversational pathways dynamically makes BTs a good candidate for designing and implementing conversation for creating explanation experiences.
DisCERN:Discovering Counterfactual Explanations using Relevance Features from Neighbourhoods
Wiratunga, Nirmalie, Wijekoon, Anjana, Nkisi-Orji, Ikechukwu, Martin, Kyle, Palihawadana, Chamath, Corsar, David
Counterfactual explanations focus on "actionable knowledge" to help end-users understand how a machine learning outcome could be changed to a more desirable outcome. For this purpose a counterfactual explainer needs to discover input dependencies that relate to outcome changes. Identifying the minimum subset of feature changes needed to action an output change in the decision is an interesting challenge for counterfactual explainers. The DisCERN algorithm introduced in this paper is a case-based counter-factual explainer. Here counterfactuals are formed by replacing feature values from a nearest unlike neighbour (NUN) until an actionable change is observed. We show how widely adopted feature relevance-based explainers (i.e. LIME, SHAP), can inform DisCERN to identify the minimum subset of "actionable features". We demonstrate our DisCERN algorithm on five datasets in a comparative study with the widely used optimisation-based counterfactual approach DiCE. Our results demonstrate that DisCERN is an effective strategy to minimise actionable changes necessary to create good counterfactual explanations.
The Crowd and the Web of Linked Data: A Provenance Perspective
Markovic, Milan (University of Aberdeen) | Edwards, Peter (University of Aberdeen) | Corsar, David (University of Aberdeen) | Pan, Jeff Z. (University of Aberdeen)
The usefulness of intelligent applications/services reasoning with linked data is dependent on the availability and correctness of this data. The crowd potentially has an important role to play in performing the non-trivial tasks of creating, validating, and maintaining the online linked data sets used by applications and services. Additional information captured within a provenance record can be used in these tasks and others, such as evaluating the performance of the crowd and its members. In this paper we describe two roles for the crowd in the web of linked data (creation and maintenance), and argue that incorporating provenance into these tasks is beneficial especially in scenarios when the population of available workers is small. We also identify several challenges for the use of provenance in this context and define a set of requirements for a provenance model to address these challenges.
Report on the Fourth International Conference on Knowledge Capture (K-CAP 2007)
Sleeman, Derek (University of Aberdeen) | Barker, Ken (University of Texas) | Corsar, David (University of Aberdeen)
The Fourth International Conference on Knowledge Capture was held October 28-31, 2007 in Whistler, British Columbia. K-CAP 2007 included two invited talks, technical papers, posters, and demonstrations. Topics included knowledge engineering and modeling methodologies, knowledge engineering and the semantic web, mixed-initiative planning and decision-support tools, acquisition of problem-solving knowledge, knowledge-based markup techniques, knowledge extraction systems, knowledge acquisition tools, and advice taking systems.
Report on the Fourth International Conference on Knowledge Capture (K-CAP 2007)
Sleeman, Derek (University of Aberdeen) | Barker, Ken (University of Texas) | Corsar, David (University of Aberdeen)
The Fourth International Conference on Knowledge Capture was held October 28-31, 2007, in Whistler, British Columbia. The topics covered in the invited talks, technical papers, posters, and demonstrations included knowledge engineering and modeling methodologies, knowledge engineering and the semantic web, mixedinitiative planning and decision-support tools, acquisition of problem-solving knowledge, knowledge-based markup techniques, knowledge extraction systems, knowledge acquisition tools, and advice-taking systems. These events, which were from web-based game-playing systems. The title of his talk was "Human Ken Barker and John Gennari Derek Sleeman noted in his introductory Etzioni's invited talk and had primary responsibilities for comments, knowledge capture is gave some technical details of the systems the conference and workshop programs. In the The best technical paper Since the K-CAP series was initiated, last decade or so, knowledge capture award was presented to Kai Eckert, the K-CAP and European Knowledge has again expanded its horizons significantly Heiner Stuckenschmidt, and Magnus Acquisition Workshop (EKAW) meetings to embrace information-extraction Pfeffer for their paper "Interactive have been held in alternate years, techniques, and more recently Thesaurus Assessment for Automatic with the K-CAP meetings taking place the web and enhanced connectivity Document Annotation."