Hitzler, Pascal
Concept Induction using LLMs: a user experiment for assessment
Barua, Adrita, Widmer, Cara, Hitzler, Pascal
Explainable Artificial Intelligence (XAI) poses a significant challenge in providing transparent and understandable insights into complex AI models. Traditional post-hoc algorithms, while useful, often struggle to deliver interpretable explanations. Concept-based models offer a promising avenue by incorporating explicit representations of concepts to enhance interpretability. However, existing research on automatic concept discovery methods is often limited by lower-level concepts, costly human annotation requirements, and a restricted domain of background knowledge. In this study, we explore the potential of a Large Language Model (LLM), specifically GPT-4, by leveraging its domain knowledge and common-sense capability to generate high-level concepts that are meaningful as explanations for humans, for a specific setting of image classification. We use minimal textual object information available in the data via prompting to facilitate this process. To evaluate the output, we compare the concepts generated by the LLM with two other methods: concepts generated by humans and the ECII heuristic concept induction system. Since there is no established metric to determine the human understandability of concepts, we conducted a human study to assess the effectiveness of the LLM-generated concepts. Our findings indicate that while human-generated explanations remain superior, concepts derived from GPT-4 are more comprehensible to humans compared to those generated by ECII.
Towards Complex Ontology Alignment using Large Language Models
Amini, Reihaneh, Norouzi, Sanaz Saki, Hitzler, Pascal, Amini, Reza
Ontology alignment, a critical process in the Semantic Web for detecting relationships between different ontologies, has traditionally focused on identifying so-called "simple" 1-to-1 relationships through class labels and properties comparison. The more practically useful exploration of more complex alignments remains a hard problem to automate, and as such is largely underexplored, i.e. in application practice it is usually done manually by ontology and domain experts. Recently, the surge in Natural Language Processing (NLP) capabilities, driven by advancements in Large Language Models (LLMs), presents new opportunities for enhancing ontology engineering practices, including ontology alignment tasks. This paper investigates the application of LLM technologies to tackle the complex ontology alignment challenge. Leveraging a prompt-based approach and integrating rich ontology content - so-called modules - our work constitutes a significant advance towards automating the complex alignment task.
On the Psychology of GPT-4: Moderately anxious, slightly masculine, honest, and humble
Barua, Adrita, Brase, Gary, Dong, Ke, Hitzler, Pascal, Vasserman, Eugene
The capability of Large Language Models (LLMs) such as GPT-4 to engage in conversation with humans presents a significant leap in Artificial Intelligence (AI) development that is broadly considered to be disruptive for certain technological areas. A human interacting with an LLM may indeed perceive the LLM as an agent with a personality, to the extent that some have even called them sentient (De Cosmo, 2022). While we, of course, do not subscribe to the notion that LLMs are sentient - nor do we believe it is as yet clear what it means to even ask whether an LLM has a personality - there is still the appearance of agency and personality to the human user interacting with the system. Subjecting an LLM to psychometric tests is thus, in our view, less an assessment of some actual personality that the LLM may or may not have, but rather an assessment of the personality or personalities perceived by the human user. As such, our interest is not only in the actual personality profile(s) resulting from the tests, but also in the question whether the profiles are stable over re-tests and how they vary with different (relevant) parameter settings. At the same time, the results beg the question why the results of the tests are what they are.
Neural Fuzzy Extractors: A Secure Way to Use Artificial Neural Networks for Biometric User Authentication
Jana, Abhishek, Paudel, Bipin, Sarker, Md Kamruzzaman, Ebrahimi, Monireh, Hitzler, Pascal, Amariucai, George T
Powered by new advances in sensor development and artificial intelligence, the decreasing cost of computation, and the pervasiveness of handheld computation devices, biometric user authentication (and identification) is rapidly becoming ubiquitous. Modern approaches to biometric authentication, based on sophisticated machine learning techniques, cannot avoid storing either trained-classifier details or explicit user biometric data, thus exposing users' credentials to falsification. In this paper, we introduce a secure way to handle user-specific information involved with the use of vector-space classifiers or artificial neural networks for biometric authentication. Our proposed architecture, called a Neural Fuzzy Extractor (NFE), allows the coupling of pre-existing classifiers with fuzzy extractors, through a artificial-neural-network-based buffer called an expander, with minimal or no performance degradation. The NFE thus offers all the performance advantages of modern deep-learning-based classifiers, and all the security of standard fuzzy extractors. We demonstrate the NFE retrofit to a classic artificial neural network for a simple scenario of fingerprint-based user authentication.
Conversational Ontology Alignment with ChatGPT
Norouzi, Sanaz Saki, Mahdavinejad, Mohammad Saeid, Hitzler, Pascal
This study evaluates the applicability and efficiency of ChatGPT for ontology alignment using a naive approach. ChatGPT's output is compared to the results of the Ontology Alignment Evaluation Initiative 2022 campaign using conference track ontologies. This comparison is intended to provide insights into the capabilities of a conversational large language model when used in a naive way for ontology matching, and to investigate the potential advantages and disadvantages of this approach.
Understanding CNN Hidden Neuron Activations Using Structured Background Knowledge and Deductive Reasoning
Dalal, Abhilekha, Sarker, Md Kamruzzaman, Barua, Adrita, Vasserman, Eugene, Hitzler, Pascal
A major challenge in Explainable AI is in correctly interpreting activations of hidden neurons: accurate interpretations would provide insights into the question of what a deep learning system has internally detected as relevant on the input, demystifying the otherwise black-box character of deep learning systems. The state of the art indicates that hidden node activations can, in some cases, be interpretable in a way that makes sense to humans, but systematic automated methods that would be able to hypothesize and verify interpretations of hidden neuron activations are underexplored. In this paper, we provide such a method and demonstrate that it provides meaningful interpretations. Our approach is based on using large-scale background knowledge approximately 2 million classes curated from the Wikipedia concept hierarchy together with a symbolic reasoning approach called Concept Induction based on description logics, originally developed for applications in the Semantic Web field. Our results show that we can automatically attach meaningful labels from the background knowledge to individual neurons in the dense layer of a Convolutional Neural Network through a hypothesis and verification process.
A Modular Ontology for MODS -- Metadata Object Description Schema
Rayan, Rushrukh, Shimizu, Cogan, Sieverding, Heidi, Hitzler, Pascal
The Metadata Object Description Schema (MODS) was developed to describe bibliographic concepts and metadata and is maintained by the Library of Congress. Its authoritative version is given as an XML schema based on an XML mindset which means that it has significant limitations for use in a knowledge graphs context. We have therefore developed the Modular MODS Ontology (MMODS-O) which incorporates all elements and attributes of the MODS XML schema. In designing the ontology, we adopt the recent Modular Ontology Design Methodology (MOMo) with the intention to strike a balance between modularity and quality ontology design on the one hand, and conservative backward compatibility with MODS on the other.
An Ontology Design Pattern for Role-Dependent Names
Rayan, Rushrukh, Shimizu, Cogan, Hitzler, Pascal
We present an ontology design pattern for modeling Names as part of Roles, to capture scenarios where an Agent performs different Roles using different Names associated with the different Roles. Examples of an Agent performing a Role using different Names are rather ubiquitous, e.g., authors who write under different pseudonyms, or different legal names for citizens of more than one country. The proposed pattern is a modified merger of a standard Agent Role and a standard Name pattern stub.
Explaining Deep Learning Hidden Neuron Activations using Concept Induction
Dalal, Abhilekha, Sarker, Md Kamruzzaman, Barua, Adrita, Hitzler, Pascal
One of the current key challenges in Explainable AI is in correctly interpreting activations of hidden neurons. It seems evident that accurate interpretations thereof would provide insights into the question what a deep learning system has internally \emph{detected} as relevant on the input, thus lifting some of the black box character of deep learning systems. The state of the art on this front indicates that hidden node activations appear to be interpretable in a way that makes sense to humans, at least in some cases. Yet, systematic automated methods that would be able to first hypothesize an interpretation of hidden neuron activations, and then verify it, are mostly missing. In this paper, we provide such a method and demonstrate that it provides meaningful interpretations. It is based on using large-scale background knowledge -- a class hierarchy of approx. 2 million classes curated from the Wikipedia Concept Hierarchy -- together with a symbolic reasoning approach called \emph{concept induction} based on description logics that was originally developed for applications in the Semantic Web field. Our results show that we can automatically attach meaningful labels from the background knowledge to individual neurons in the dense layer of a Convolutional Neural Network through a hypothesis and verification process.
Towards Human-Compatible XAI: Explaining Data Differentials with Concept Induction over Background Knowledge
Widmer, Cara, Sarker, Md Kamruzzaman, Nadella, Srikanth, Fiechter, Joshua, Juvina, Ion, Minnery, Brandon, Hitzler, Pascal, Schwartz, Joshua, Raymer, Michael
Concept induction, which is based on formal logical reasoning over description logics, has been used in ontology engineering in order to create ontology (TBox) axioms from the base data (ABox) graph. In this paper, we show that it can also be used to explain data differentials, for example in the context of Explainable AI (XAI), and we show that it can in fact be done in a way that is meaningful to a human observer.