Hinder, Fabian
Conceptualizing Uncertainty
Roberts, Isaac, Schulz, Alexander, Schroeder, Sarah, Hinder, Fabian, Hammer, Barbara
While advances in deep learning in the last years have led to impressive performance in many domains, such models are not always reliable, particularly when it comes to generalizing to new environments or adversarial attacks. To improve on that, numerous methods have been developed in the field of explainable artificial intelligence (xAI) [5] to provide insights into model behavior and facilitate actionable modifications. However, the majority of methods focus on explaining model predictions, which can help understand misclassifications but do not explicitly address predictive uncertainty(See Figure 1). Understanding uncertainty is crucial for detecting potential model weaknesses, particularly in dynamic environments. Since uncertainty quantification is useful in various applications, including active learning [20], classification with rejects [17], adversarial example detection [26], and reinforcement learning [24], a significant body of work aims to improve the quantification of predictive uncertainty using Bayesian deep learning (BDL) and approximations thereof [15,9,14]. In contrast, the literature on understanding the sources of uncertainty for a given model via explanations is limited, focusing on methods for feature attribution [28,27] (see section 2.4 for more related
Continual Learning Should Move Beyond Incremental Classification
Mitchell, Rupert, Alliegro, Antonio, Camoriano, Raffaello, Carrión-Ojeda, Dustin, Carta, Antonio, Chalvatzaki, Georgia, Churamani, Nikhil, D'Eramo, Carlo, Hamidi, Samin, Hesse, Robin, Hinder, Fabian, Kamath, Roshni Ramanna, Lomonaco, Vincenzo, Paul, Subarnaduti, Pistilli, Francesca, Tuytelaars, Tinne, van de Ven, Gido M, Kersting, Kristian, Schaub-Meyer, Simone, Mundt, Martin
Continual learning (CL) is the sub-field of machine learning concerned with accumulating knowledge in dynamic environments. So far, CL research has mainly focused on incremental classification tasks, where models learn to classify new categories while retaining knowledge of previously learned ones. Here, we argue that maintaining such a focus limits both theoretical development and practical applicability of CL methods. Through a detailed analysis of concrete examples - including multi-target classification, robotics with constrained output spaces, learning in continuous task domains, and higher-level concept memorization - we demonstrate how current CL approaches often fail when applied beyond standard classification. We identify three fundamental challenges: (C1) the nature of continuity in learning problems, (C2) the choice of appropriate spaces and metrics for measuring similarity, and (C3) the role of learning objectives beyond classification. For each challenge, we provide specific recommendations to help move the field forward, including formalizing temporal dynamics through distribution processes, developing principled approaches for continuous task spaces, and incorporating density estimation and generative objectives. In so doing, this position paper aims to broaden the scope of CL research while strengthening its theoretical foundations, making it more applicable to real-world problems.
An Algorithm-Centered Approach To Model Streaming Data
Hinder, Fabian, Vaquet, Valerie, Komnick, David, Hammer, Barbara
Besides the classical offline setup of machine learning, stream learning constitutes a well-established setup where data arrives over time in potentially non-stationary environments. Concept drift, the phenomenon that the underlying distribution changes over time poses a significant challenge. Yet, despite high practical relevance, there is little to no foundational theory for learning in the drifting setup comparable to classical statistical learning theory in the offline setting. This can be attributed to the lack of an underlying object comparable to a probability distribution as in the classical setup. While there exist approaches to transfer ideas to the streaming setup, these start from a data perspective rather than an algorithmic one. In this work, we suggest a new model of data over time that is aimed at the algorithm's perspective. Instead of defining the setup using time points, we utilize a window-based approach that resembles the inner workings of most stream learning algorithms. We compare our framework to others from the literature on a theoretical basis, showing that in many cases both model the same situation. Furthermore, we perform a numerical evaluation and showcase an application in the domain of critical infrastructure.
Adversarial Attacks for Drift Detection
Hinder, Fabian, Vaquet, Valerie, Hammer, Barbara
Data from the real world is often subject to continuous changes known as concept drift [1, 2, 3]. Such can be caused by seasonal changes, changed demands, aging of sensors, etc. Concept drift not only poses a problem for maintaining high performance in learning models [2, 3] but also plays a crucial role in system monitoring [1]. In the latter case, the detection of concept drift is crucial as it enables the detection of anomalous behavior. Examples include machine malfunctions or failures, network security, environmental changes, and critical infrastructures. This is done by detecting irregular shifts [4, 1, 5]. In these contexts, the ability to robustly detect drift is essential. In addition to problems such as noise and sampling error, which challenge all statistical methods, drift detection faces a special kind of difficulty when the drift follows certain patterns that evade detection. In this work, we study those specific drifts that we will refer to as "drift adversarials". Similar to adversarial attacks, drift adversarials exploit weaknesses in the detection methods, and thus allow significant concept drift to occur without triggering alarms posing major issues for monitoring systems.
Challenges, Methods, Data -- a Survey of Machine Learning in Water Distribution Networks
Vaquet, Valerie, Hinder, Fabian, Artelt, André, Ashraf, Inaam, Strotherm, Janine, Vaquet, Jonas, Brinkrolf, Johannes, Hammer, Barbara
Research on methods for planning and controlling water distribution networks gains increasing relevance as the availability of drinking water will decrease as a consequence of climate change. So far, the majority of approaches is based on hydraulics and engineering expertise. However, with the increasing availability of sensors, machine learning techniques constitute a promising tool. This work presents the main tasks in water distribution networks, discusses how they relate to machine learning and analyses how the particularities of the domain pose challenges to and can be leveraged by machine learning approaches. Besides, it provides a technical toolkit by presenting evaluation benchmarks and a structured survey of the exemplary task of leakage detection and localization.
FairGLVQ: Fairness in Partition-Based Classification
Störck, Felix, Hinder, Fabian, Brinkrolf, Johannes, Paassen, Benjamin, Vaquet, Valerie, Hammer, Barbara
Fairness is an important objective throughout society. From the distribution of limited goods such as education, over hiring and payment, to taxes, legislation, and jurisprudence. Due to the increasing importance of machine learning approaches in all areas of daily life including those related to health, security, and equity, an increasing amount of research focuses on fair machine learning. In this work, we focus on the fairness of partition- and prototype-based models. The contribution of this work is twofold: 1) we develop a general framework for fair machine learning of partition-based models that does not depend on a specific fairness definition, and 2) we derive a fair version of learning vector quantization (LVQ) as a specific instantiation. We compare the resulting algorithm against other algorithms from the literature on theoretical and real-world data showing its practical relevance.
Localizing Anomalies in Critical Infrastructure using Model-Based Drift Explanations
Vaquet, Valerie, Hinder, Fabian, Vaquet, Jonas, Lammers, Kathrin, Quakernack, Lars, Hammer, Barbara
Facing climate change, the already limited availability of drinking water will decrease in the future rendering drinking water an increasingly scarce resource. Considerable amounts of it are lost through leakages in water transportation and distribution networks. Thus, anomaly detection and localization, in particular for leakages, are crucial but challenging tasks due to the complex interactions and changing demands in water distribution networks. In this work, we analyze the effects of anomalies on the dynamics of critical infrastructure systems by modeling the networks employing Bayesian networks. We then discuss how the problem is connected to and can be considered through the lens of concept drift. In particular, we argue that model-based explanations of concept drift are a promising tool for localizing anomalies given limited information about the network. The methodology is experimentally evaluated using realistic benchmark scenarios. To showcase that our methodology applies to critical infrastructure more generally, in addition to considering leakages and sensor faults in water systems, we showcase the suitability of the derived technique to localize sensor faults in power systems.
Semantic Properties of cosine based bias scores for word embeddings
Schröder, Sarah, Schulz, Alexander, Hinder, Fabian, Hammer, Barbara
In the domain of Natural Language Processing (NLP), many works have investigated social biases in terms of associations in the embeddings space. Early works [1, 2] introduced methods to measure and mitigate social biases based on cosine similarity in word embeddigs. With NLP research progressing to large language models and contextualized embeddings, doubts have been raised whether these methods are still suitable for fairness evaluation [3] and other works criticize that for instance the Word Embedding Association Test (WEAT) [2] fails to detect some kinds of biases [4, 5]. Overall there exists a great deal of bias measures in the literature, which not necessarily detect the same biases [6, 4, 5]. In general, researchers are questioning the usability of model intrinsic bias measures, such as cosine based methods [7, 8, 9]. There exist few papers that compare the performance of different bias scores [10, 11] and works that evaluate experimental setups for bias measurement [12]. However, to our knowledge, only two works investigate the properties of intrinsic bias scores on a theoretical level [5, 13]. To further close this gap, we evaluate the semantic properties of cosine based bias scores, focusing on bias quantification as opposed to bias detection. We make the following contributions: (i) We formalize the properties of trustworthiness and comparability as requirements for cosine based bias scores.
Investigating the Suitability of Concept Drift Detection for Detecting Leakages in Water Distribution Networks
Vaquet, Valerie, Hinder, Fabian, Hammer, Barbara
Leakages are a major risk in water distribution networks as they cause water loss and increase contamination risks. Leakage detection is a difficult task due to the complex dynamics of water distribution networks. In particular, small leakages are hard to detect. From a machine-learning perspective, leakages can be modeled as concept drift. Thus, a wide variety of drift detection schemes seems to be a suitable choice for detecting leakages. In this work, we explore the potential of model-loss-based and distribution-based drift detection methods to tackle leakage detection. We additionally discuss the issue of temporal dependencies in the data and propose a way to cope with it when applying distribution-based detection. We evaluate different methods systematically for leakages of different sizes and detection times. Additionally, we propose a first drift-detection-based technique for localizing leakages.
A Remark on Concept Drift for Dependent Data
Hinder, Fabian, Vaquet, Valerie, Hammer, Barbara
Concept drift, i.e., the change of the data generating distribution, can render machine learning models inaccurate. Several works address the phenomenon of concept drift in the streaming context usually assuming that consecutive data points are independent of each other. To generalize to dependent data, many authors link the notion of concept drift to time series. In this work, we show that the temporal dependencies are strongly influencing the sampling process. Thus, the used definitions need major modifications. In particular, we show that the notion of stationarity is not suited for this setup and discuss alternatives. We demonstrate that these alternative formal notions describe the observable learning behavior in numerical experiments.