Emde, Cornelius
Shh, don't say that! Domain Certification in LLMs
Emde, Cornelius, Paren, Alasdair, Arvind, Preetham, Kayser, Maxime, Rainforth, Tom, Lukasiewicz, Thomas, Ghanem, Bernard, Torr, Philip H. S., Bibi, Adel
Large language models (LLMs) are often deployed to perform constrained tasks, with narrow domains. For example, customer support bots can be built on top of LLMs, relying on their broad language understanding and capabilities to enhance performance. However, these LLMs are adversarially susceptible, potentially generating outputs outside the intended domain. To formalize, assess, and mitigate this risk, we introduce domain certification; a guarantee that accurately characterizes the out-of-domain behavior of language models. We then propose a simple yet effective approach, which we call VALID that provides adversarial bounds as a certificate. Finally, we evaluate our method across a diverse set of datasets, demonstrating that it yields meaningful certificates, which bound the probability of out-of-domain samples tightly with minimum penalty to refusal behavior.
Fool Me Once? Contrasting Textual and Visual Explanations in a Clinical Decision-Support Setting
Kayser, Maxime, Menzat, Bayar, Emde, Cornelius, Bercean, Bogdan, Novak, Alex, Espinosa, Abdala, Papiez, Bartlomiej W., Gaube, Susanne, Lukasiewicz, Thomas, Camburu, Oana-Maria
The growing capabilities of AI models are leading to their wider use, including in safety-critical domains. Explainable AI (XAI) aims to make these models safer to use by making their inference process more transparent. However, current explainability methods are seldom evaluated in the way they are intended to be used: by real-world end users. To address this, we conducted a large-scale user study with 85 healthcare practitioners in the context of human-AI collaborative chest X-ray analysis. We evaluated three types of explanations: visual explanations (saliency maps), natural language explanations, and a combination of both modalities. We specifically examined how different explanation types influence users depending on whether the AI advice and explanations are factually correct. We find that text-based explanations lead to significant over-reliance, which is alleviated by combining them with saliency maps. We also observe that the quality of explanations, that is, how much factually correct information they entail, and how much this aligns with AI correctness, significantly impacts the usefulness of the different explanation types.
Benchmarking Predictive Coding Networks -- Made Simple
Pinchetti, Luca, Qi, Chang, Lokshyn, Oleh, Olivers, Gaspard, Emde, Cornelius, Tang, Mufeng, M'Charrak, Amine, Frieder, Simon, Menzat, Bayar, Bogacz, Rafal, Lukasiewicz, Thomas, Salvatori, Tommaso
In this work, we tackle the problems of efficiency and scalability for predictive coding networks in machine learning. To do so, we first propose a library called PCX, whose focus lies on performance and simplicity, and provides a user-friendly, deep-learning oriented interface. Second, we use PCX to implement a large set of benchmarks for the community to use for their experiments. As most works propose their own tasks and architectures, do not compare one against each other, and focus on small-scale tasks, a simple and fast open-source library adopted by the whole community would address all of these concerns. Third, we perform extensive benchmarks using multiple algorithms, setting new state-of-the-art results in multiple tasks and datasets, as well as highlighting limitations inherent to PC that should be addressed. Thanks to the efficiency of PCX, we are able to analyze larger architectures than commonly used, providing baselines to galvanize community efforts towards one of the main open problems in the field: scalability. The code for PCX is available at https://github.com/liukidar/pcax.
Towards Certification of Uncertainty Calibration under Adversarial Attacks
Emde, Cornelius, Pinto, Francesco, Lukasiewicz, Thomas, Torr, Philip H. S., Bibi, Adel
Since neural classifiers are known to be sensitive to adversarial perturbations that alter their accuracy, \textit{certification methods} have been developed to provide provable guarantees on the insensitivity of their predictions to such perturbations. Furthermore, in safety-critical applications, the frequentist interpretation of the confidence of a classifier (also known as model calibration) can be of utmost importance. This property can be measured via the Brier score or the expected calibration error. We show that attacks can significantly harm calibration, and thus propose certified calibration as worst-case bounds on calibration under adversarial perturbations. Specifically, we produce analytic bounds for the Brier score and approximate bounds via the solution of a mixed-integer program on the expected calibration error. Finally, we propose novel calibration attacks and demonstrate how they can improve model calibration through \textit{adversarial calibration training}.
Incremental Predictive Coding: A Parallel and Fully Automatic Learning Algorithm
Salvatori, Tommaso, Song, Yuhang, Millidge, Beren, Xu, Zhenghua, Sha, Lei, Emde, Cornelius, Bogacz, Rafal, Lukasiewicz, Thomas
In recent years, deep learning has reached and surpassed human-level performance in a multitude of tasks, such as game playing [Silver et al., 2017, 2016], image recognition [Krizhevsky et al., 2012, He et al., 2016], natural language processing [Chen et al., 2020], and image generation [Ramesh et al., 2022]. These successes are achieved entirely using deep artificial neural networks trained via backpropagation (BP), which is a learning algorithm that is often criticized for its biological implausibilities [Grossberg, 1987, Crick, 1989, Abdelghani et al., 2008, Lillicrap et al., 2016, Roelfsema and Holtmaat, 2018, Whittington and Bogacz, 2019], such as lacking local plasticity and autonomy. In fact, backpropagation requires a global control signal required to trigger computations, since gradients must be sequentially computed backwards through the computation graph. These properties are not only important for biological plausibility: parallelization, locality, and automation are key to build efficient models that can be trained end-to-end on non Von-Neumann machines, such as analog chips [Kendall et al., 2020]. A learning algorithm with most of the above properties is predictive coding (PC). PC is an influential theory of information processing in the brain [Mumford, 1992, Friston, 2005], where learning happens by minimizing the prediction error of every neuron. PC can be shown to approximate backpropagation in layered networks [Whittington and Bogacz, 2017], as well as on any other model [Millidge et al., 2020], and can exactly replicate its weight update if some external control is added [Salvatori et al., 2022a]. Also the differences with BP are interesting, as PC allows for a much more flexible training and testing [Salvatori et al., 2022b], has a rich mathematical formulation [Friston, 2005, Millidge et al., 2022], and is an energy-based model [Bogacz, 2017].