Goto

Collaborating Authors

 input example


5d0d5594d24f0f955548f0fc0ff83d10-Supplemental.pdf

Neural Information Processing Systems

Onemightconsider"2V 7 V"and "V 84V"tobedifferent patterns orinvariants butatahigher levelofabstraction theycan both represent the concept of a repeated symbol irrespective of the position of the repeating item.




Neural Path Features and Neural Path Kernel : Understanding the role of gates in deep learning

Neural Information Processing Systems

Rectified linear unit (ReLU) activations can also be thought of as'gates', which, either pass or stop their pre-activation input when they are'on' (when the pre-activation input is positive) or'off' (when the pre-activation input is negative) respectively. A deep neural network (DNN) with ReLU activations has many gates, and the on/off status of each gate changes across input examples as well as network weights. For a given input example, only a subset of gates are'active', i.e., on, and the sub-network of weights connected to these active gates is responsible for producing the output. At randomised initialisation, the active sub-network corresponding to a given input example is random. During training, as the weights are learnt, the active sub-networks are also learnt, and could hold valuable information.


Neural Path Features and Neural Path Kernel: Understanding the role of gates in deep learning Chandrashekar Lakshminarayanan and Amit Vikram Singh

Neural Information Processing Systems

A deep neural network (DNN) with ReLU activations has many gates, and the on/off status of each gate changes across input examples as well as network weights. For a given input example, only a subset of gates are active, i.e., on, and the sub-network of weights connected to these active gates is responsible for producing


mhealth_ood_neurips_2021.pdf

Neural Information Processing Systems

In this section, we provide screenshots and list of examples that were used in the user study. Note that the name of the institution is redacted for the review. This shows an example interface for skin cancer classifier. Figure 4: Interface to display different input data types. Figure 5: List of input examples used in the user study.


Program Semantic Inequivalence Game with Large Language Models

Miceli-Barone, Antonio Valerio, Belle, Vaishak, Payani, Ali

arXiv.org Artificial Intelligence

Large Language Models (LLMs) can achieve strong performance on everyday coding tasks, but they can fail on complex tasks that require non-trivial reasoning about program semantics. Finding training examples to teach LLMs to solve these tasks can be challenging. In this work, we explore a method to synthetically generate code reasoning training data based on a semantic inequivalence game SInQ: a generator agent creates program variants that are semantically distinct, derived from a dataset of real-world programming tasks, while an evaluator agent has to identify input examples that cause the original programs and the generated variants to diverge in their behaviour, with the agents training each other semi-adversarially. We prove that this setup enables theoretically unlimited improvement through self-play in the limit of infinite computational resources. We evaluated our approach on multiple code generation and understanding benchmarks, including cross-language vulnerability detection (Lu et al., 2021), where our method improves vulnerability detection in C/C++ code despite being trained exclusively on Python code, and the challenging Python builtin identifier swap benchmark (Miceli-Barone et al., 2023), showing that whereas modern LLMs still struggle with this benchmark, our approach yields substantial improvements. We release the code needed to replicate the experiments, as well as the generated synthetic data, which can be used to fine-tune LLMs.


Neural Path Features and Neural Path Kernel : Understanding the role of gates in deep learning

Neural Information Processing Systems

Rectified linear unit (ReLU) activations can also be thought of as'gates', which, either pass or stop their pre-activation input when they are'on' (when the pre-activation input is positive) or'off' (when the pre-activation input is negative) respectively. A deep neural network (DNN) with ReLU activations has many gates, and the on/off status of each gate changes across input examples as well as network weights. For a given input example, only a subset of gates are'active', i.e., on, and the sub-network of weights connected to these active gates is responsible for producing the output. At randomised initialisation, the active sub-network corresponding to a given input example is random. During training, as the weights are learnt, the active sub-networks are also learnt, and could hold valuable information. To this end, we encode the on/off state of the gates for a given input in a novel'neural path feature' (NPF), and the weights of the DNN are encoded in a novel'neural path value' (NPV).


Do Llamas Work in English? On the Latent Language of Multilingual Transformers

Wendler, Chris, Veselovsky, Veniamin, Monea, Giovanni, West, Robert

arXiv.org Artificial Intelligence

We ask whether multilingual language models trained on unbalanced, English-dominated corpora use English as an internal pivot language -- a question of key importance for understanding how language models function and the origins of linguistic bias. Focusing on the Llama-2 family of transformer models, our study uses carefully constructed non-English prompts with a unique correct single-token continuation. From layer to layer, transformers gradually map an input embedding of the final prompt token to an output embedding from which next-token probabilities are computed. Tracking intermediate embeddings through their high-dimensional space reveals three distinct phases, whereby intermediate embeddings (1) start far away from output token embeddings; (2) already allow for decoding a semantically correct next token in the middle layers, but give higher probability to its version in English than in the input language; (3) finally move into an input-language-specific region of the embedding space. We cast these results into a conceptual model where the three phases operate in "input space", "concept space", and "output space", respectively. Crucially, our evidence suggests that the abstract "concept space" lies closer to English than to other languages, which may have important consequences regarding the biases held by multilingual language models.


MaNtLE: Model-agnostic Natural Language Explainer

Menon, Rakesh R., Zaman, Kerem, Srivastava, Shashank

arXiv.org Artificial Intelligence

Understanding the internal reasoning behind the predictions of machine learning systems is increasingly vital, given their rising adoption and acceptance. While previous approaches, such as LIME, generate algorithmic explanations by attributing importance to input features for individual examples, recent research indicates that practitioners prefer examining language explanations that explain sub-groups of examples. In this paper, we introduce MaNtLE, a model-agnostic natural language explainer that analyzes multiple classifier predictions and generates faithful natural language explanations of classifier rationale for structured classification tasks. MaNtLE uses multi-task training on thousands of synthetic classification tasks to generate faithful explanations. Simulated user studies indicate that, on average, MaNtLE-generated explanations are at least 11% more faithful compared to LIME and Anchors explanations across three tasks. Human evaluations demonstrate that users can better predict model behavior using explanations from MaNtLE compared to other techniques