Not enough data to create a plot.
Try a different view from the menu above.
Information Technology
Meta-Gradient Reinforcement Learning
Zhongwen Xu, Hado P. van Hasselt, David Silver
The goal of reinforcement learning algorithms is to estimate and/or optimise the value function. However, unlike supervised learning, no teacher or oracle is available to provide the true value function. Instead, the majority of reinforcement learning algorithms estimate and/or optimise a proxy for the value function. This proxy is typically based on a sampled and bootstrapped approximation to the true value function, known as a return. The particular choice of return is one of the chief components determining the nature of the algorithm: the rate at which future rewards are discounted; when and how values should be bootstrapped; or even the nature of the rewards themselves.
Grid Saliency for Context Explanations of Semantic Segmentation
Lukas Hoyer, Mauricio Munoz, Prateek Katiyar, Anna Khoreva, Volker Fischer
Recently, there has been a growing interest in developing saliency methods that provide visual explanations of network predictions. Still, the usability of existing methods is limited to image classification models. To overcome this limitation, we extend the existing approaches to generate grid saliencies, which provide spatially coherent visual explanations for (pixel-level) dense prediction networks. As the proposed grid saliency allows to spatially disentangle the object and its context, we specifically explore its potential to produce context explanations for semantic segmentation networks, discovering which context most influences the class predictions inside a target object area. We investigate the effectiveness of grid saliency on a synthetic dataset with an artificially induced bias between objects and their context as well as on the real-world Cityscapes dataset using state-of-the-art segmentation networks. Our results show that grid saliency can be successfully used to provide easily interpretable context explanations and, moreover, can be employed for detecting and localizing contextual biases present in the data.
Diff-eRank: A Novel Rank-Based Metric for Evaluating Large Language Models
Large Language Models (LLMs) have transformed natural language processing and extended their powerful capabilities to multi-modal domains. As LLMs continue to advance, it is crucial to develop diverse and appropriate metrics for their evaluation. In this paper, we introduce a novel rank-based metric, Diff-eRank, grounded in information theory and geometry principles. Diff-eRank assesses LLMs by analyzing their hidden representations, providing a quantitative measure of how efficiently they eliminate redundant information during training. We demonstrate the applicability of Diff-eRank in both single-modal (e.g., language) and multimodal settings. For language models, our results show that Diff-eRank increases with model size and correlates well with conventional metrics such as loss and accuracy. In the multi-modal context, we propose an alignment evaluation method based on the eRank, and verify that contemporary multi-modal LLMs exhibit strong alignment performance based on our method.
Dendritic cortical microcircuits approximate the backpropagation algorithm
Joรฃo Sacramento, Rui Ponte Costa, Yoshua Bengio, Walter Senn
Deep learning has seen remarkable developments over the last years, many of them inspired by neuroscience. However, the main learning mechanism behind these advances - error backpropagation - appears to be at odds with neurobiology. Here, we introduce a multilayer neuronal network model with simplified dendritic compartments in which error-driven synaptic plasticity adapts the network towards a global desired output. In contrast to previous work our model does not require separate phases and synaptic learning is driven by local dendritic prediction errors continuously in time. Such errors originate at apical dendrites and occur due to a mismatch between predictive input from lateral interneurons and activity from actual top-down feedback. Through the use of simple dendritic compartments and different cell-types our model can represent both error and normal activity within a pyramidal neuron. We demonstrate the learning capabilities of the model in regression and classification tasks, and show analytically that it approximates the error backpropagation algorithm. Moreover, our framework is consistent with recent observations of learning between brain areas and the architecture of cortical microcircuits. Overall, we introduce a novel view of learning on dendritic cortical circuits and on how the brain may solve the long-standing synaptic credit assignment problem.
Trump signs executive orders to spur US 'nuclear energy renaissance'
Donald Trump signed a series of executive orders on Friday intended to spur a "nuclear energy renaissance" through the construction of new reactors he said would satisfy the electricity demands of data centers for artificial intelligence and other emerging industries. The orders represented the president's latest foray into the policy underlying America's electricity supply. Trump declared a national energy emergency on his first day in office over and moved to undo a ban implemented by Joe Biden on new natural gas export terminals and expand oil and gas drilling in Alaska. Nuclear does not carry oil and gas's carbon emissions, but produces radioactive waste that the United States lacks a facility to permanently store. Some environmental groups have safety concerns over the reactors and their supply chain. Trump signed four orders intended to speed up the approval of nuclear reactors for defense and AI purposes, reform the Nuclear Regulatory Commission with the goal of quadrupling production of electricity over the next 25 years, revamp the regulatory process to have three experimental reactors operating by 4 July 2026 and boost investment in the technology's industrial base.
D-Wave revives 'quantum supremacy' claims for new Advantage2 computer
Quantum computing pioneer D-Wave Quantum on Tuesday announced the general availability of its sixth-generation quantum computer, the Advantage2. The company said the Advantage2 offers orders-of-magnitude greater performance compared to its prior system, expanding the tasks the company can accomplish in optimization problems. The machine even achieves the long-sought goal of quantum "supremacy," says the company, despite that term's highly controversial past. "This is a really historic moment for both D-Wave and the quantum computing industry," said D-Wave CEO Alan Baratz in an interview via Zoom. "Fundamentally, our technology is doing something that can't be touched classically."
On the Power of Small-size Graph Neural Networks for Linear Programming
Graph neural networks (GNNs) have recently emerged as powerful tools for addressing complex optimization problems. It has been theoretically demonstrated that GNNs can universally approximate the solution mapping functions of linear programming (LP) problems. However, these theoretical results typically require GNNs to have large parameter sizes. Conversely, empirical experiments have shown that relatively small GNNs can solve LPs effectively, revealing a significant discrepancy between theoretical predictions and practical observations. In this work, we aim to bridge this gap by providing a theoretical foundation for the effectiveness of smaller GNNs. We prove that polylogarithmic-depth, constant-width GNNs are sufficient to solve packing and covering LPs, two widely used classes of LPs. Our proof leverages the capability of GNNs to simulate a variant of the gradient descent algorithm on a carefully selected potential function. Additionally, we introduce a new GNN architecture, termed GD-Net. Experimental results demonstrate that GD-Net significantly outperforms conventional GNN structures while using fewer parameters.