Goto

Collaborating Authors

 voltage


NeuralFuse: Learning to Recover the Accuracy of Access-Limited Neural Network Inference in Low-Voltage Regimes Hao-Lun Sun

Neural Information Processing Systems

Energy-efficient computing is of primary importance to the effective deployment of deep neural networks (DNNs), particularly in edge devices and in on-chip AI systems. Increasing DNN computation's energy efficiency and lowering its carbon footprint require iterative efforts from both chip designers and algorithm developers.


Scalable Bayesian inference of dendritic voltage via spatiotemporal recurrent state space models

Ruoxi Sun, Scott Linderman, Ian Kinsella, Liam Paninski

Neural Information Processing Systems

Recent progress in the development of voltage indicators [1-8] has brought us closer to a longstanding goal incellular neuroscience: imaging the full spatiotemporal voltageonadendritic tree. These recordings have the potential (pun not intended) to resolve fundamental questions about the computations performed by dendrites -- questions that have remained open for more than a century[9,10].





Medieval plague victims likely found in mass grave in Germany

Popular Science

Archaeologists say they located a Black Death burial site containing some of a village's 12,000 dead. Breakthroughs, discoveries, and DIY tips sent six days a week. The Black Death () killed as much as half of Europe's total population between 1346 and 1353, so there are a of bodies buried across the continent. For example, contemporary accounts from Thuringia--a state in central Germany--report that about 12,000 plague victims died around Erfurt amid the city's outbreak in 1350. But despite multiple accounts attesting to this devastation, none of the 11 mass graves could be pinpointed for centuries.


Scalable Bayesian inference of dendritic voltage via spatiotemporal recurrent state space models

Neural Information Processing Systems

Recent advances in optical voltage sensors have brought us closer to a critical goal in cellular neuroscience: imaging the full spatiotemporal voltage on a dendritic tree. However, current sensors and imaging approaches still face significant limitations in SNR and sampling frequency; therefore statistical denoising and interpolation methods remain critical for understanding single-trial spatiotemporal dendritic voltage dynamics. Previous denoising approaches were either based on an inadequate linear voltage model or scaled poorly to large trees. Here we introduce a scalable fully Bayesian approach. We develop a generative nonlinear model that requires few parameters per compartment of the cell but is nonetheless flexible enough to sample realistic spatiotemporal data. The model captures different dynamics in each compartment and leverages biophysical knowledge to constrain intra-and inter-compartmental dynamics. We obtain a full posterior distribution over spatiotemporal voltage via an augmented Gibbs sampling algorithm. The nonlinear smoother model outperforms previously developed linear methods, and scales to much larger systems than previous methods based on sequential Monte Carlo approaches.


Random Feature Spiking Neural Networks

Gollwitzer, Maximilian, Dietrich, Felix

arXiv.org Artificial Intelligence

Spiking Neural Networks (SNNs) as Machine Learning (ML) models have recently received a lot of attention as a potentially more energy-efficient alternative to conventional Artificial Neural Networks. The non-differentiability and sparsity of the spiking mechanism can make these models very difficult to train with algorithms based on propagating gradients through the spiking non-linearity. We address this problem by adapting the paradigm of Random Feature Methods (RFMs) from Artificial Neural Networks (ANNs) to Spike Response Model (SRM) SNNs. This approach allows training of SNNs without approximation of the spike function gradient. Concretely, we propose a novel data-driven, fast, high-performance, and interpretable algorithm for end-to-end training of SNNs inspired by the SWIM algorithm for RFM-ANNs, which we coin S-SWIM. We provide a thorough theoretical discussion and supplementary numerical experiments showing that S-SWIM can reach high accuracies on time series forecasting as a standalone strategy and serve as an effective initialisation strategy before gradient-based training. Additional ablation studies show that our proposed method performs better than random sampling of network weights.


GAVINA: flexible aggressive undervolting for bit-serial mixed-precision DNN acceleration

Fornt, Jordi, Fontova-Musté, Pau, Gras, Adrian, Lahyani, Omar, Caro, Martí, Abella, Jaume, Moll, Francesc, Altet, Josep

arXiv.org Artificial Intelligence

Voltage overscaling, or undervolting, is an enticing approximate technique in the context of energy-efficient Deep Neural Network (DNN) acceleration, given the quadratic relationship between power and voltage. Nevertheless, its very high error rate has thwarted its general adoption. Moreover, recent undervolting accelerators rely on 8-bit arithmetic and cannot compete with state-of-the-art low-precision (<8b) architectures. To overcome these issues, we propose a new technique called Guarded Aggressive underVolting (GAV), which combines the ideas of undervolting and bit-serial computation to create a flexible approximation method based on aggressively lowering the supply voltage on a select number of least significant bit combinations. Based on this idea, we implement GAVINA (GAV mIxed-precisioN Accelerator), a novel architecture that supports arbitrary mixed precision and flexible undervolting, with an energy efficiency of up to 89 TOP/sW in its most aggressive configuration. By developing an error model of GAVINA, we show that GAV can achieve an energy efficiency boost of 20% via undervolting, with negligible accuracy degradation on ResNet-18.


Enhancing Large Language Models for Automated Homework Assessment in Undergraduate Circuit Analysis

Chen, Liangliang, Xie, Huiru, Qin, Zhihao, Guo, Yiming, Rohde, Jacqueline, Zhang, Ying

arXiv.org Artificial Intelligence

This research full paper presents an enhancement pipeline for large language models (LLMs) in assessing homework for an undergraduate circuit analysis course, aiming to improve LLMs' capacity to provide personalized support to electrical engineering students. Existing evaluations have demonstrated that GPT-4o possesses promising capabilities in assessing student homework in this domain. Building on these findings, we enhance GPT-4o's performance through multi-step prompting, contextual data augmentation, and the incorporation of targeted hints. These strategies effectively address common errors observed in GPT-4o's responses when using simple prompts, leading to a substantial improvement in assessment accuracy. Specifically, the correct response rate for GPT-4o increases from 74.71% to 97.70% after applying the enhanced prompting and augmented data on entry-level circuit analysis topics. This work lays a foundation for the effective integration of LLMs into circuit analysis instruction and, more broadly, into engineering education.