Goto

Collaborating Authors

 seneca


From RISC-V Cores to Neuromorphic Arrays: A Tutorial on Building Scalable Digital Neuromorphic Processors

Yousefzadeh, Amirreza

arXiv.org Artificial Intelligence

Digital neuromorphic processors are emerging as a promising computing substrate for low-power, always-on EdgeAI applications. In this tutorial paper, we outline the main architectural design principles behind fully digital neuromorphic processors and illustrate them using the SENECA platform as a running example. Starting from a flexible array of tiny RISC-V processing cores connected by a simple Network-on-Chip (NoC), we show how to progressively evolve the architecture: from a baseline event-driven implementation of fully connected networks, to versions with dedicated Neural Processing Elements (NPEs) and a loop controller that offloads fine-grained control from the general-purpose cores. Along the way, we discuss software and mapping techniques such as spike grouping, event-driven depth-first convolution for convolutional networks, and hard-attention style processing for high-resolution event-based vision. The focus is on architectural trade-offs, performance and energy bottlenecks, and on leveraging flexibility to incrementally add domain-specific acceleration. This paper assumes familiarity with basic neuromorphic concepts (spikes, event-driven computation, sparse activation) and deep neural network workloads. It does not present new experimental results; instead, it synthesizes and contextualizes findings previously reported in our SENECA publications to provide a coherent, step-by-step architectural perspective for students and practitioners who wish to design their own digital neuromorphic processors.


Preparation Meets Opportunity: Enhancing Data Preprocessing for ML Training With Seneca

Desai, Omkar, Jiao, Ziyang, Pei, Shuyi, Bhimani, Janki, Kim, Bryan S.

arXiv.org Artificial Intelligence

Input data preprocessing is a common bottleneck when concurrently training multimedia machine learning (ML) models in modern systems. To alleviate these bottlenecks and reduce the training time for concurrent jobs, we present Seneca, a data loading system that optimizes cache partitioning and data sampling for the data storage and ingestion (DSI) pipeline. The design of Seneca contains two key techniques. First, Seneca uses a performance model for the data pipeline to optimally partition the cache for three different forms of data (encoded, decoded, and augmented). Second, Seneca opportunistically serves cached data over uncached ones during random batch sampling so that concurrent jobs benefit from each other. We implement Seneca by modifying PyTorch and demonstrate its effectiveness by comparing it against several state-of-the-art caching systems for DNN training. Seneca reduces the makespan by 45.23% compared to PyTorch and increases data processing throughput by up to 3.45x compared to the next best dataloader.


Efficient Synaptic Delay Implementation in Digital Event-Driven AI Accelerators

Meijer, Roy, Detterer, Paul, Yousefzadeh, Amirreza, Patino-Saucedo, Alberto, Tang, Guanghzi, Vadivel, Kanishkan, Xu, Yinfu, Gomony, Manil-Dev, Corradi, Federico, Linares-Barranco, Bernabe, Sifalakis, Manolis

arXiv.org Artificial Intelligence

Synaptic delay parameterization of neural network models have remained largely unexplored but recent literature has been showing promising results, suggesting the delay parameterized models are simpler, smaller, sparser, and thus more energy efficient than similar performing (e.g. task accuracy) non-delay parameterized ones. We introduce Shared Circular Delay Queue (SCDQ), a novel hardware structure for supporting synaptic delays on digital neuromorphic accelerators. Our analysis and hardware results show that it scales better in terms of memory, than current commonly used approaches, and is more amortizable to algorithm-hardware co-optimizations, where in fact, memory scaling is modulated by model sparsity and not merely network size. Next to memory we also report performance on latency area and energy per inference.


Hardware-aware training of models with synaptic delays for digital event-driven neuromorphic processors

Patino-Saucedo, Alberto, Meijer, Roy, Yousefzadeh, Amirreza, Gomony, Manil-Dev, Corradi, Federico, Detteter, Paul, Garrido-Regife, Laura, Linares-Barranco, Bernabe, Sifalakis, Manolis

arXiv.org Artificial Intelligence

Configurable synaptic delays are a basic feature in many neuromorphic neural network hardware accelerators. However, they have been rarely used in model implementations, despite their promising impact on performance and efficiency in tasks that exhibit complex (temporal) dynamics, as it has been unclear how to optimize them. In this work, we propose a framework to train and deploy, in digital neuromorphic hardware, highly performing spiking neural network models (SNNs) where apart from the synaptic weights, the per-synapse delays are also co-optimized. Leveraging spike-based back-propagation-through-time, the training accounts for both platform constraints, such as synaptic weight precision and the total number of parameters per core, as a function of the network size. In addition, a delay pruning technique is used to reduce memory footprint with a low cost in performance. We evaluate trained models in two neuromorphic digital hardware platforms: Intel Loihi and Imec Seneca. Loihi offers synaptic delay support using the so-called Ring-Buffer hardware structure. Seneca does not provide native hardware support for synaptic delays. A second contribution of this paper is therefore a novel area- and memory-efficient hardware structure for acceleration of synaptic delays, which we have integrated in Seneca. The evaluated benchmark involves several models for solving the SHD (Spiking Heidelberg Digits) classification task, where minimal accuracy degradation during the transition from software to hardware is demonstrated. To our knowledge, this is the first work showcasing how to train and deploy hardware-aware models parameterized with synaptic delays, on multicore neuromorphic hardware accelerators.


Louisiana Jeffrey Dahmer copycat sentenced for Grindr dating app scheme to kidnap, murder men

FOX News

On a recent episode of Dr. Phil, the host spoke with some of Jeffrey Dahmer's victims and showed them an interview he filmed with the father of one of America's most infamous serial killers. A 21-year-old Louisiana man has been sentenced to 45 years in prison after plotting a Jeffrey Dahmer-like scheme to meet men on the gay dating app Grindr and kill them, according to federal officials. Chance Seneca of Lafayette Parish targeted one particular victim, as well as other gay men, through the app in 2020 because of their sexual orientation and gender, the Justice Department said. "The facts of this case are truly shocking, and the defendant's decision to specifically target gay men is a disturbing reminder of the unique prejudices and dangers facing the LGBTQ community today," Assistant Attorney General Kristen Clarke of the Justice Department's Civil Rights Division said in a Wednesday statement. Clarke continued: "The internet should be accessible and safe for all Americans, regardless of their gender or sexual orientation. We will continue to identify and intercept the predators who weaponize online platforms to target LGBTQ victims and carry out acts of violence and hate."


Empirical Analysis of Foundational Distinctions in Linked Open Data

Asprino, Luigi, Basile, Valerio, Ciancarini, Paolo, Presutti, Valentina

arXiv.org Artificial Intelligence

The Web and its Semantic extension (i.e. Linked Open Data) contain open global-scale knowledge and make it available to potentially intelligent machines that want to benefit from it. Nevertheless, most of Linked Open Data lack ontological distinctions and have sparse axiomatisation. For example, distinctions such as whether an entity is inherently a class or an individual, or whether it is a physical object or not, are hardly expressed in the data, although they have been largely studied and formalised by foundational ontologies (e.g. DOLCE, SUMO). These distinctions belong to common sense too, which is relevant for many artificial intelligence tasks such as natural language understanding, scene recognition, and the like. There is a gap between foundational ontologies, that often formalise or are inspired by pre-existing philosophical theories and are developed with a top-down approach, and Linked Open Data that mostly derive from existing databases or crowd-based effort (e.g. DBpedia, Wikidata). We investigate whether machines can learn foundational distinctions over Linked Open Data entities, and if they match common sense. We want to answer questions such as "does the DBpedia entity for dog refer to a class or to an instance?". We report on a set of experiments based on machine learning and crowdsourcing that show promising results.


Automated Attribution and Intertextual Analysis

Brofos, James, Kannan, Ajay, Shu, Rui

arXiv.org Machine Learning

In this work, we employ quantitative methods from the realm of statistics and machine learning to develop novel methodologies for author attribution and textual analysis. In particular, we develop techniques and software suitable for applications to Classical study, and we illustrate the efficacy of our approach in several interesting open questions in the field. We apply our numerical analysis techniques to questions of authorship attribution in the case of the Greek tragedian Euripides, to instances of intertextuality and influence in the poetry of the Roman statesman Seneca the Younger, and to cases of "interpolated" text with respect to the histories of Livy.