Learning in Log-Domain: Subthreshold Analog AI Accelerator Based on Stochastic Gradient Descent
Tageldeen, Momen K, Belgaid, Yacine, Mohan, Vivek, Wang, Zhou, Drakakis, Emmanuel M
–arXiv.org Artificial Intelligence
In recent years, artificial intelligence (AI) has become an integral part of daily life, serving as a transformative tool across various professional domains [1] and driving personal applications through advancements in transformer models that power large language models (LLMs) [2]. However, both training and inference of AI models demand substantial computational and energy resources, which are becoming increasingly challenging to access [3, 4]. While server-class GPUs are effective for training, their energy inefficiency [5] and high costs present significant barriers [6]. Additionally, the environmental impact of energy-intensive AI systems has raised critical concerns about their role in exacerbating climate change [4]. Amdahl's law predicts that performance and efficiency gains are best achieved through innovative application-specific accelerator architectures rather than scaling up multi-core general-purpose processors [7]. Consequently, applicationspecific integrated circuits (ASICs), both digital and analog, have emerged as critical solutions for enabling highefficiency training and inference of artificial neural networks [7, 8, 9]. Digital accelerators are widely adopted for training workloads. Notable examples include the Brainwave Neural Processing Unit (NPU) [10], Google's Tensor Processing Unit (TPU) [11], and low-precision inference accelerators such as YodaNN [5], the Unified Neural Processing Unit (UNPU) [12], and BRein Memory [13].
arXiv.org Artificial Intelligence
Jan-22-2025
- Country:
- Europe > Italy (0.28)
- North America > United States
- Massachusetts (0.14)
- Genre:
- Research Report (0.50)
- Industry:
- Energy (0.86)
- Technology: