redistribution
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Asia > Middle East > Jordan (0.04)
- (5 more...)
- Education (0.46)
- Leisure & Entertainment > Games (0.30)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (0.34)
Redistribution of Weights and Activations for AdderNet Quantization
Adder Neural Network (AdderNet) provides a new way for developing energy-efficient neural networks by replacing the expensive multiplications in convolution with cheaper additions (i.e., L1-norm). To achieve higher hardware efficiency, it is necessary to further study the low-bit quantization of AdderNet. Due to the limitation that the commutative law in multiplication does not hold in L1-norm, the well-established quantization methods on convolutional networks cannot be applied on AdderNets. Thus, the existing AdderNet quantization techniques propose to use only one shared scale to quantize both the weights and activations simultaneously. Admittedly, such an approach can keep the commutative law in the L1-norm quantization process, while the accuracy drop after low-bit quantization cannot be ignored.
Why universal basic income still can't meet the challenges of an AI economy
A person holds a fake $1,000 bill signed by former Democratic presidential candidate Andrew Yang following a campaign event in Iowa City, Iowa, on 29 January 2020. A person holds a fake $1,000 bill signed by former Democratic presidential candidate Andrew Yang following a campaign event in Iowa City, Iowa, on 29 January 2020. Why universal basic income still can't meet the challenges of an AI economy Andrew Yang's revived pitch suits the automation debate, but UBI can't fix inequalities concentrated tech wealth drives Universal basic income (UBI) is back, like a space zombie in a sci-fi movie, resurrected from policy oblivion, hungry for policymakers' attention: brains! Andrew Yang, whose "Yang Gang" enthusiasm briefly shook up the Democratic presidential nomination in 2020 promoting a "Freedom Dividend" to save workers from automation - $1,000 a month for every American adult - is again the main carrier of the bug: offering UBI to save the nation when robots eat all our jobs. This time Chat GPT, Yang hopes, will help his argument land: if artificial intelligence truly makes human labor redundant, as so many citizens of the tech bubble in Silicon Valley expect, society will need something other than employment for all of us to make ends meet.
- North America > United States > Iowa > Johnson County > Iowa City (0.45)
- North America > United States > California (0.25)
- Oceania > Australia (0.05)
- Europe > Ukraine (0.05)
Physically-Based Simulation of Automotive LiDAR
Dudzik, L., Roschani, M., Sielemann, A., Trampert, K., Ziehn, J., Beyerer, J., Neumann, C.
Abstract--We present an analytic model for simulating automotive time-of-flight (T oF) LiDAR that includes blooming, echo pulse width, and ambient light, along with steps to determine model parameters systematically through optical laboratory measurements. The model uses physically based rendering (PBR) in the near-infrared domain. It assumes single-bounce reflections and retroreflections over rasterized rendered images from shading or ray tracing, including light emitted from the sensor as well as stray light from other, non-correlated sources such as sunlight. Beams from the sensor and sensitivity of the receiving diodes are modeled with flexible beam steering patterns and with non-vanishing diameter . Different (all non-real time) computational approaches can be chosen based on system properties, computing capabilities, and desired output properties. Model parameters include system-specific properties, namely the physical spread of the LiDAR beam, combined with the sensitivity of the receiving diode; the intensity of the emitted light; the conversion between the intensity of reflected light and the echo pulse width; and scenario parameters such as environment lighting, positioning, and surface properties of the target(s) in the relevant infrared domain. System-specific properties of the model are determined from laboratory measurements of the photometric luminance on different target surfaces aligned with a goniometer at 0.01 resolution, which marks the best available resolution for measuring the beam pattern. The approach is calibrated for and tested on two automotive LiDAR systems, the V aleo Scala Gen. 2 and the Blickfeld Cube 1. Both systems differ notably in their properties and available interfaces, but the relevant model parameters could be extracted successfully.
- Europe > Germany > Baden-Württemberg > Karlsruhe Region > Karlsruhe (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Transportation > Ground > Road (0.68)
- Automobiles & Trucks (0.68)
- Information Technology > Robotics & Automation (0.46)
- Aerospace & Defense > Aircraft (0.46)
Hey AI, Generate Me a Hardware Code! Agentic AI-based Hardware Design & Verification
Gadde, Deepak Narayan, Radhakrishna, Keerthan Kopparam, Viswambharan, Vaisakh Naduvodi, Kumar, Aman, Lettnin, Djones, Kunz, Wolfgang, Simon, Sebastian
Modern Integrated Circuits (ICs) are becoming increasingly complex, and so is their development process. Hardware design verification entails a methodical and disciplined approach to the planning, development, execution, and sign-off of functionally correct hardware designs. This tedious process requires significant effort and time to ensure a bug-free tape-out. The field of Natural Language Processing has undergone a significant transformation with the advent of Large Language Models (LLMs). These powerful models, often referred to as Generative AI (GenAI), have revolutionized how machines understand and generate human language, enabling unprecedented advancements in a wide array of applications, including hardware design verification. This paper presents an agentic AI-based approach to hardware design verification, which empowers AI agents, in collaboration with Humain-in-the-Loop (HITL) intervention, to engage in a more dynamic, iterative, and self-reflective process, ultimately performing end-to-end hardware design and verification. This methodology is evaluated on five open-source designs, achieving over 95% coverage with reduced verification time while demonstrating superior performance, adaptability, and configurability.
- Asia > India (0.05)
- South America > Brazil > Amazonas > Manaus (0.04)
- Europe > Germany > Rhineland-Palatinate > Landau (0.04)
- Europe > Germany > Rhineland-Palatinate > Kaiserslautern (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.34)
CAMP-HiVe: Cyclic Pair Merging based Efficient DNN Pruning with Hessian-Vector Approximation for Resource-Constrained Systems
Uddin, Mohammad Helal, Ghanta, Sai Krishna, Seymour, Liam, Baidya, Sabur
Deep learning algorithms are becoming an essential component of many artificial intelligence (AI) driven applications, many of which run on resource-constrained and energy-constrained systems. For efficient deployment of these algorithms, although different techniques for the compression of neural network models are proposed, neural pruning is one of the fastest and effective methods, which can provide a high compression gain with minimal cost. To harness enhanced performance gain with respect to model complexity, we propose a novel neural network pruning approach utilizing Hessian-vector products that approximate crucial curvature information in the loss function, which significantly reduces the computation demands. By employing a power iteration method, our algorithm effectively identifies and preserves the essential information, ensuring a balanced trade-off between model accuracy and computational efficiency. Herein, we introduce CAMP-HiVe, a cyclic pair merging-based pruning with Hessian Vector approximation by iteratively consolidating weight pairs, combining significant and less significant weights, thus effectively streamlining the model while preserving its performance. This dynamic, adaptive framework allows for real-time adjustment of weight significance, ensuring that only the most critical parameters are retained. Our experimental results demonstrate that our proposed method achieves significant reductions in computational requirements while maintaining high performance across different neural network architectures, e.g., ResNet18, ResNet56, and MobileNetv2, on standard benchmark datasets, e.g., CIFAR10, CIFAR-100, and ImageNet, and it outperforms the existing state-of-the-art neural pruning methods.
- North America > United States > Georgia > Clarke County > Athens (0.14)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > Kentucky > Jefferson County > Louisville (0.04)
- (2 more...)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (0.92)
- Health & Medicine > Therapeutic Area > Immunology > HIV (0.43)
Data for Inclusion: The Redistributive Power of Data Economics
While credit is often portrayed as the fuel of development, access to credi t is unevenly distributed -- not merely as a function of income or collateral, but increasingly as a function of data visibility. In this context, the core hypothesis of this paper is that data, when governed ethically and reused efficiently, operates as a re distributive economic asset. The idea that being poor is more expensive is not new; it has been conceptualized as the "poverty premium" -- where low - income individuals pay higher effective prices for credit, insurance, and other services (Carrière - Swallow & Haksar, 2019). Y et what has ch anged is the infrastructure of decision - making: creditworthiness is increasingly determined by algorithmic systems whose inputs are not equitably distributed. Individuals with limited credit histories or fragmented digital footprints remain invisible, not due to financial incapacity, but due to informational exclusion. This asymmetry is not merely a market failure -- it is a structural inequality encoded in data regimes. W e argue that positive credit data -- payment histories, utilization patterns, and account stability -- constitutes a nonrival input that, once generated, can be reused across institutions at near - zero marginal cost without diminishing its value (Jones & Tonetti, 2020; Acemoglu et al., 2023). However, the ability to extract value from such data remains highly uneven. In traditional credit markets, the absence of negative signals penalizes borrowers more than the presence of positive behavior benefits them.
- South America > Uruguay (0.06)
- North America > United States > Tennessee > Davidson County > Nashville (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (3 more...)
- Information Technology > Artificial Intelligence (1.00)
- Information Technology > Data Science > Data Mining (0.46)
Decentralized Collective World Model for Emergent Communication and Coordination
Nomura, Kentaro, Aoki, Tatsuya, Taniguchi, Tadahiro, Horii, Takato
We propose a fully decentralized multi-agent world model that enables both symbol emergence for communication and coordinated behavior through temporal extension of collective predictive coding. Unlike previous research that focuses on either communication or coordination separately, our approach achieves both simultaneously. Our method integrates world models with communication channels, enabling agents to predict environmental dynamics, estimate states from partial observations, and share critical information through bidirectional message exchange with contrastive learning for message alignment. Using a two-agent trajectory drawing task, we demonstrate that our communication-based approach outperforms non-communicative models when agents have divergent perceptual capabilities, achieving the second-best coordination after centralized models. Importantly, our decentralized approach with constraints preventing direct access to other agents' internal states facilitates the emergence of more meaningful symbol systems that accurately reflect environmental states. These findings demonstrate the effectiveness of decentralized communication for supporting coordination while developing shared representations of the environment.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Asia > Japan > Honshū > Kansai > Osaka Prefecture > Osaka (0.04)
- (2 more...)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Asia > Middle East > Jordan (0.04)
- (5 more...)
- Education (0.46)
- Leisure & Entertainment > Games (0.30)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (0.34)
Meaningless Tokens, Meaningful Gains: How Activation Shifts Enhance LLM Reasoning
Shi, Zeru, Wan, Yingjia, Wang, Zhenting, Wang, Qifan, Yang, Fan, Kreiss, Elisa, Tang, Ruixiang
Motivated by the puzzling observation that inserting long sequences of meaningless tokens before the query prompt can consistently enhance LLM reasoning performance, this work analyzes the underlying mechanism driving this phenomenon and based on these insights proposes a more principled method that allows for similar performance gains. First, we find that the improvements arise from a redistribution of activations in the LLM's MLP layers, where near zero activations become less frequent while large magnitude activations increase. This redistribution enhances the model's representational capacity by suppressing weak signals and promoting stronger, more informative ones. Building on this insight, we propose the Activation Redistribution Module (ARM), a lightweight inference-time technique that modifies activations directly without altering the input sequence. ARM adaptively identifies near-zero activations after the non-linear function and shifts them outward, implicitly reproducing the beneficial effects of meaningless tokens in a controlled manner. Extensive experiments across diverse benchmarks and model architectures clearly show that ARM consistently improves LLM performance on reasoning tasks while requiring only a few lines of simple code to implement. Our findings deliver both a clear mechanistic explanation for the unexpected benefits of meaningless tokens and a simple yet effective technique that harnesses activation redistribution to further improve LLM performance.