xnor
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- North America > United States > Tennessee > Knox County > Knoxville (0.04)
- Europe > Italy > Tuscany > Florence (0.04)
- Asia > Middle East > Jordan (0.04)
- Research Report > New Finding (0.92)
- Research Report > Experimental Study (0.92)
Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators
The choice of activation functions and their motivation is a long-standing issue within the neural network community. Neuronal representations within artificial neural networks are commonly understood as logits, representing the log-odds score of presence of features within the stimulus. We derive logit-space operators equivalent to probabilistic Boolean logic-gates AND, OR, and XNOR for independent probabilities. Such theories are important to formalize more complex dendritic operations in real neurons, and these operations can be used as activation functions within a neural network, introducing probabilistic Boolean-logic as the core operation of the neural network. Since these functions involve taking multiple exponents and logarithms, they are computationally expensive and not well suited to be directly used within neural networks. Consequently, we construct efficient approximations named $\text{AND}_\text{AIL}$ (the AND operator Approximate for Independent Logits), $\text{OR}_\text{AIL}$, and $\text{XNOR}_\text{AIL}$, which utilize only comparison and addition operations, have well-behaved gradients, and can be deployed as activation functions in neural networks. Like MaxOut, $\text{AND}_\text{AIL}$ and $\text{OR}_\text{AIL}$ are generalizations of ReLU to two-dimensions. While our primary aim is to formalize dendritic computations within a logit-space probabilistic-Boolean framework, we deploy these new activation functions, both in isolation and in conjunction to demonstrate their effectiveness on a variety of tasks including tabular classification, image classification, transfer learning, abstract reasoning, and compositional zero-shot learning.
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- North America > United States > Tennessee > Knox County > Knoxville (0.04)
- Europe > Italy > Tuscany > Florence (0.04)
- Asia > Middle East > Jordan (0.04)
- Research Report > New Finding (0.92)
- Research Report > Experimental Study (0.92)
Highly Efficient and Effective LLMs with Multi-Boolean Architectures
Tran, Ba-Hien, Nguyen, Van Minh
Weight binarization has emerged as a promising strategy to drastically reduce the complexity of large language models (LLMs). It is mainly classified into two approaches: post-training binarization and finetuning with training-aware binarization methods. The first approach, while having low complexity, leads to significant loss of information from the original LLMs, resulting in poor performance. The second approach, on the other hand, relies heavily on full-precision latent weights for gradient approximation of binary weights, which not only remains suboptimal but also introduces substantial complexity. In this paper, we introduce a novel framework that effectively transforms LLMs into multi-kernel Boolean parameters, for the first time, finetunes them directly in the Boolean domain, eliminating the need for expensive latent weights. This significantly reduces complexity during both finetuning and inference. Through extensive and insightful experiments across a wide range of LLMs, we demonstrate that our method outperforms recent ultra low-bit quantization and binarization methods.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Russia (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- (8 more...)
Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators
The choice of activation functions and their motivation is a long-standing issue within the neural network community. Neuronal representations within artificial neural networks are commonly understood as logits, representing the log-odds score of presence of features within the stimulus. We derive logit-space operators equivalent to probabilistic Boolean logic-gates AND, OR, and XNOR for independent probabilities. Such theories are important to formalize more complex dendritic operations in real neurons, and these operations can be used as activation functions within a neural network, introducing probabilistic Boolean-logic as the core operation of the neural network. Since these functions involve taking multiple exponents and logarithms, they are computationally expensive and not well suited to be directly used within neural networks.
Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators
The choice of activation functions and their motivation is a long-standing issue within the neural network community. Neuronal representations within artificial neural networks are commonly understood as logits, representing the log-odds score of presence of features within the stimulus. We derive logit-space operators equivalent to probabilistic Boolean logic-gates AND, OR, and XNOR for independent probabilities. Such theories are important to formalize more complex dendritic operations in real neurons, and these operations can be used as activation functions within a neural network, introducing probabilistic Boolean-logic as the core operation of the neural network. Since these functions involve taking multiple exponents and logarithms, they are computationally expensive and not well suited to be directly used within neural networks.
BOLD: Boolean Logic Deep Learning
Nguyen, Van Minh, Ocampo, Cristian, Askri, Aymen, Leconte, Louis, Tran, Ba-Hien
Deep learning is computationally intensive, with significant efforts focused on reducing arithmetic complexity, particularly regarding energy consumption dominated by data movement. While existing literature emphasizes inference, training is considerably more resource-intensive. This paper proposes a novel mathematical principle by introducing the notion of Boolean variation such that neurons made of Boolean weights and inputs can be trained -- for the first time -- efficiently in Boolean domain using Boolean logic instead of gradient descent and real arithmetic. We explore its convergence, conduct extensively experimental benchmarking, and provide consistent complexity evaluation by considering chip architecture, memory hierarchy, dataflow, and arithmetic precision. Our approach achieves baseline full-precision accuracy in ImageNet classification and surpasses state-of-the-art results in semantic segmentation, with notable performance in image super-resolution, and natural language understanding with transformer-based models. Moreover, it significantly reduces energy consumption during both training and inference.
- North America > United States > Tennessee > Knox County > Knoxville (0.14)
- North America > Canada > Ontario > Toronto (0.14)
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- Europe > Italy > Tuscany > Florence (0.04)
Boolean Variation and Boolean Logic BackPropagation
Deep learning has become a commonplace solution to numerous modern problems, occupying a central spot of today's technological and social attention. The recipe of its power is the combining effect of unprecedented large dimensions and learning process based on gradient backpropagation [LeCun et al., 1998]. In particular, thanks to the simplicity of neuron model that is decomposed into a weighted linear sum followed by a non-linear activation function, the gradient of weights is solely determined by their respective input without involving cross-parameter dependency. Hence, in terms of computational process, gradient backpropagation is automated by gradient chain rule and only requires a buffering of the forward input data. However, deep learning is computationally intensive.
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Romania > Sud-Est Development Region > Constanța County > Constanța (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- (3 more...)
Apple On A Buying Spree, Banking On AI Startups
For many years now, we are witnessing a tectonic shift in how technology is being used to revolutionise major sectors around the world. One of the most prominent keywords that keep popping up now and then is the'artificial intelligence revolution.' Tech giants such as Google, Microsoft, Facebook, Apple, etc., have shown deep interest in not just adopting AI but also acquiring AI startups, which has become the new normal. Apple is winning the race with the'You Start, I Purchase' mindset. The number of acquisitions it has made, without much pomp and shows over a short span of three to four years, is remarkable. Nevertheless, it is a win-win situation for both Apple as well as the startups. "Apple buys a new company every two or three weeks."
Apple acquires Inductiv, a machine learning startup
Apple recently acquired Inductiv Inc. Inductiv, a machine learning startup, is co-founded by Professor Ihab Francis Ilyas, University of Waterloo. It has other professor co-founders from the University of Wisconsin and Stanford University. Inductiv uses artificial intelligence to correct errors in data. Apple has so far acquired four companies in 2020– DarkSky, Voysis, Xnor.ai and NextVR. Inductiv is the 5th acquisition in 2020.