Light Differentiable Logic Gate Networks
Rüttgers, Lukas, Aczel, Till, Plesner, Andreas, Wattenhofer, Roger
–arXiv.org Artificial Intelligence
Differentiable logic gate networks (DLGNs) exhibit extraordinary efficiency at inference while sustaining competitive accuracy. But vanishing gradients, discretization errors, and high training cost impede scaling these networks. Even with dedicated parameter initialization schemes from subsequent works, increasing depth still harms accuracy. We show that the root cause of these issues lies in the underlying parametrization of logic gate neurons themselves. To overcome this issue, we propose a reparametrization that also shrinks the parameter size logarithmically in the number of inputs per gate. For binary inputs, this already reduces the model size by 4x, speeds up the backward pass by up to 1.86x, and converges in 8.5x fewer training steps. On top of that, we show that the accuracy on CIFAR-100 remains stable and sometimes superior to the original parametrization.
arXiv.org Artificial Intelligence
Oct-7-2025
- Country:
- Asia > Japan
- Honshū
- Kantō > Kanagawa Prefecture
- Yokohama (0.04)
- Tōhoku > Fukushima Prefecture
- Fukushima (0.04)
- Kantō > Kanagawa Prefecture
- Honshū
- Europe
- Switzerland > Zürich
- Zürich (0.85)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Switzerland > Zürich
- North America
- Canada
- British Columbia > Metro Vancouver Regional District
- Vancouver (0.04)
- Ontario > Toronto (0.14)
- British Columbia > Metro Vancouver Regional District
- United States
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Maryland > Baltimore (0.04)
- Nevada > Clark County
- Las Vegas (0.04)
- Louisiana > Orleans Parish
- Canada
- Asia > Japan
- Genre:
- Research Report (1.00)
- Technology: