Neural Networks Use Distance Metrics
We present empirical evidence that neural networks with ReLU and Absolute Value activations learn distance-based representations. We independently manipulate both distance and intensity properties of internal activations in trained models, finding that both architectures are highly sensitive to small distance-based perturbations while maintaining robust performance under large intensity-based perturbations. These findings challenge the prevailing intensity-based interpretation of neural network activations and offer new insights into their learning and decision-making processes.
Nov-26-2024
- Country:
- Asia > India (0.04)
- Europe > United Kingdom (0.04)
- North America > Canada
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (0.94)
- Research Report
- Technology: