Nuclear Norm Regularization for Deep Learning
–Neural Information Processing Systems
Penalizing the nuclear norm of a function's Jacobian encourages it to locally behave like a low-rank linear map. Such functions vary locally along only a handful of directions, making the Jacobian nuclear norm a natural regularizer for machine learning problems. However, this regularizer is intractable for high-dimensional problems, as it requires computing a large Jacobian matrix and taking its SVD. We show how to efficiently penalize the Jacobian nuclear norm using techniques tailormade for deep learning. We prove that for functions parametrized as compositions f = g h, one may equivalently penalize the average squared Frobenius norms of Jg and Jh. We then propose a denoising-style approximation that avoids Jacobian computations altogether. Our method is simple, efficient, and accurate, enabling Jacobian nuclear norm regularization to scale to high-dimensional deep learning problems. We complement our theory with an empirical study of our regularizer's performance and investigate applications to denoising and representation learning.
Neural Information Processing Systems
Mar-27-2025, 09:13:57 GMT
- Country:
- North America > United States (0.46)
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Education > Focused Education
- Special Education (0.45)
- Health & Medicine (0.45)
- Information Technology (0.67)
- Education > Focused Education
- Technology: