Power Laws in Deep Learning 2: Universality

#artificialintelligence 

Editor's note: You can read the previous post in this series, Power Laws in Deep Learning, here. In a previous post, we saw that the Fully Connected (FC) layers of the most common pre-trained Deep Learning display power law behavior. Remarkably, the FC matrices all lie within the Universality Class of Fat Tailed Random Matrices! We define a random matrix by defining a matrix of size, and drawing the matrix elements from a random distribution. In either case, Random Matrix Theory tells us what the asymptotic form of ESD should look like. But first, let's see what model works best.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found