Knowledge Isomorphism between Neural Networks
Liang, Ruofan, Li, Tianlin, Li, Longfei, Zhang, Quanshi
This paper aims to analyze knowledge isomorphism between pre-trained deep neural networks. We propose a generic definition for knowledge isomorphism between neural networks at different fuzziness levels, and design a task-agnostic and model-agnostic method to disentangle and quantify isomorphic features from intermediate layers of a neural network. As a generic tool, our method can be broadly used for different applications. In preliminary experiments, we have used knowledge isomorphism as a tool to diagnose feature representations of neural networks. Knowledge isomorphism provides new insights to explain the success of existing deep-learning techniques, such as knowledge distillation and network compression. More crucially, it has been shown that knowledge isomorphism can also be used to refine pre-trained networks and boost performance.
Aug-5-2019
- Country:
- Asia > China
- North America > United States
- California (0.04)
- Genre:
- Research Report > New Finding (0.46)
- Technology: