Abstraction hierarchy in deep learning neural networks
Abstract
We develop a methodology to assess knowledge representation in deep neural networks trained to recognize classes of objects. We measure the abstraction level by studying correlations between the neuron activation levels of different layers based on image class. The approach is developed and tested using CIFAR-10 dataset and MatConvNet toolbox. The results show that different kinds of layers, convolutional or pooling, have different effect on the representation. The observations also point to a tendency for incremental increase in the abstraction measure, sometimes interrupted by more significant jumps, which may indicate a qualitative transition between abstraction levels. We describe and interpret current results and outline the direction of future work.
Publication Title
Proceedings of the International Joint Conference on Neural Networks
Recommended Citation
Ilin, R., Watson, T., & Kozma, R. (2017). Abstraction hierarchy in deep learning neural networks. Proceedings of the International Joint Conference on Neural Networks, 2017-May, 768-774. https://doi.org/10.1109/IJCNN.2017.7965929