An Information Theoretic Interpretation to Deep Neural Networks

With the unprecedented performance achieved by deep learning, it is commonly believed that deep neural networks (DNNs) attempt to extract informative features for learning tasks. To formalize this intuition, we apply the local information geometric analysis and establish an information-theoretic fra...

Full description

Bibliographic Details
Main Authors: Xu, Xiangxiang (Author), Huang, Shao-Lun (Author), Zheng, Lizhong (Author), Wornell, Gregory W. (Author)
Format: Article
Language:English
Published: Multidisciplinary Digital Publishing Institute, 2022-01-20T19:32:08Z.
Subjects:
Online Access:Get fulltext
LEADER 01410 am a22001573u 4500
001 139647
042 |a dc 
100 1 0 |a Xu, Xiangxiang  |e author 
700 1 0 |a Huang, Shao-Lun  |e author 
700 1 0 |a Zheng, Lizhong  |e author 
700 1 0 |a Wornell, Gregory W.  |e author 
245 0 0 |a An Information Theoretic Interpretation to Deep Neural Networks 
260 |b Multidisciplinary Digital Publishing Institute,   |c 2022-01-20T19:32:08Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/139647 
520 |a With the unprecedented performance achieved by deep learning, it is commonly believed that deep neural networks (DNNs) attempt to extract informative features for learning tasks. To formalize this intuition, we apply the local information geometric analysis and establish an information-theoretic framework for feature selection, which demonstrates the information-theoretic optimality of DNN features. Moreover, we conduct a quantitative analysis to characterize the impact of network structure on the feature extraction process of DNNs. Our investigation naturally leads to a performance metric for evaluating the effectiveness of extracted features, called the H-score, which illustrates the connection between the practical training process of DNNs and the information-theoretic framework. Finally, we validate our theoretical results by experimental designs on synthesized data and the ImageNet dataset. 
655 7 |a Article