• DocumentCode
    715499
  • Title

    Deep learning and the information bottleneck principle

  • Author

    Tishby, Naftali ; Zaslavsky, Noga

  • Author_Institution
    Edmond & Lilly Safra Center for Brain Sci., Hebrew Univ. of Jerusalem, Jerusalem, Israel
  • fYear
    2015
  • fDate
    April 26 2015-May 1 2015
  • Firstpage
    1
  • Lastpage
    5
  • Abstract
    Deep Neural Networks (DNNs) are analyzed via the theoretical framework of the information bottleneck (IB) principle. We first show that any DNN can be quantified by the mutual information between the layers and the input and output variables. Using this representation we can calculate the optimal information theoretic limits of the DNN and obtain finite sample generalization bounds. The advantage of getting closer to the theoretical limit is quantifiable both by the generalization bound and by the network´s simplicity. We argue that both the optimal architecture, number of layers and features/connections at each layer, are related to the bifurcation points of the information bottleneck tradeoff, namely, relevant compression of the input layer with respect to the output layer. The hierarchical representations at the layered network naturally correspond to the structural phase transitions along the information curve. We believe that this new insight can lead to new optimality bounds and deep learning algorithms.
  • Keywords
    bifurcation; data compression; learning (artificial intelligence); neural nets; DNN; IB principle; bifurcation points; deep learning algorithms; deep neural networks; finite sample generalization bounds; hierarchical layered network representations; information bottleneck principle; information curve; input layer compression; layer connections; layer features; mutual information; optimal architecture; optimal information theoretic limits; optimality bounds; output layer; output variables; structural phase transitions; Bifurcation; Complexity theory; Computer architecture; Distortion; Feature extraction; Mutual information; Training;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Theory Workshop (ITW), 2015 IEEE
  • Conference_Location
    Jerusalem
  • Print_ISBN
    978-1-4799-5524-4
  • Type

    conf

  • DOI
    10.1109/ITW.2015.7133169
  • Filename
    7133169