• DocumentCode
    476986
  • Title

    Dimensionally distributed learning models and algorithm

  • Author

    Zheng, Haipeng ; Kulkarni, Sanjeev R. ; Poor, H. Vincent

  • Author_Institution
    Dept. of Electr. Eng., Princeton Univ., Princeton, NJ
  • fYear
    2008
  • fDate
    June 30 2008-July 3 2008
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    This paper introduces a framework for regression with dimensionally distributed data with a fusion center. A cooperative learning algorithm, the iterative conditional expectation algorithm (ICEA), is designed within this framework. The algorithm can effectively discover linear combinations of individual estimators trained by each agent without transferring and storing large amount of data amongst the agents and the fusion center. The convergence of ICEA is explored. Specifically, for a two agent system, each complete round of ICEA is guaranteed to be a non-expansive map on the function space of each agent. The advantages and limitations of ICEA are also discussed for data sets with various distributions and various hidden rules. Moreover, several techniques are also designed to leverage the algorithm to effectively learn more complex hidden rules that are not linearly decomposable.
  • Keywords
    distributed algorithms; iterative methods; learning (artificial intelligence); multi-agent systems; regression analysis; ICEA convergence; agent system; cooperative learning; dimensionally distributed data; dimensionally distributed learning model; fusion center; iterative conditional expectation algorithm; nonexpansive map; regression analysis; Distributed learning; estimation; heterogeneous data; regression;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Fusion, 2008 11th International Conference on
  • Conference_Location
    Cologne
  • Print_ISBN
    978-3-8007-3092-6
  • Electronic_ISBN
    978-3-00-024883-2
  • Type

    conf

  • Filename
    4632362