15th European Conference on Artificial Intelligence
  July 21-26 2002     Lyon     France  
   

ECAI-2002 Conference Paper

[PDF] [full paper] [prev] [tofc] [next]

Effective Stacking of Distributed Classifiers

Grigorios Tsoumakas, Ioannis Vlahavas

One of the most promising lines of research towards discovering global predictive models from physically distributed data sets is local learning and model integration. Local learning avoids moving raw data around the distributed nodes and minimizes communication, coordination and synchronization cost. However, the integration of local models is not a straightforward process. Majority Voting is a simple solution that works well in some domains, but it does not always offer the best predictive performance. Stacking on the other hand, offers flexibility in modeling, but brings along the problem of how to train on sufficient and at the same time independent data without the cost of moving raw data around the distributed nodes. In addition, the scalability of Stacking with respect to the number of distributed nodes is another important issue that has not yet been substantially investigated. This paper presents a framework for constructing a global predictive model from local classifiers that does not require moving raw data around, achieves high predictive accuracy and scales up efficiently with respect to large numbers of distributed data sets.

Keywords: Data Mining and Knowledge Discovery, Machine Learning, Distributed AI

Citation: Grigorios Tsoumakas, Ioannis Vlahavas: Effective Stacking of Distributed Classifiers. In F. van Harmelen (ed.): ECAI2002, Proceedings of the 15th European Conference on Artificial Intelligence, IOS Press, Amsterdam, 2002, pp.340-344.


[prev] [tofc] [next]


ECAI-2002 is organised by the European Coordinating Committee for Artificial Intelligence (ECCAI) and hosted by the Université Claude Bernard and INSA, Lyon, on behalf of Association Française pour l'Intelligence Artificielle.