ECAI-2000 Logo

ECAI-2000 Conference Paper

[PDF] [full paper] [prev] [tofc] [next]

Learning Efficiently with Neural Networks: A Theoretical Comparison between Structured and Flat Representations

Marco Gori, Paolo Frasconi, Alessandro Sperduti

We are interested in the relationship between learning efficiency and representation in the case of supervised neural networks for pattern classification trained by continuous error minimization techniques, such as gradient descent. In particular, we focus our attention on a recently introduced architecture called recursive neural network (RNN) which is able to learn class membership of patterns represented as labeled directed ordered acyclic graphs (DOAG). RNNs offer several benefits compared to feedforward and recurrent networks for sequences. However, how RNNs compare to these models in terms of learning efficiency still needs investigation. In this paper we give a theoretical answer by giving a set of results concerning the shape of the error surface and critically discussing the implications of these results on the relative difficulty of learning with different data representations. The message of this paper is that, whenever structured representations are available, they should be preferred to ``flat'' (array based) representations because they are likely to simplify learning in terms of time complexity.

Keywords: Machine Learning, Neural Networks

Citation: Marco Gori, Paolo Frasconi, Alessandro Sperduti: Learning Efficiently with Neural Networks: A Theoretical Comparison between Structured and Flat Representations. In W.Horn (ed.): ECAI2000, Proceedings of the 14th European Conference on Artificial Intelligence, IOS Press, Amsterdam, 2000, pp.301-305.


[prev] [tofc] [next]


ECAI-2000 is organised by the European Coordinating Committee for Artificial Intelligence (ECCAI) and hosted by the Humboldt University on behalf of Gesellschaft für Informatik.