[full paper] |
Marco Gori, Paolo Frasconi, Alessandro Sperduti
We are interested in the relationship between learning efficiency and representation in the case of supervised neural networks for pattern classification trained by continuous error minimization techniques, such as gradient descent. In particular, we focus our attention on a recently introduced architecture called recursive neural network (RNN) which is able to learn class membership of patterns represented as labeled directed ordered acyclic graphs (DOAG). RNNs offer several benefits compared to feedforward and recurrent networks for sequences. However, how RNNs compare to these models in terms of learning efficiency still needs investigation. In this paper we give a theoretical answer by giving a set of results concerning the shape of the error surface and critically discussing the implications of these results on the relative difficulty of learning with different data representations. The message of this paper is that, whenever structured representations are available, they should be preferred to ``flat'' (array based) representations because they are likely to simplify learning in terms of time complexity.
Keywords: Machine Learning, Neural Networks
Citation: Marco Gori, Paolo Frasconi, Alessandro Sperduti: Learning Efficiently with Neural Networks: A Theoretical Comparison between Structured and Flat Representations. In W.Horn (ed.): ECAI2000, Proceedings of the 14th European Conference on Artificial Intelligence, IOS Press, Amsterdam, 2000, pp.301-305.