Talk
The approximation capabilities of neural networks
- Kai Fong Ernest Chong (Singapore University of Technology and Design)
Abstract
The universal approximation theorem says that a standard feedforward neural network with one hidden layer is able to uniformly approximate any continuous multivariate function f to any given approximation threshold ε, provided that the activation function is continuous and non-polynomial. In this talk, we shall give a quantitative refinement of the theorem via a direct algebraic approach. In particular, when f is polynomial, we give an explicit finite upper bound (independent of ε) for the number of hidden units required. We shall discuss how ideas from algebraic geometry, algebraic combinatorics and approximation theory are combined in our algebraic proof.