TY - UNPB
T1 - Optimization in an Error Backpropagation Neural
Network Environment with a Performance Test
on a Pattern Classification Problem
AU - Fischer, Manfred M.
AU - Staufer-Steinnocher, Petra
PY - 1998/3/1
Y1 - 1998/3/1
N2 - Various techniques of optimizing the multiple class cross-entropy error function
to train single hidden layer neural network classifiers with softmax output transfer
functions are investigated on a real-world multispectral pixel-by-pixel classification
problem that is of fundamental importance in remote sensing. These techniques
include epoch-based and batch versions of backpropagation of gradient descent,
PR-conjugate gradient and BFGS quasi-Newton errors. The method of choice
depends upon the nature of the learning task and whether one wants to optimize
learning for speed or generalization performance. It was found that, comparatively
considered, gradient descent error backpropagation provided the best and most stable
out-of-sample performance results across batch and epoch-based modes of operation.
If the goal is to maximize learning speed and a sacrifice in generalisation is acceptable,
then PR-conjugate gradient error backpropagation tends to be superior. If the
training set is very large, stochastic epoch-based versions of local optimizers should
be chosen utilizing a larger rather than a smaller epoch size to avoid inacceptable
instabilities in the generalization results. (authors' abstract)
AB - Various techniques of optimizing the multiple class cross-entropy error function
to train single hidden layer neural network classifiers with softmax output transfer
functions are investigated on a real-world multispectral pixel-by-pixel classification
problem that is of fundamental importance in remote sensing. These techniques
include epoch-based and batch versions of backpropagation of gradient descent,
PR-conjugate gradient and BFGS quasi-Newton errors. The method of choice
depends upon the nature of the learning task and whether one wants to optimize
learning for speed or generalization performance. It was found that, comparatively
considered, gradient descent error backpropagation provided the best and most stable
out-of-sample performance results across batch and epoch-based modes of operation.
If the goal is to maximize learning speed and a sacrifice in generalisation is acceptable,
then PR-conjugate gradient error backpropagation tends to be superior. If the
training set is very large, stochastic epoch-based versions of local optimizers should
be chosen utilizing a larger rather than a smaller epoch size to avoid inacceptable
instabilities in the generalization results. (authors' abstract)
U2 - 10.57938/5b7541c1-fefe-4eb1-af1a-34b4aae86d28
DO - 10.57938/5b7541c1-fefe-4eb1-af1a-34b4aae86d28
M3 - WU Working Paper
T3 - Discussion Papers of the Institute for Economic Geography and GIScience
BT - Optimization in an Error Backpropagation Neural
Network Environment with a Performance Test
on a Pattern Classification Problem
PB - WU Vienna University of Economics and Business
CY - Vienna
ER -