To compare the performance of the backpropagation and the combined algorithm on large and highly nonlinear problems, both algorithms were used to train a network for the 1-Norm of a 7-dimensional binary vector to an error of 0.001. 4 runs of the sequential versions of both algorithms have been performed.
All 4 programs used online learning at a learn rate of 1. The impulse constants 0 and 0.8 were used on two programs each. The simulations were started by the following commands.
cntback -N7 -O1 -e0.001 -f1 -g250000 -o1
cntback -N7 -O1 -e0.001 -f1 -g250000 -o1 -a0.8
Both simulations with failed, while
the other two programs succeeded after 55097 and
45974 iterations (11661 s and 9800 s)
The combined algorithm used a population size of 50 with 5 backpropagation steps per generation and was started by the following command
cntseq -N7 -O1 -e0.001 -f1 -g1000 -p50 -b5
All 4 programs succeeded after an average of 240 generations
and an average computation time of 18797.5 s.
The following weight matrices and
were found by one of the programs:
Since the encoding range of the weights is by default defined
as , the occurrences of
and
indicate range
overflows caused by the backpropagation steps.
Considering that the parallel version at this problem size
() would run with an efficiency of effectively 100%,
the combined algorithm would definitely be the better choice
for this problem.