These two factors are not independent, so we will perform an analysis
using all the possible pairs (,
) of the Cartesian
product of the two sets. For each pair we will perform 30 runs of the
genetic algorithm with different random seeds. Table
2 shows the average value and standard deviation of
the 30 runs for each experiment.
The study of the results has been made by means of an analysis of
variance ANOVA II [DC74,Mil81,SC80],
with the fitness of the best individuals, , as test variable. This
fitness is obtained independently in 30 runs and depending on two
fixed factors and their interaction. The fixed factors are: the
confidence coefficient
with four levels and the number of best
individuals
with five levels. The linear model has the form:
![]() |
(14) | ||
![]() ![]() |
where:
The hypothesis tests try to determine the effect of each term over the
fitness of the best individuals, . We have carried out tests for
every factor and for the interaction among the factors. This and
subsequent tests are performed with a confidence level of
. The
coefficient
of the linear model tells us the percentage of
variance of
that is explained by the model.
For determining whether there are significant differences among the various levels of a factor, we perform a multiple comparison test of the average fitness obtained with the different levels of each factor. First, we carry out a Levene test [Mil96,Lev60] for evaluating the equality of variances. If the hypothesis that the variances are equal is accepted, we perform a Bonferroni test [Mil96] for ranking the means of each level of the factor. Our aim is to find the level of each factor whose average fitness is significantly better than the average fitness of the rest of the levels of the factor. If the test of Levene results in rejecting the equality of covariance matrixes, we perform a Tamhane test [TD00] instead of a Bonferroni test. Tables 9, 12, and 13 in Appendix A show the results obtained following the above methodology.
For Sphere function, the significant levels of each term of
the linear model on Table 9 show that none of the
factors of the linear model has a significant effect on the model
built to explain the variance of the fitness
. This effect is due
to the fact that
is easy to optimize and the fitness behaves
as a singular random variable with sample variance near 0. We can see
in Table 2 that the best results are obtained with
the pair
. The multiple comparison test of Table
12 confirms that the means obtained with the value
are significatively better than the means obtained with other
values. In the same way, the average fitness for
is
significantly the best one. The results show that, for any value of
, the best value of
, in general, is
. Due
to the simple form of
, the best parameters of the crossover
show a high exploitative component with a fast shifting towards the
region of the best individuals.
For the unimodal and non-separable functions and
, both factors and their interaction are significant in the
linear model that explains the sample variance of
with a
determination coefficient around
. Table 2
shows that the best results are obtained with
; the Tamhane test
shows that the means obtained with this value of
are
significatively better than the means obtained with other values. The
results for the value of the confidence coefficient are less
conclusive. In fact, for
there are no significant
differences among the different values of
, although the
best results are obtained with
. For
the
average fitness for
is the best one, but without
significant differences with
.
together with
is the one that shows the best results. We can conclude that the
feature of non-separability of the functions does not imply a notable
change in the parameters of the crossover with respect to the
parameters used for
.
For and
, which are separable and multimodal, the
most adequate pair of parameters is
. For
, the
test shows that the performance of this pair is significantly better.
However, for
, the best mean is obtained with
with
results that are significantly better than these obtained with other
values, with the exception of
. There are no significant
differences among
,
and
. The three
factors of the linear model are significant with quite large
determination coefficients of
for
and
for
. This means that the factors and their interaction
explain a high percentage of the variance of the fitness
.
For , the best results are obtained with the pair
. The Tamhane test confirms that
is the most suitable value,
while there are no significant differences among
,
and
. For
the best results are
obtained with the pair
. The test shows that large values
of
are the most suitable for the optimization of this function.
There are no significant differences among the performance of the
different values of
. For both functions the determination
coefficient of the linear model is low, showing that the linear model
does not explain the variance of the fitness. The lack of a linear
relation among
,
and the fitness makes it more difficult
to determine the best value of the parameters of the crossover.
The case of and
is similar, as the linear model
hardly gives any information about the effect of the parameters on the
fitness. The most adequate pair for the optimization of these two
functions is
. The test shows that the best values of
are
and
. On the other hand, there are no significant
differences among the performance of the crossover for the different
values of
.
The overall results show that the selection of the best
individuals of the population would suffice for obtaining a
localization estimator good enough to guide the search process even
for multimodal functions where a small value of
could favor the
convergence to local optima. However, if the virtual parents have
a worse fitness than the parent from the population, the offspring is
generated near the latter, and the domain can be explored in multiple
directions. In this way, the premature convergence to suboptimal
virtual parents is avoided.
However, if the best individuals are concentrated in a local
optimum the algorithm will very likely converge to such optimum. That
is the reason why in complex functions a larger value of
may be
reasonable, adding to the confidence interval individuals located in
or near different optima. As an example of this, the case of
for which the best results are achieved with
and
is noteworthy.
The confidence coefficient bounds the error in the determination of
the localization parameter and is responsible for focussing the
search. The multiple comparison tests show that the value
is the best for 6 problems, and is, as least, no worse
than the best one in the other problems. So it can be chosen as the
most adequate value of the parameter.
Domingo 2005-07-11