Here we use non-parametric tests to validate the hypotheses H4 (that the crossover operator has an influence on both the GA-based ANN training and the testing performances; Table X) and H5 (that the generation of effective training and validation sets has an influence on both RT-based ANN training and testing performances). As for the first experiment, we used the real dataset (the original telecom data) without preprocessing the data (first preprocessing approach).
We found a very weak support: two pair-vectors differ significantly at a level of significance of 0.1, namely GAO versus GAM and GAM versus GAU, both in the case of training phase. Also, we found no evidence to differentiate between the three RT mechanisms.
In the first four experiments, when we validate our hypotheses, we relied exclusively on nonparametric tests. We argued that the parametric tests (like t-test, univariate ANOVA, etc.) require the vectors analysed to satisfy different assumptions. For instance, when applying ANOVA one should check the following assumptions: that the observations are independent, that the sample data have a normal distribution, and that scores in different groups have homogeneous variances. The first assumption is satisfied, since all other factors besides preprocessing, distribution and training mechanism that could influence the classifiers’ performances are fixed. For the second assumption we argue that ANOVA is robust against normality assumptions if the sample size is large. Regarding the third assumption, SPSS (the software that we used) incorporates the case when the variances between groups are assumed to be non-equal.
In order to give more strength to our results from the previous experiments and, at the same time, to validate our main hypothesis, we finally performed a three-way ANOVA analysis having as grouping variables the technique used (GAO, GAM, GAA, GAU, RT1, RT2, and RT3), the preprocessing method (PR1, i.e. no preprocessing; PR2, i.e. normalization; PR3, i.e. dividing the variables by the maximum absolute values), and the data distribution (REAL, UNIF, NORM, LOG, and LAP). With the third preprocessing method we obtained values between −1 and +1. We used the vectors’ means to fill in our accuracy rates data. Tables XI and XII include the data we used to perform the three-way ANOVA.
Next, the results of three-way ANOVA for both training and test accuracy rates are shown in Tables XIII and XIV.
As the tables show, all the factors are statistically significant. In other words, they have an individual and combined influence on both training and testing performances. The last column (partial η2) reports the ‘practical’ significance of each term, based upon the ratio of the variation (sum of squares) accounted for by the term to the sum of the variation accounted for by the term and the variation left to error. Larger values of partial η2 indicate a greater amount of variation accounted for by the model term, to a maximum of unity. Here, the individual factors and their combinations, while statistically significant, have great effect on classifier accuracy. Consequently, the main hypothesis, i.e. H6, is validated.
In the next three tables we present the pairs’ comparisons for the training performances. The second hypothesis (H2) is validated (Table XV) and ‘normalization’ is the best preprocessing approach, followed in order by ‘maximum absolute values’ and ‘no preprocessing’. Concerning the third hypothesis (H3), the best performance was obtained when data were normally distributed (Table XVI). The next best distribution was that of the real data, followed by uniform, logistic and
Table XI. Accuracy rates for training
PREPROC |
DISTRIB |
Technique |
||||||
GAO |
GA GAM |
GAA |
GAU |
RT |
||||
RT1 |
RT2 |
RT3 |
||||||
Unpreprocessed |
REAL |
93.02 |
92.86 |
92.92 |
93.30 |
92.22 |
92.43 |
92.41 |
UNIF |
95.84 |
95.84 |
95.78 |
95.76 |
94.92 |
95.04 |
94.84 |
|
NORM |
96.39 |
96.60 |
96.49 |
96.43 |
95.48 |
95.45 |
95.52 |
|
LOG |
94.82 |
94.99 |
94.80 |
94.82 |
92.70 |
92.70 |
92.53 |
|
LAP |
90.09 |
90.00 |
90.16 |
90.19 |
88.12 |
88.06 |
88.06 |
|
Normalization |
REAL |
99.43 |
99.49 |
99.46 |
99.33 |
99.11 |
99.10 |
99.08 |
UNIF |
99.79 |
99.81 |
99.79 |
99.79 |
99.80 |
99.80 |
99.80 |
|
NORM |
97.90 |
97.90 |
97.90 |
97.90 |
98.07 |
98.03 |
97.97 |
|
LOG |
99.11 |
99.06 |
98.98 |
98.98 |
98.95 |
98.98 |
98.96 |
|
LAP |
98.08 |
98.13 |
98.01 |
98.10 |
98.02 |
98.01 |
98.06 |
|
Max. of |
REAL |
99.68 |
99.68 |
99.68 |
99.68 |
99.69 |
99.69 |
99.69 |
absolute values |
UNIF |
97.79 |
97.77 |
97.73 |
97.84 |
97.77 |
97.77 |
97.89 |
NORM |
96.91 |
97.00 |
97.02 |
97.02 |
96.93 |
96.90 |
96.91 |
|
LOG |
96.50 |
96.52 |
96.52 |
96.52 |
96.68 |
96.59 |
96.60 |
|
LAP |
95.64 |
95.83 |
95.76 |
95.81 |
95.26 |
95.47 |
95.23 |
|
Table XII. Accuracy rates for testing |
||||||||
PREPROC |
DISTRIB |
Technique |
||||||
GAO |
GA GAM |
GAA |
GAU |
RT |
||||
RT1 |
RT2 |
RT3 |
||||||
Unpreprocessed |
REAL |
85.24 |
85.48 |
85.48 |
85.95 |
83.49 |
83.97 |
93.81 |
UNIF |
88.57 |
87.86 |
88.93 |
88.75 |
89.11 |
88.57 |
89.05 |
|
NORM |
89.46 |
89.64 |
90.00 |
89.82 |
89.46 |
89.52 |
89.52 |
|
LOG |
81.19 |
81.43 |
81.43 |
81.19 |
79.92 |
79.92 |
79.29 |
|
LAP |
80.41 |
81.02 |
81.22 |
81.02 |
78.10 |
77.76 |
77.89 |
|
Normalization |
REAL |
85.71 |
86.19 |
85.71 |
85.71 |
85.79 |
85.63 |
85.79 |
UNIF |
92.86 |
93.04 |
92.86 |
92.86 |
92.86 |
92.86 |
92.92 |
|
NORM |
96.43 |
96.43 |
96.43 |
96.43 |
96.49 |
96.19 |
96.43 |
|
LOG |
88.10 |
88.10 |
88.10 |
88.10 |
88.10 |
88.25 |
88.25 |
|
LAP |
92.25 |
92.45 |
91.84 |
92.25 |
91.36 |
91.50 |
91.56 |
|
Max. of |
REAL |
97.62 |
97.62 |
97.62 |
97.62 |
97.54 |
97.70 |
97.62 |
absolute values |
UNIF |
95.00 |
95.36 |
95.71 |
95.36 |
96.31 |
96.43 |
96.07 |
NORM |
93.21 |
93.57 |
93.93 |
93.57 |
92.86 |
93.15 |
93.27 |
|
LOG |
88.10 |
88.10 |
87.86 |
88.10 |
88.25 |
88.10 |
88.25 |
|
LAP |
88.37 |
89.18 |
88.98 |
89.59 |
89.86 |
89.93 |
89.86 |
Уважаемый посетитель!
Чтобы распечатать файл, скачайте его (в формате Word).
Ссылка на скачивание - внизу страницы.