Intelligent systems in accounting, finance and managementassessing predictive performance of ann-based classifiers, страница 18

6.4.  Experiment 4

Here we use non-parametric tests to validate the hypotheses H4 (that the crossover operator has an influence on both the GA-based ANN training and the testing performances; Table X) and H5 (that the generation of effective training and validation sets has an influence on both RT-based ANN training and testing performances). As for the first experiment, we used the real dataset (the original telecom data) without preprocessing the data (first preprocessing approach).

We found a very weak support: two pair-vectors differ significantly at a level of significance of 0.1, namely GAO versus GAM and GAM versus GAU, both in the case of training phase. Also, we found no evidence to differentiate between the three RT mechanisms.

6.5.  Experiment 5

In the first four experiments, when we validate our hypotheses, we relied exclusively on nonparametric tests. We argued that the parametric tests (like t-test, univariate ANOVA, etc.) require the vectors analysed to satisfy different assumptions. For instance, when applying ANOVA one should check the following assumptions: that the observations are independent, that the sample data have a normal distribution, and that scores in different groups have homogeneous variances. The first assumption is satisfied, since all other factors besides preprocessing, distribution and training mechanism that could influence the classifiers’ performances are fixed. For the second assumption we argue that ANOVA is robust against normality assumptions if the sample size is large. Regarding the third assumption, SPSS (the software that we used) incorporates the case when the variances between groups are assumed to be non-equal.

In order to give more strength to our results from the previous experiments and, at the same time, to validate our main hypothesis, we finally performed a three-way ANOVA analysis having as grouping variables the technique used (GAO, GAM, GAA, GAU, RT1, RT2, and RT3), the preprocessing method (PR1, i.e. no preprocessing; PR2, i.e. normalization; PR3, i.e. dividing the variables by the maximum absolute values), and the data distribution (REAL, UNIF, NORM, LOG, and LAP). With the third preprocessing method we obtained values between −1 and +1. We used the vectors’ means to fill in our accuracy rates data. Tables XI and XII include the data we used to perform the three-way ANOVA.

Next, the results of three-way ANOVA for both training and test accuracy rates are shown in Tables XIII and XIV.

As the tables show, all the factors are statistically significant. In other words, they have an individual and combined influence on both training and testing performances. The last column (partial η2) reports the ‘practical’ significance of each term, based upon the ratio of the variation (sum of squares) accounted for by the term to the sum of the variation accounted for by the term and the variation left to error. Larger values of partial η2 indicate a greater amount of variation accounted for by the model term, to a maximum of unity. Here, the individual factors and their combinations, while statistically significant, have great effect on classifier accuracy. Consequently, the main hypothesis, i.e. H6, is validated.

In the next three tables we present the pairs’ comparisons for the training performances. The second hypothesis (H2) is validated (Table XV) and ‘normalization’ is the best preprocessing approach, followed in order by ‘maximum absolute values’ and ‘no preprocessing’. Concerning the third hypothesis (H3), the best performance was obtained when data were normally distributed (Table XVI). The next best distribution was that of the real data, followed by uniform, logistic and

Table XI. Accuracy rates for training

PREPROC

DISTRIB

Technique

GAO

GA

GAM

GAA

GAU

RT

RT1

RT2

RT3

Unpreprocessed

REAL

93.02

92.86

92.92

93.30

92.22

92.43

92.41

UNIF

95.84

95.84

95.78

95.76

94.92

95.04

94.84

NORM

96.39

96.60

96.49

96.43

95.48

95.45

95.52

LOG

94.82

94.99

94.80

94.82

92.70

92.70

92.53

LAP

90.09

90.00

90.16

90.19

88.12

88.06

88.06

Normalization

REAL

99.43

99.49

99.46

99.33

99.11

99.10

99.08

UNIF

99.79

99.81

99.79

99.79

99.80

99.80

99.80

NORM

97.90

97.90

97.90

97.90

98.07

98.03

97.97

LOG

99.11

99.06

98.98

98.98

98.95

98.98

98.96

LAP

98.08

98.13

98.01

98.10

98.02

98.01

98.06

Max. of

REAL

99.68

99.68

99.68

99.68

99.69

99.69

99.69

absolute values

UNIF

97.79

97.77

97.73

97.84

97.77

97.77

97.89

NORM

96.91

97.00

97.02

97.02

96.93

96.90

96.91

LOG

96.50

96.52

96.52

96.52

96.68

96.59

96.60

LAP

95.64

95.83

95.76

95.81

95.26

95.47

95.23

Table XII. Accuracy rates for testing

PREPROC

DISTRIB

Technique

GAO

GA

GAM

GAA

GAU

RT

RT1

RT2

RT3

Unpreprocessed

REAL

85.24

85.48

85.48

85.95

83.49

83.97

93.81

UNIF

88.57

87.86

88.93

88.75

89.11

88.57

89.05

NORM

89.46

89.64

90.00

89.82

89.46

89.52

89.52

LOG

81.19

81.43

81.43

81.19

79.92

79.92

79.29

LAP

80.41

81.02

81.22

81.02

78.10

77.76

77.89

Normalization

REAL

85.71

86.19

85.71

85.71

85.79

85.63

85.79

UNIF

92.86

93.04

92.86

92.86

92.86

92.86

92.92

NORM

96.43

96.43

96.43

96.43

96.49

96.19

96.43

LOG

88.10

88.10

88.10

88.10

88.10

88.25

88.25

LAP

92.25

92.45

91.84

92.25

91.36

91.50

91.56

Max. of

REAL

97.62

97.62

97.62

97.62

97.54

97.70

97.62

absolute values

UNIF

95.00

95.36

95.71

95.36

96.31

96.43

96.07

NORM

93.21

93.57

93.93

93.57

92.86

93.15

93.27

LOG

88.10

88.10

87.86

88.10

88.25

88.10

88.25

LAP

88.37

89.18

88.98

89.59

89.86

89.93

89.86