Professional Documents
Culture Documents
(1)
where i denotes the i-th layer (i.e., i = 0 is the input layer
and i = n+1 is the output layer), is the internal bias of the
-th neuron in the i-th layer, is the weight between the
-th neuron in the -th layer and the -th neuron in the
i-th layer, and for (i.e., the i-th layer has
neurons). This general DNN with N hidden neurons and M
output neurons can use a maximum of N+M different
activation functions.
Each neuron in each hidden or output layer of the general
DNN can have a different activation function. Hence, if there
are 100 neurons and 3 different activation functions, then
there are a total of 3100 combinations of functions for all the
neurons. The advantage is that a global optimization
algorithm can be used to find an optimal or near-optimal
function combination in addition to only locally optimizing
the parameters among all the neurons.
I.
INTRODUCTION
2849
TABLE II.
PERFORMANCE OF GDNN WITH SINGLE ACTIVATION
FUNCTION FOR 30-YEAR TREASURY CONSTANT MATURITY RATE
No.
Hidden
Layers
4
0.0101
0.0066
0.0078
0.0043
0.0051
0.0102
0.0135
0.0104
0.0120
0.0161
0.0172
0.0028
0.0243
0.0169
0.0162
0.0359
0.0201
0.0080
10
0.0242
0.0379
0.0239
0.0077
0.0597
0.0223
12
0.0360
0.0399
0.0316
0.0700
0.0365
0.0635
Avg.
0.0216
0.0223
0.0183
0.0268
0.0277
0.0213
TABLE III.
TRAINING PERFORMANCE OF GDNN WITH TWO
ACTIVATION FUNCTIONS FOR DOW JONES INDUSTRIAL AVERAGE
No.
Hidden
Layers
4
B-U
B-H
U-B
U-H
H-U
H-B
0.0292
0.0381
0.0401
0.0279
0.0202
0.0359
0.0358
0.0367
0.0306
0.0219
0.0281
0.0374
0.0385
0.0431
0.0430
0.0392
0.0458
0.0420
10
0.0446
0.0512
0.0515
0.0399
0.0394
0.0418
12
0.0436
0.0502
0.0501
0.0559
0.0324
0.0446
14
0.0576
0.0645
0.0526
0.0554
0.0540
0.0645
16
0.0645
0.0645
0.0645
0.0645
0.0645
0.0645
18
0.0580
0.0573
0.0645
0.0645
0.0645
0.0645
20
0.0645
0.0645
0.0645
0.0645
0.0645
0.0645
Avg.
0.0485
0.0522
0.0513
0.0482
0.0459
0.0511
0.0238
0.0247
0.0391
0.0098
0.0368
0.0211
0.0442
0.0442
0.0212
0.0163
0.0387
0.0349
0.0380
0.0386
0.0436
0.0330
0.0833
0.0411
10
0.0482
0.0603
0.0478
0.0720
0.1066
0.0164
12
14
0.0608
0.0520
Testing MSE
0.0645
0.1071
0.0601
0.0645
0.0645
0.0207
0.1024
0.1016
16
0.0361
0.0645
0.0645
0.0164
0.1022
0.1031
18
0.0563
0.0645
0.0645
0.0188
0.1022
0.1019
20
0.0645
0.0645
0.0645
0.1019
0.1023
0.1026
Avg.
0.0484
0.0531
0.0527
0.0440
0.0816
0.0637
B-U
B-H
U-B
U-H
H-U
H-B
0.0511
0.0479
0.0550
0.0491
0.0583
0.0532
0.0503
0.0633
Training MSE
TABLE I.
PERFORMANCE OF GDNN WITH SINGLE ACTIVATION
FUNCTION FOR DOW JONES INDUSTRIAL AVERAGE
Training MSE
Testing MSE
of the -th output neuron for the -th input data for K data
and L output neurons.
In this paper, all of the neurons in the same hidden layer
or the same output layer have the same activation function.
No.
Hidden
Layers
4
Training MSE
2850
No.
Hidden
Layers
4
B-U
B-H
U-B
U-H
H-U
H-B
0.0065
0.0060
0.0079
0.0079
0.0127
0.0101
0.0113
0.0117
0.0097
0.0108
0.0173
0.0106
0.0245
0.0189
0.0237
0.0138
0.0156
0.0202
10
0.0226
0.0240
0.0305
0.0272
0.0236
0.0147
12
0.0227
0.0200
0.0334
0.0304
0.0399
0.0391
Avg.
0.0175
0.0161
0.0211
0.0180
0.0218
0.0189
Training MSE
B-H
U-B
U-H
H-U
H-B
0.0132
0.0152
0.0154
0.0230
0.0302
0.0161
B-H-U
B-U-H
H-B-U
H-U-B
U-B-H
U-H-B
0.0387
0.0225
0.0222
0.0334
0.0230
0.0365
0.0492
0.0381
0.0349
0.0228
0.0256
0.0370
0.0470
0.0474
0.0372
0.0393
0.0404
0.0391
10
0.0500
0.0359
0.0380
0.0477
0.0411
0.0512
12
0.0572
0.0442
0.0541
0.0574
0.0492
0.0483
14
0.0645
0.0645
0.0515
0.0618
0.0430
0.0645
16
0.0645
0.0513
0.0645
0.0645
0.0550
0.0606
18
0.0645
0.0645
0.0645
0.0645
0.0645
0.0645
20
0.0645
0.0645
0.0645
0.0645
0.0645
0.0645
Avg.
0.0556
0.0481
0.0479
0.0507
0.0451
0.0518
Training MSE
B-U-H
H-B-U
H-U-B
U-B-H
U-H-B
0.0648
0.0742
0.0524
0.0762
0.0351
0.0694
B-H-U
B-U-H
H-B-U
H-U-B
U-B-H
U-H-B
0.0082
0.0081
0.0071
0.0074
0.0396
0.0058
0.0101
0.0101
0.0128
0.0128
0.0159
0.0103
0.0246
0.0073
0.0201
0.0156
0.0089
0.0190
10
0.0340
0.0307
0.0354
0.0235
0.0282
0.0273
12
0.0304
0.0152
0.0330
0.0396
0.0369
0.0262
Avg.
0.0214
0.0143
0.0217
0.0198
0.0259
0.0177
Training MSE
B-U-H
H-B-U
H-U-B
U-B-H
U-H-B
0.0234
0.0209
0.0279
0.0189
0.0172
0.0162
IV.
CONCLUSIONS
2851