Commit 467c4be8 authored by Margret A. Riegert's avatar Margret A. Riegert
Browse files

Remove fft epochs


Former-commit-id: 1324a94c2a0ad85911d8e120c311e52c93082563 [formerly d7f60ba914979f903f3a879f110de94c0f046b4f]
Former-commit-id: aeac5fd6613f0e808e6868ba49ff64cf3e6aee7c
parent 2018ed97
225 1 2
0.0
-0.0 1.0
0.345703125
-0.824589312077 -0.565731823444
0.375
-0.707106769085 -0.707106769085
0.34375
-0.831469595432 -0.555570244789
0.44921875
-0.31368175149 -0.949528157711
0.0
-0.0 1.0
0.3056640625
-0.939459204674 -0.342660725117
0.125
-0.707106769085 0.707106769085
0.0
-0.0 1.0
0.453125
-0.290284663439 -0.956940352917
0.25
-1.0 6.12323426293e-17
0.3125
-0.923879504204 -0.382683426142
0.4765625
-0.146730467677 -0.989176511765
0.0205078125
-0.128498107195 0.991709768772
0.1796875
-0.903989315033 0.427555084229
0.25
-1.0 6.12323426293e-17
0.2001953125
-0.951435029507 0.307849645615
0.0
-0.0 1.0
0.25
-1.0 6.12323426293e-17
0.01953125
-0.122410677373 0.992479562759
0.28125
-0.980785250664 -0.195090323687
0.0
-0.0 1.0
0.25
-1.0 6.12323426293e-17
0.455078125
-0.278519690037 -0.960430502892
0.25
-1.0 6.12323426293e-17
0.3134765625
-0.921514034271 -0.38834503293
0.0
-0.0 1.0
0.453125
-0.290284663439 -0.956940352917
0.125
-0.707106769085 0.707106769085
0.3125
-0.923879504204 -0.382683426142
0.125
-0.707106769085 0.707106769085
0.28125
-0.980785250664 -0.195090323687
0.203125
-0.956940352917 0.290284663439
0.375
-0.707106769085 -0.707106769085
0.0
-0.0 1.0
0.1953125
-0.941544055939 0.336889863014
0.375
-0.707106769085 -0.707106769085
0.1015625
-0.595699310303 0.80320751667
0.3671875
-0.740951120853 -0.671558976173
0.18359375
-0.914209783077 0.405241310596
0.453125
-0.290284663439 -0.956940352917
0.4296875
-0.427555084229 -0.903989315033
0.2421875
-0.998795449734 0.0490676760674
0.125
-0.707106769085 0.707106769085
0.486328125
-0.0857973098755 -0.996312618256
0.3125
-0.923879504204 -0.382683426142
0.25
-1.0 6.12323426293e-17
0.40625
-0.555570244789 -0.831469595432
0.47265625
-0.170961886644 -0.98527765274
0.421875
-0.471396744251 -0.881921291351
0.0
-0.0 1.0
0.0
-0.0 1.0
0.17578125
-0.893224298954 0.449611335993
0.318359375
-0.90916800499 -0.416429549456
0.396484375
-0.605511069298 -0.795836925507
0.25
-1.0 6.12323426293e-17
0.28125
-0.980785250664 -0.195090323687
0.125
-0.707106769085 0.707106769085
0.4765625
-0.146730467677 -0.989176511765
0.125
-0.707106769085 0.707106769085
0.03125
-0.195090323687 0.980785250664
0.34375
-0.831469595432 -0.555570244789
0.2265625
-0.989176511765 0.146730467677
0.0
-0.0 1.0
0.125
-0.707106769085 0.707106769085
0.31640625
-0.914209783077 -0.405241310596
0.0
-0.0 1.0
0.234375
-0.995184719563 0.098017141223
0.0
-0.0 1.0
0.25
-1.0 6.12323426293e-17
0.1025390625
-0.600616455078 0.799537241459
0.0
-0.0 1.0
0.310546875
-0.928506076336 -0.371317207813
0.0
-0.0 1.0
0.4921875
-0.0490676760674 -0.998795449734
0.193359375
-0.937339007854 0.348418682814
0.23046875
-0.992479562759 0.122410677373
0.2421875
-0.998795449734 0.0490676760674
0.0380859375
-0.237023606896 0.971503913403
0.12890625
-0.724247097969 0.689540565014
0.375
-0.707106769085 -0.707106769085
0.0361328125
-0.225083917379 0.974339365959
0.498046875
-0.0122715383768 -0.999924719334
0.0
-0.0 1.0
0.1474609375
-0.799537241459 0.600616455078
0.0
-0.0 1.0
0.3125
-0.923879504204 -0.382683426142
0.453125
-0.290284663439 -0.956940352917
0.25
-1.0 6.12323426293e-17
0.0
-0.0 1.0
0.0751953125
-0.455083578825 0.890448749065
0.0
-0.0 1.0
0.162109375
-0.851355195045 0.524589657784
0.34375
-0.831469595432 -0.555570244789
0.1875
-0.923879504204 0.382683426142
0.2255859375
-0.988257586956 0.152797192335
0.365234375
-0.749136388302 -0.662415802479
0.0068359375
-0.0429382584989 0.999077737331
0.34375
-0.831469595432 -0.555570244789
0.0
-0.0 1.0
0.0
-0.0 1.0
0.451171875
-0.302005946636 -0.953306019306
0.296875
-0.956940352917 -0.290284663439
0.0625
-0.382683426142 0.923879504204
0.09375
-0.555570244789 0.831469595432
0.25
-1.0 6.12323426293e-17
0.33203125
-0.870086967945 -0.492898195982
0.0
-0.0 1.0
0.4375
-0.382683426142 -0.923879504204
0.041015625
-0.254865646362 0.966976463795
0.314453125
-0.919113874435 -0.393992036581
0.015625
-0.098017141223 0.995184719563
0.15625
-0.831469595432 0.555570244789
0.0498046875
-0.307849645615 0.951435029507
0.234375
-0.995184719563 0.098017141223
0.40625
-0.555570244789 -0.831469595432
0.07421875
-0.449611335993 0.893224298954
0.447265625
-0.32531028986 -0.945607304573
0.1875
-0.923879504204 0.382683426142
0.0625
-0.382683426142 0.923879504204
0.0078125
-0.0490676760674 0.998795449734
0.4482421875
-0.319502025843 -0.947585582733
0.09765625
-0.575808167458 0.817584812641
0.3046875
-0.941544055939 -0.336889863014
0.09375
-0.555570244789 0.831469595432
0.10107421875
-0.593232274055 0.805031359196
0.09375
-0.555570244789 0.831469595432
0.4580078125
-0.260794103146 -0.965394437313
0.078125
-0.471396744251 0.881921291351
0.146484375
-0.795836925507 0.605511069298
0.3125
-0.923879504204 -0.382683426142
0.25
-1.0 6.12323426293e-17
0.375
-0.707106769085 -0.707106769085
0.390625
-0.634393274784 -0.77301043272
0.375
-0.707106769085 -0.707106769085
0.0
-0.0 1.0
0.453125
-0.290284663439 -0.956940352917
0.0
-0.0 1.0
0.125
-0.707106769085 0.707106769085
0.0
-0.0 1.0
0.25
-1.0 6.12323426293e-17
0.25
-1.0 6.12323426293e-17
0.125
-0.707106769085 0.707106769085
0.0
-0.0 1.0
0.109375
-0.634393274784 0.77301043272
0.25
-1.0 6.12323426293e-17
0.4609375
-0.242980182171 -0.970031261444
0.0
-0.0 1.0
0.125
-0.707106769085 0.707106769085
0.4375
-0.382683426142 -0.923879504204
0.0
-0.0 1.0
0.404296875
-0.565731823444 -0.824589312077
0.0625
-0.382683426142 0.923879504204
0.4375
-0.382683426142 -0.923879504204
0.0
-0.0 1.0
0.078125
-0.471396744251 0.881921291351
0.02734375
-0.170961886644 0.98527765274
0.34375
-0.831469595432 -0.555570244789
0.0
-0.0 1.0
0.15625
-0.831469595432 0.555570244789
0.49609375
-0.0245412290096 -0.99969881773
0.3896484375
-0.639124453068 -0.769103348255
0.34375
-0.831469595432 -0.555570244789
0.0
-0.0 1.0
0.37109375
-0.724247097969 -0.689540565014
0.359375
-0.77301043272 -0.634393274784
0.0
-0.0 1.0
0.23828125
-0.997290432453 0.0735645666718
0.25
-1.0 6.12323426293e-17
0.0
-0.0 1.0
0.0625
-0.382683426142 0.923879504204
0.125
-0.707106769085 0.707106769085
0.02392578125
-0.149764537811 0.98872166872
0.078125
-0.471396744251 0.881921291351
0.0
-0.0 1.0
0.109375
-0.634393274784 0.77301043272
0.25
-1.0 6.12323426293e-17
0.0859375
-0.514102756977 0.857728600502
0.3359375
-0.857728600502 -0.514102756977
0.3125
-0.923879504204 -0.382683426142
0.2421875
-0.998795449734 0.0490676760674
0.0
-0.0 1.0
0.0
-0.0 1.0
0.0
-0.0 1.0
0.138671875
-0.765167236328 0.643831551075
0.09375
-0.555570244789 0.831469595432
0.0625
-0.382683426142 0.923879504204
0.0
-0.0 1.0
0.09375
-0.555570244789 0.831469595432
0.4375
-0.382683426142 -0.923879504204
0.265625
-0.995184719563 -0.098017141223
0.09375
-0.555570244789 0.831469595432
0.3232421875
-0.895966231823 -0.444122135639
0.08935546875
-0.532403111458 0.84649091959
0.46875
-0.195090323687 -0.980785250664
0.0
-0.0 1.0
0.380859375
-0.680601000786 -0.73265427351
0.28125
-0.980785250664 -0.195090323687
0.421875
-0.471396744251 -0.881921291351
0.25
-1.0 6.12323426293e-17
0.0
-0.0 1.0
0.119140625
-0.680601000786 0.73265427351
0.125
-0.707106769085 0.707106769085
0.0
-0.0 1.0
0.0
-0.0 1.0
0.0
-0.0 1.0
0.0
-0.0 1.0
0.1875
-0.923879504204 0.382683426142
0.49560546875
-0.0276081450284 -0.999618828297
0.0
-0.0 1.0
0.466796875
-0.207111373544 -0.978317379951
0.25
-1.0 6.12323426293e-17
0.03125
-0.195090323687 0.980785250664
0.0
-0.0 1.0
0.0
-0.0 1.0
0.078125
-0.471396744251 0.881921291351
0.3125
-0.923879504204 -0.382683426142
0.1953125
-0.941544055939 0.336889863014
0.1875
-0.923879504204 0.382683426142
0.03125
-0.195090323687 0.980785250664
0.078125
-0.471396744251 0.881921291351
0.15673828125
-0.833170175552 0.553016722202
0.3984375
-0.595699310303 -0.80320751667
0.24609375
-0.99969881773 0.0245412290096
0.375
-0.707106769085 -0.707106769085
FANN_FLO_2.1
num_layers=3
learning_rate=0.500000
connection_rate=1.000000
network_type=0
learning_momentum=0.000000
training_algorithm=2
train_error_function=1
train_stop_function=0
cascade_output_change_fraction=0.010000
quickprop_decay=-0.000100
quickprop_mu=1.750000
rprop_increase_factor=1.200000
rprop_decrease_factor=0.500000
rprop_delta_min=0.000000
rprop_delta_max=50.000000
rprop_delta_zero=0.100000
cascade_output_stagnation_epochs=12
cascade_candidate_change_fraction=0.010000
cascade_candidate_stagnation_epochs=12
cascade_max_out_epochs=150
cascade_min_out_epochs=50
cascade_max_cand_epochs=150
cascade_min_cand_epochs=50
cascade_num_candidate_groups=2
bit_fail_limit=3.49999994039535522461e-01
cascade_candidate_limit=1.00000000000000000000e+03
cascade_weight_multiplier=4.00000005960464477539e-01
cascade_activation_functions_count=10
cascade_activation_functions=3 5 7 8 10 11 14 15 16 17
cascade_activation_steepnesses_count=4
cascade_activation_steepnesses=2.50000000000000000000e-01 5.00000000000000000000e-01 7.50000000000000000000e-01 1.00000000000000000000e+00
layer_sizes=2 3 3
scale_included=0
neurons (num_inputs, activation_function, activation_steepness)=(0, 0, 0.00000000000000000000e+00) (0, 0, 0.00000000000000000000e+00) (2, 3, 5.00000000000000000000e-01) (2, 3, 5.00000000000000000000e-01) (0, 3, 0.00000000000000000000e+00) (3, 0, 5.00000000000000000000e-01) (3, 0, 5.00000000000000000000e-01) (0, 0, 0.00000000000000000000e+00)
connections (connected_to_neuron, weight)=(0, -1.37027778625488281250e+01) (1, 4.49666070938110351562e+00) (0, 1.32289533615112304688e+01) (1, -1.72329843044281005859e+00) (2, -3.60841751098632812500e+00) (3, -4.28367900848388671875e+00) (4, 4.13665485382080078125e+00) (2, 2.87105202674865722656e+00) (3, -1.95669329166412353516e+00) (4, -4.67934131622314453125e-01)
FANN_FLO_2.1
num_layers=3
learning_rate=0.500000
connection_rate=1.000000
network_type=0
learning_momentum=0.000000
training_algorithm=2
train_error_function=1
train_stop_function=0
cascade_output_change_fraction=0.010000
quickprop_decay=-0.000100
quickprop_mu=1.750000
rprop_increase_factor=1.200000
rprop_decrease_factor=0.500000
rprop_delta_min=0.000000
rprop_delta_max=50.000000
rprop_delta_zero=0.100000
cascade_output_stagnation_epochs=12
cascade_candidate_change_fraction=0.010000
cascade_candidate_stagnation_epochs=12
cascade_max_out_epochs=150
cascade_min_out_epochs=50
cascade_max_cand_epochs=150
cascade_min_cand_epochs=50
cascade_num_candidate_groups=2
bit_fail_limit=3.49999994039535522461e-01
cascade_candidate_limit=1.00000000000000000000e+03
cascade_weight_multiplier=4.00000005960464477539e-01
cascade_activation_functions_count=10
cascade_activation_functions=3 5 7 8 10 11 14 15 16 17
cascade_activation_steepnesses_count=4
cascade_activation_steepnesses=2.50000000000000000000e-01 5.00000000000000000000e-01 7.50000000000000000000e-01 1.00000000000000000000e+00
layer_sizes=2 3 3
scale_included=0
neurons (num_inputs, activation_function, activation_steepness)=(0, 0, 0.00000000000000000000e+00) (0, 0, 0.00000000000000000000e+00) (2, 3, 5.00000000000000000000e-01) (2, 3, 5.00000000000000000000e-01) (0, 3, 0.00000000000000000000e+00) (3, 0, 5.00000000000000000000e-01) (3, 0, 5.00000000000000000000e-01) (0, 0, 0.00000000000000000000e+00)
connections (connected_to_neuron, weight)=(0, 1.35890951156616210938e+01) (1, -1.73230099678039550781e+00) (0, -1.26018352508544921875e+01) (1, 4.08841896057128906250e+00) (2, -4.37116289138793945312e+00) (3, -3.77689194679260253906e+00) (4, 4.29474639892578125000e+00) (2, -1.69275057315826416016e+00) (3, 3.20263123512268066406e+00) (4, -8.28253805637359619141e-01)
../../../../../../../fann.template/train_1_hidden_layers_rprop
\ No newline at end of file
FANN_FLO_2.1
num_layers=3
learning_rate=0.500000
connection_rate=1.000000
network_type=0
learning_momentum=0.000000
training_algorithm=2
train_error_function=1
train_stop_function=0
cascade_output_change_fraction=0.010000
quickprop_decay=-0.000100
quickprop_mu=1.750000
rprop_increase_factor=1.200000
rprop_decrease_factor=0.500000
rprop_delta_min=0.000000
rprop_delta_max=50.000000
rprop_delta_zero=0.100000
cascade_output_stagnation_epochs=12
cascade_candidate_change_fraction=0.010000
cascade_candidate_stagnation_epochs=12
cascade_max_out_epochs=150
cascade_min_out_epochs=50
cascade_max_cand_epochs=150
cascade_min_cand_epochs=50
cascade_num_candidate_groups=2
bit_fail_limit=3.49999994039535522461e-01
cascade_candidate_limit=1.00000000000000000000e+03
cascade_weight_multiplier=4.00000005960464477539e-01
cascade_activation_functions_count=10
cascade_activation_functions=3 5 7 8 10 11 14 15 16 17
cascade_activation_steepnesses_count=4
cascade_activation_steepnesses=2.50000000000000000000e-01 5.00000000000000000000e-01 7.50000000000000000000e-01 1.00000000000000000000e+00
layer_sizes=2 3 3
scale_included=0
neurons (num_inputs, activation_function, activation_steepness)=(0, 0, 0.00000000000000000000e+00) (0, 0, 0.00000000000000000000e+00) (2, 3, 5.00000000000000000000e-01) (2, 3, 5.00000000000000000000e-01) (0, 3, 0.00000000000000000000e+00) (3, 0, 5.00000000000000000000e-01) (3, 0, 5.00000000000000000000e-01) (0, 0, 0.00000000000000000000e+00)
connections (connected_to_neuron, weight)=(0, 1.30167951583862304688e+01) (1, -1.69916963577270507812e+00) (0, -1.38468322753906250000e+01) (1, 4.56416225433349609375e+00) (2, -4.30477476119995117188e+00) (3, -3.60770773887634277344e+00) (4, 4.15533781051635742188e+00) (2, -2.02127599716186523438e+00) (3, 2.82119965553283691406e+00) (4, -4.01479542255401611328e-01)
../../../../../../../fann.template/train_1_hidden_layers_rprop
\ No newline at end of file
FANN_FLO_2.1
num_layers=3
learning_rate=0.500000
connection_rate=1.000000
network_type=0
learning_momentum=0.000000