Scaling
|
{ None, Z-Score, Min-Max Scaling }
|
Training algorithm
|
{ SGD, AdaGrad, RMSProp, Adam }
|
Activation function
|
{ Sigmoid, ReLU, tanh, linear (in the output layer) }
|
Hours of input data
|
{ 24 × 3, 24 × 5, 24 × 7, 24 × 9 }
|
Learning rate
|
{lrd×10−1,lrd,lrd×101,lrd×102} with (lrd) = default learning rate of the
|
|
corresponding optimiser as implemented in the python keras api
|
Hidden layers
|
{1,2,3,4}
|
Decay
|
{0,0.0001,0.001,0.01}
|
Patience of early stopping
|
{10,20,30}
|
Test split
|
{0.25,0.3,0.35}
|
L2−Regularisation
|
λ∈{ 0, 0.001, 0.01, 0.1} |
Dropout
|
{ 0.1, 0.2, 0.3 }
|