StochasticGradient
has several field which have an impact on a call to train()
.
-
learningRate
- This is the learning rate used during training. The update of the parameters will be
parameters = parameters - learningRate * parameters_gradient
Default value is 0.01
.
-
learningRateDecay
: The learning rate decay. If non-zero, the learning rate (note - the field learningRate
will not change value) will be computed after each iteration (pass over the dataset) with:
current_learning_rate =
learningRate / (1 + iteration * learningRateDecay)
-
maxIteration
- The maximum number of iteration (passes over the dataset). Default is
25
.
-
shuffleIndices
- Boolean which says if the examples will be randomly sampled or not. Default is
true
.
If false
, the examples will be taken in the order of the dataset.
-
hookExample
- A possible hook function which will be called (if non-nil) during training after each example forwarded and backwarded
through the network. The function takes
(self, example)
as parameters. Default is nil
.
-
hookIteration
- A possible hook function which will be called (if non-nil) during training after a complete pass
over the dataset. The function takes
(self, iteration)
as parameters. Default is nil
.