If the module has parameters, this will update these parameters, according
to the accumulation of the gradients with respect to these parameters,
accumulated through backward()
calls.
The update is basically:
parameters = parameters - learningRate * gradients_wrt_parametersIf the module does not have parameters, it does nothing.