#
normalization
It's important to look at how you initialize the values for your neurons. You essentially want them to be gaussian at the beginning so everything has a fair chance of training. If you look at the following, already after one layer, our standard deviation has almost tripled after performing the dot product.
g = torch.Generator().manual_seed(1)
x = torch.randn(1000, 10, generator=g)
w = torch.randn(10, 200, generator=g)
print(w)
y = x @ w
print(x.mean(), x.std())
# tensor(-0.0038) tensor(1.0057)
print(w.mean(), w.std())
# tensor(0.0003) tensor(1.0115)
print(y.mean(), y.std())
# tensor(-0.0064) tensor(3.2211)
To fix this, the referenced paper above ended up in the addition of the kaiming_normal operation, which divides the weights by the square root of the number of inputs to each neuron. It's a bit different depending on exactly which non-linearity you're using. For simplicity though you can commonly divide your weights by the square root of the fan-in, or the number of inputs to your neuron.
g = torch.Generator().manual_seed(1)
x = torch.randn(1000, 10, generator=g)
w = torch.randn(10, 200, generator=g) / 10**.5
print(w)
y = x @ w
print(x.mean(), x.std())
# tensor(-0.0038) tensor(1.0057)
print(w.mean(), w.std())
# tensor(9.0598e-05) tensor(0.3199)
print(y.mean(), y.std())
# tensor(-0.0020) tensor(1.0186)
Multiplying your weights by this changes the standard deviation because if you look at a layer with standard deviation 1 and multiply it by .2
the standard deviation is now .2
.
g = torch.Generator().manual_seed(1)
w = torch.randn(10, 200, generator=g)
print(w.std())
# tensor(1.0293)
w = w * .2
print(w.std())
# tensor(0.2059)
So following that, we can set the standard deviation of a gaussian distribution by multiplying it by whatever our ideal distribution is. The kaiming documentation gives us the equation
std = \frac{gain}{\sqrt{fan\_mode}}
where gain is precalculated by them in the docs. For tanh it is \frac{5}{3}. And fan_mode
is the number of axons of the neuron. Applying this to our function
g = torch.Generator().manual_seed(1)
w = torch.randn(10, 200, generator=g) * ((5/3)/10**.5)
print(w.std())
# tensor(.3084)
So if doing it manually we would want to multiply our initial weights of a neuron with 10 axons by .3084
to have a more normal standard deviation after passing through the activation function. This helps with tanh because it's a squashing function, so we want to amplify the values a bit, to fight back against it squashing everything inward. This normalization is typically fixed by applying something like
#
Batch normalization
A concept published in this paper, provides a mechanism so that you don't have to normalize your weights at initialization but instead normalize them through the training cycle.
It's pretty complex and I don't fully understand it, but one critique of it is since it's working with the mean and standard deviation of each layer, it's coupling all of the inputs during training. So then independent inputs now have a dependency on each other and can affect training. What does that mean? I have no clue.
Another weird trait is that due to the way you calculate batch normalization, the biases actually ended up being effectively zeroed out. So when doing batch normalization you don't use bias in your neurons. It's important to remember the bias values wouldn't actually be zero, but because you're adding a constant value, the subtraction of the mean in the normalization removes any effect that bias would have.
You apply batch normalization on any layers of your choosing before or after the activation functions have been calculated, it's most common to do before though. You generate the normalization with the following
z = \frac{x - m}{s}\newline output = (z * g) + b\newline x = \text{data point}\newline m = \text{mean of the data set}\newline s = \text{standard deviation}\newline g = \text{arbitrary parameter}\newline b = \text{arbitrary parameter}
Another important fact is that since you are using this hidden layer of normalization, with its own weights and bias, you need to use those weights and biases when running inference passes on your network. Pytorch calculates a rolling average of these so that you can access them after training, and apply them where appropriate in your forward passes.
#
Inputs
It's also common to preform normalization and standardization of your inputs, so that they follow a more equal distribution as well. A good analogy is miles driven, if some users drove 100 miles and others drove 100,000 that creates a significantly large data set that could have large imapcts on the gradients of the neural network. We can normalize them by squashing them down to 0-1 and standardize them by the equation in
This is discussed in this video. Normalization and standardization seem to be used somewhat interchangeably.