Layer normalization is one of the three most common normalization strategies used in neural networks, the other two being batch normalization and group normalization. The name “layer normalization” is a bit of a misnomer, since you’re really normalizing across the features in an instance. To compute the layer normalization of an instance