More detailed documentation on Deep Learning Toolbox normalization layers?
Show older comments
Below are excerpts from the Deep Learning Toolbox documentation describing several neural network layer types that perform different kinds of input data normalization. For me, the decriptions are a bit too terse and textual to understand clearly the differences between the normalization operations that these different layers apply (both at training time and at test time). Is there any additional documentation to be found somewhere, with a more mathematical description, and perhaps illustrative figures?
imageInputLayer An image input layer inputs 2-D images to a network and applies data normalization.
batchNormalizationLayer A batch normalization layer normalizes a mini-batch of data across all observations for each channel independently. To speed up training of the convolutional neural network and reduce the sensitivity to network initialization, use batch normalization layers between convolutional layers and nonlinearities, such as ReLU layers.
groupNormalizationLayer A group normalization layer normalizes a mini-batch of data across grouped subsets of channels for each observation independently. To speed up training of the convolutional neural network and reduce the sensitivity to network initialization, use group normalization layers between convolutional layers and nonlinearities, such as ReLU layers.
instanceNormalizationLayer An instance normalization layer normalizes a mini-batch of data across each channel for each observation independently. To improve the convergence of training the convolutional neural network and reduce the sensitivity to network hyperparameters, use instance normalization layers between convolutional layers and nonlinearities, such as ReLU layers.
layerNormalizationLayer A layer normalization layer normalizes a mini-batch of data across all channels for each observation independently. To speed up training of recurrent and multilayer perceptron neural networks and reduce the sensitivity to network initialization, use layer normalization layers after the learnable layers, such as LSTM and fully connected layers.
Accepted Answer
More Answers (0)
Categories
Find more on Deep Learning Toolbox in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!


