lstmLayer
Long shortterm memory (LSTM) layer
Description
An LSTM layer learns longterm dependencies between time steps in time series and sequence data.
The layer performs additive interactions, which can help improve gradient flow over long sequences during training.
Creation
Description
creates an LSTM layer and sets the layer
= lstmLayer(numHiddenUnits
)NumHiddenUnits
property.
sets additional layer
= lstmLayer(numHiddenUnits
,Name,Value
)OutputMode
, Activations, State, Parameters and Initialization, Learning Rate and Regularization, and
Name
properties using one or more namevalue pair arguments. You can specify multiple
namevalue pair arguments. Enclose each property name in quotes.
Properties
LSTM
NumHiddenUnits
— Number of hidden units
positive integer
Number of hidden units (also known as the hidden size), specified as a positive integer.
The number of hidden units corresponds to the amount of information remembered between time steps (the hidden state). The hidden state can contain information from all previous time steps, regardless of the sequence length. If the number of hidden units is too large, then the layer might overfit to the training data. This value can vary from a few dozen to a few thousand.
The hidden state does not limit the number of time steps that are processed in an
iteration. To split your sequences into smaller sequences for training, use the
'SequenceLength'
option in trainingOptions
.
Example: 200
OutputMode
— Output mode
'sequence'
(default)  'last'
Output mode, specified as one of the following:
'sequence'
– Output the complete sequence.'last'
– Output the last time step of the sequence.
HasStateInputs
— Flag for state inputs to layer
0
(false) (default)  1
(true)
Flag for state inputs to the layer, specified as 0
(false) or
1
(true).
If the HasStateInputs
property is 0
(false), then the
layer has one input with name 'in'
, which corresponds to the input data.
In this case, the layer uses the HiddenState
and
CellState
properties for the layer operation.
If the HasStateInputs
property is 1
(true), then the
layer has three inputs with names 'in'
, 'hidden'
, and
'cell'
, which correspond to the input data, hidden state, and cell
state respectively. In this case, the layer uses the values passed to these inputs for the
layer operation. If HasStateInputs
is
1
(true), then the HiddenState
and
CellState
properties must be empty.
HasStateOutputs
— Flag for state outputs from layer
0
(false) (default)  1
(true)
Flag for state outputs from the layer, specified as 0
(false) or
1
(true).
If the HasStateOutputs
property is 0
(false), then the
layer has one output with name 'out'
, which corresponds to the output
data.
If the HasStateOutputs
property is 1
(true), then the
layer has three outputs with names 'out'
,
'hidden'
, and 'cell'
, which correspond
to the output data, hidden state, and cell state, respectively. In this case, the
layer also outputs the state values computed during the layer operation.
InputSize
— Input size
'auto'
(default)  positive integer
Input size, specified as a positive integer or 'auto'
. If InputSize
is 'auto'
, then the software automatically assigns the input size at training time.
Example: 100
Activations
StateActivationFunction
— Activation function to update the cell and hidden state
'tanh'
(default)  'softsign'
Activation function to update the cell and hidden state, specified as one of the following:
'tanh'
– Use the hyperbolic tangent function (tanh).'softsign'
– Use the softsign function $$\text{softsign}(x)=\frac{x}{1+\leftx\right}$$.
The layer uses this option as the function $${\sigma}_{c}$$ in the calculations to update the cell and hidden state. For more information on how activation functions are used in an LSTM layer, see Long ShortTerm Memory Layer.
GateActivationFunction
— Activation function to apply to the gates
'sigmoid'
(default)  'hardsigmoid'
Activation function to apply to the gates, specified as one of the following:
'sigmoid'
– Use the sigmoid function $$\sigma (x)={(1+{e}^{x})}^{1}$$.'hardsigmoid'
– Use the hard sigmoid function$$\sigma (x)=\{\begin{array}{cc}\begin{array}{l}0\hfill \\ 0.2x+0.5\hfill \\ 1\hfill \end{array}& \begin{array}{l}\text{if}x2.5\hfill \\ \text{if}2.5\le x\le 2.5\hfill \\ \text{if}x2.5\hfill \end{array}\end{array}.$$
The layer uses this option as the function $${\sigma}_{g}$$ in the calculations for the layer gates.
State
CellState
— Cell state
numeric vector
Cell state to use in the layer operation, specified as a
NumHiddenUnits
by1 numeric vector. This value
corresponds to the initial cell state when data is passed to the
layer.
After setting this property manually, calls to the
resetState
function set the cell state to this
value.
If HasStateInputs
is
true
, then the CellState
property must be empty.
HiddenState
— Hidden state
numeric vector
Hidden state to use in the layer operation, specified as a
NumHiddenUnits
by1 numeric vector. This value corresponds to the
initial hidden state when data is passed to the layer.
After setting this property manually, calls to the resetState
function set the hidden state to this value.
If HasStateInputs
is true
, then
the HiddenState
property must be empty.
Parameters and Initialization
InputWeightsInitializer
— Function to initialize input weights
'glorot'
(default)  'he'
 'orthogonal'
 'narrownormal'
 'zeros'
 'ones'
 function handle
Function to initialize the input weights, specified as one of the following:
'glorot'
– Initialize the input weights with the Glorot initializer [4] (also known as Xavier initializer). The Glorot initializer independently samples from a uniform distribution with zero mean and variance2/(InputSize + numOut)
, wherenumOut = 4*NumHiddenUnits
.'he'
– Initialize the input weights with the He initializer [5]. The He initializer samples from a normal distribution with zero mean and variance2/InputSize
.'orthogonal'
– Initialize the input weights with Q, the orthogonal matrix given by the QR decomposition of Z = QR for a random matrix Z sampled from a unit normal distribution. [6]'narrownormal'
– Initialize the input weights by independently sampling from a normal distribution with zero mean and standard deviation 0.01.'zeros'
– Initialize the input weights with zeros.'ones'
– Initialize the input weights with ones.Function handle – Initialize the input weights with a custom function. If you specify a function handle, then the function must be of the form
weights = func(sz)
, wheresz
is the size of the input weights.
The layer only initializes the input weights when the
InputWeights
property is empty.
Data Types: char
 string
 function_handle
RecurrentWeightsInitializer
— Function to initialize recurrent weights
'orthogonal'
(default)  'glorot'
 'he'
 'narrownormal'
 'zeros'
 'ones'
 function handle
Function to initialize the recurrent weights, specified as one of the following:
'orthogonal'
– Initialize the recurrent weights with Q, the orthogonal matrix given by the QR decomposition of Z = QR for a random matrix Z sampled from a unit normal distribution. [6]'glorot'
– Initialize the recurrent weights with the Glorot initializer [4] (also known as Xavier initializer). The Glorot initializer independently samples from a uniform distribution with zero mean and variance2/(numIn + numOut)
, wherenumIn = NumHiddenUnits
andnumOut = 4*NumHiddenUnits
.'he'
– Initialize the recurrent weights with the He initializer [5]. The He initializer samples from a normal distribution with zero mean and variance2/NumHiddenUnits
.'narrownormal'
– Initialize the recurrent weights by independently sampling from a normal distribution with zero mean and standard deviation 0.01.'zeros'
– Initialize the recurrent weights with zeros.'ones'
– Initialize the recurrent weights with ones.Function handle – Initialize the recurrent weights with a custom function. If you specify a function handle, then the function must be of the form
weights = func(sz)
, wheresz
is the size of the recurrent weights.
The layer only initializes the recurrent weights when the
RecurrentWeights
property is empty.
Data Types: char
 string
 function_handle
BiasInitializer
— Function to initialize bias
'unitforgetgate'
(default)  'narrownormal'
 'ones'
 function handle
Function to initialize the bias, specified as one of the following:
'unitforgetgate'
– Initialize the forget gate bias with ones and the remaining biases with zeros.'narrownormal'
– Initialize the bias by independently sampling from a normal distribution with zero mean and standard deviation 0.01.'ones'
– Initialize the bias with ones.Function handle – Initialize the bias with a custom function. If you specify a function handle, then the function must be of the form
bias = func(sz)
, wheresz
is the size of the bias.
The layer only initializes the bias when the Bias
property is
empty.
Data Types: char
 string
 function_handle
InputWeights
— Input weights
[]
(default)  matrix
Input weights, specified as a matrix.
The input weight matrix is a concatenation of the four input weight matrices for the components (gates) in the LSTM layer. The four matrices are concatenated vertically in the following order:
Input gate
Forget gate
Cell candidate
Output gate
The input weights are learnable parameters. When training a network, if InputWeights
is nonempty, then trainNetwork
uses the InputWeights
property as the initial value. If InputWeights
is empty, then trainNetwork
uses the initializer specified by InputWeightsInitializer
.
At training time, InputWeights
is a
4*NumHiddenUnits
byInputSize
matrix.
RecurrentWeights
— Recurrent weights
[]
(default)  matrix
Recurrent weights, specified as a matrix.
The recurrent weight matrix is a concatenation of the four recurrent weight matrices for the components (gates) in the LSTM layer. The four matrices are vertically concatenated in the following order:
Input gate
Forget gate
Cell candidate
Output gate
The recurrent weights are learnable parameters. When training a network, if RecurrentWeights
is nonempty, then trainNetwork
uses the RecurrentWeights
property as the initial value. If RecurrentWeights
is empty, then trainNetwork
uses the initializer specified by RecurrentWeightsInitializer
.
At training time RecurrentWeights
is a
4*NumHiddenUnits
byNumHiddenUnits
matrix.
Bias
— Layer biases
[]
(default)  numeric vector
Layer biases for the LSTM layer, specified as a numeric vector.
The bias vector is a concatenation of the four bias vectors for the components (gates) in the LSTM layer. The four vectors are concatenated vertically in the following order:
Input gate
Forget gate
Cell candidate
Output gate
The layer biases are learnable parameters. When you train a
network, if Bias
is nonempty, then trainNetwork
uses the Bias
property as the
initial value. If Bias
is empty, then
trainNetwork
uses the initializer specified by BiasInitializer
.
At training time, Bias
is a
4*NumHiddenUnits
by1 numeric vector.
Learning Rate and Regularization
InputWeightsLearnRateFactor
— Learning rate factor for input weights
1 (default)  numeric scalar  1by4 numeric vector
Learning rate factor for the input weights, specified as a numeric scalar or a 1by4 numeric vector.
The software multiplies this factor by the global learning rate to determine the learning rate factor for the input weights of the layer. For example, if InputWeightsLearnRateFactor
is 2, then the learning rate factor for the input weights of the layer is twice the current global learning rate. The software determines the global learning rate based on the settings specified with the trainingOptions
function.
To control the value of the learning rate factor for the four
individual matrices in InputWeights
, specify a
1by4 vector. The entries of
InputWeightsLearnRateFactor
correspond to the
learning rate factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the matrices, specify a nonnegative scalar.
Example: 2
Example:
[1 2 1 1]
RecurrentWeightsLearnRateFactor
— Learning rate factor for recurrent weights
1 (default)  numeric scalar  1by4 numeric vector
Learning rate factor for the recurrent weights, specified as a numeric scalar or a 1by4 numeric vector.
The software multiplies this factor by the global learning rate to determine the learning rate for the recurrent weights of the layer. For example, if RecurrentWeightsLearnRateFactor
is 2, then the learning rate for the recurrent weights of the layer is twice the current global learning rate. The software determines the global learning rate based on the settings specified with the trainingOptions
function.
To control the value of the learning rate factor for the four
individual matrices in RecurrentWeights
, specify a
1by4 vector. The entries of
RecurrentWeightsLearnRateFactor
correspond to
the learning rate factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the matrices, specify a nonnegative scalar.
Example: 2
Example:
[1 2 1 1]
BiasLearnRateFactor
— Learning rate factor for biases
1 (default)  nonnegative scalar  1by4 numeric vector
Learning rate factor for the biases, specified as a nonnegative scalar or a 1by4 numeric vector.
The software multiplies this factor by the global learning rate
to determine the learning rate for the biases in this layer. For example, if
BiasLearnRateFactor
is 2
, then the learning rate for
the biases in the layer is twice the current global learning rate. The software determines the
global learning rate based on the settings you specify using the trainingOptions
function.
To control the value of the learning rate factor for the four
individual vectors in Bias
, specify a 1by4
vector. The entries of BiasLearnRateFactor
correspond to the learning rate factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the vectors, specify a nonnegative scalar.
Example:
2
Example:
[1 2 1 1]
InputWeightsL2Factor
— L2 regularization factor for input weights
1 (default)  numeric scalar  1by4 numeric vector
L2 regularization factor for the input weights, specified as a numeric scalar or a 1by4 numeric vector.
The software multiplies this factor by the global L2 regularization factor to determine the L2 regularization factor for the input weights of the layer. For example, if InputWeightsL2Factor
is 2, then the L2 regularization factor for the input weights of the layer is twice the current global L2 regularization factor. The software determines the L2 regularization factor based on the settings specified with the trainingOptions
function.
To control the value of the L2 regularization factor for the four
individual matrices in InputWeights
, specify a
1by4 vector. The entries of InputWeightsL2Factor
correspond to the L2 regularization factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the matrices, specify a nonnegative scalar.
Example: 2
Example:
[1 2 1 1]
RecurrentWeightsL2Factor
— L2 regularization factor for recurrent weights
1 (default)  numeric scalar  1by4 numeric vector
L2 regularization factor for the recurrent weights, specified as a numeric scalar or a 1by4 numeric vector.
The software multiplies this factor by the global L2 regularization factor to determine the L2 regularization factor for the recurrent weights of the layer. For example, if RecurrentWeightsL2Factor
is 2, then the L2 regularization factor for the recurrent weights of the layer is twice the current global L2 regularization factor. The software determines the L2 regularization factor based on the settings specified with the trainingOptions
function.
To control the value of the L2 regularization factor for the four
individual matrices in RecurrentWeights
, specify a
1by4 vector. The entries of
RecurrentWeightsL2Factor
correspond to the L2
regularization factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the matrices, specify a nonnegative scalar.
Example: 2
Example:
[1 2 1 1]
BiasL2Factor
— L2 regularization factor for biases
0 (default)  nonnegative scalar  1by4 numeric vector
L2 regularization factor for the biases, specified as a nonnegative scalar or a 1by4 numeric vector.
The software multiplies this factor by the global
L_{2} regularization factor to determine the
L_{2} regularization for the biases in this
layer. For example, if BiasL2Factor
is 2
, then the
L_{2} regularization for the biases in this layer
is twice the global L_{2} regularization factor. You can
specify the global L_{2} regularization factor using the
trainingOptions
function.
To control the value of the L2 regularization factor for the four
individual vectors in Bias
, specify a 1by4
vector. The entries of BiasL2Factor
correspond to
the L2 regularization factor of the following:
Input gate
Forget gate
Cell candidate
Output gate
To specify the same value for all the vectors, specify a nonnegative scalar.
Example:
2
Example:
[1 2 1 1]
Layer
Name
— Layer name
''
(default)  character vector  string scalar
Layer name, specified as a character vector or a string scalar.
For Layer
array input, the trainNetwork
,
assembleNetwork
, layerGraph
, and
dlnetwork
functions automatically assign names to layers with
Name
set to ''
.
Data Types: char
 string
NumInputs
— Number of inputs
1
 3
Number of inputs of the layer.
If the HasStateInputs
property is 0
(false), then the
layer has one input with name 'in'
, which corresponds to the input data.
In this case, the layer uses the HiddenState
and
CellState
properties for the layer operation.
If the HasStateInputs
property is 1
(true), then the
layer has three inputs with names 'in'
, 'hidden'
, and
'cell'
, which correspond to the input data, hidden state, and cell
state respectively. In this case, the layer uses the values passed to these inputs for the
layer operation. If HasStateInputs
is
1
(true), then the HiddenState
and
CellState
properties must be empty.
Data Types: double
InputNames
— Input names
{'in'}
 {'in','hidden','cell'}
Input names of the layer.
If the HasStateInputs
property is 0
(false), then the
layer has one input with name 'in'
, which corresponds to the input data.
In this case, the layer uses the HiddenState
and
CellState
properties for the layer operation.
If the HasStateInputs
property is 1
(true), then the
layer has three inputs with names 'in'
, 'hidden'
, and
'cell'
, which correspond to the input data, hidden state, and cell
state respectively. In this case, the layer uses the values passed to these inputs for the
layer operation. If HasStateInputs
is
1
(true), then the HiddenState
and
CellState
properties must be empty.
NumOutputs
— Number of outputs
1
 3
Number of outputs of the layer.
If the HasStateOutputs
property is 0
(false), then the
layer has one output with name 'out'
, which corresponds to the output
data.
If the HasStateOutputs
property is 1
(true), then the
layer has three outputs with names 'out'
,
'hidden'
, and 'cell'
, which correspond
to the output data, hidden state, and cell state, respectively. In this case, the
layer also outputs the state values computed during the layer operation.
Data Types: double
OutputNames
— Output names
{'out'}
 {'out','hidden','cell'}
Output names of the layer.
If the HasStateOutputs
property is 0
(false), then the
layer has one output with name 'out'
, which corresponds to the output
data.
If the HasStateOutputs
property is 1
(true), then the
layer has three outputs with names 'out'
,
'hidden'
, and 'cell'
, which correspond
to the output data, hidden state, and cell state, respectively. In this case, the
layer also outputs the state values computed during the layer operation.
Examples
Create LSTM Layer
Create an LSTM layer with the name 'lstm1'
and 100 hidden units.
layer = lstmLayer(100,'Name','lstm1')
layer = LSTMLayer with properties: Name: 'lstm1' InputNames: {'in'} OutputNames: {'out'} NumInputs: 1 NumOutputs: 1 HasStateInputs: 0 HasStateOutputs: 0 Hyperparameters InputSize: 'auto' NumHiddenUnits: 100 OutputMode: 'sequence' StateActivationFunction: 'tanh' GateActivationFunction: 'sigmoid' Learnable Parameters InputWeights: [] RecurrentWeights: [] Bias: [] State Parameters HiddenState: [] CellState: [] Show all properties
Include an LSTM layer in a Layer
array.
inputSize = 12;
numHiddenUnits = 100;
numClasses = 9;
layers = [ ...
sequenceInputLayer(inputSize)
lstmLayer(numHiddenUnits)
fullyConnectedLayer(numClasses)
softmaxLayer
classificationLayer]
layers = 5x1 Layer array with layers: 1 '' Sequence Input Sequence input with 12 dimensions 2 '' LSTM LSTM with 100 hidden units 3 '' Fully Connected 9 fully connected layer 4 '' Softmax softmax 5 '' Classification Output crossentropyex
Train Network for Sequence Classification
Train a deep learning LSTM network for sequencetolabel classification.
Load the Japanese Vowels data set as described in [1] and [2]. XTrain
is a cell array containing 270 sequences of varying length with 12 features corresponding to LPC cepstrum coefficients. Y
is a categorical vector of labels 1,2,...,9. The entries in XTrain
are matrices with 12 rows (one row for each feature) and a varying number of columns (one column for each time step).
[XTrain,YTrain] = japaneseVowelsTrainData;
Visualize the first time series in a plot. Each line corresponds to a feature.
figure plot(XTrain{1}') title("Training Observation 1") numFeatures = size(XTrain{1},1); legend("Feature " + string(1:numFeatures),'Location','northeastoutside')
Define the LSTM network architecture. Specify the input size as 12 (the number of features of the input data). Specify an LSTM layer to have 100 hidden units and to output the last element of the sequence. Finally, specify nine classes by including a fully connected layer of size 9, followed by a softmax layer and a classification layer.
inputSize = 12; numHiddenUnits = 100; numClasses = 9; layers = [ ... sequenceInputLayer(inputSize) lstmLayer(numHiddenUnits,'OutputMode','last') fullyConnectedLayer(numClasses) softmaxLayer classificationLayer]
layers = 5x1 Layer array with layers: 1 '' Sequence Input Sequence input with 12 dimensions 2 '' LSTM LSTM with 100 hidden units 3 '' Fully Connected 9 fully connected layer 4 '' Softmax softmax 5 '' Classification Output crossentropyex
Specify the training options. Specify the solver as 'adam'
and 'GradientThreshold'
as 1. Set the minibatch size to 27 and set the maximum number of epochs to 70.
Because the minibatches are small with short sequences, the CPU is better suited for training. Set 'ExecutionEnvironment'
to 'cpu'
. To train on a GPU, if available, set 'ExecutionEnvironment'
to 'auto'
(the default value).
maxEpochs = 70; miniBatchSize = 27; options = trainingOptions('adam', ... 'ExecutionEnvironment','cpu', ... 'MaxEpochs',maxEpochs, ... 'MiniBatchSize',miniBatchSize, ... 'GradientThreshold',1, ... 'Verbose',false, ... 'Plots','trainingprogress');
Train the LSTM network with the specified training options.
net = trainNetwork(XTrain,YTrain,layers,options);
Load the test set and classify the sequences into speakers.
[XTest,YTest] = japaneseVowelsTestData;
Classify the test data. Specify the same minibatch size used for training.
YPred = classify(net,XTest,'MiniBatchSize',miniBatchSize);
Calculate the classification accuracy of the predictions.
acc = sum(YPred == YTest)./numel(YTest)
acc = 0.9568
Classification LSTM Networks
To create an LSTM network for sequencetolabel classification, create a layer array containing a sequence input layer, an LSTM layer, a fully connected layer, a softmax layer, and a classification output layer.
Set the size of the sequence input layer to the number of features of the input data. Set the size of the fully connected layer to the number of classes. You do not need to specify the sequence length.
For the LSTM layer, specify the number of hidden units and the output mode 'last'
.
numFeatures = 12; numHiddenUnits = 100; numClasses = 9; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits,'OutputMode','last') fullyConnectedLayer(numClasses) softmaxLayer classificationLayer];
For an example showing how to train an LSTM network for sequencetolabel classification and classify new data, see Sequence Classification Using Deep Learning.
To create an LSTM network for sequencetosequence classification, use the same architecture as for sequencetolabel classification, but set the output mode of the LSTM layer to 'sequence'
.
numFeatures = 12; numHiddenUnits = 100; numClasses = 9; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits,'OutputMode','sequence') fullyConnectedLayer(numClasses) softmaxLayer classificationLayer];
Regression LSTM Networks
To create an LSTM network for sequencetoone regression, create a layer array containing a sequence input layer, an LSTM layer, a fully connected layer, and a regression output layer.
Set the size of the sequence input layer to the number of features of the input data. Set the size of the fully connected layer to the number of responses. You do not need to specify the sequence length.
For the LSTM layer, specify the number of hidden units and the output mode 'last'
.
numFeatures = 12; numHiddenUnits = 125; numResponses = 1; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits,'OutputMode','last') fullyConnectedLayer(numResponses) regressionLayer];
To create an LSTM network for sequencetosequence regression, use the same architecture as for sequencetoone regression, but set the output mode of the LSTM layer to 'sequence'
.
numFeatures = 12; numHiddenUnits = 125; numResponses = 1; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits,'OutputMode','sequence') fullyConnectedLayer(numResponses) regressionLayer];
For an example showing how to train an LSTM network for sequencetosequence regression and predict on new data, see SequencetoSequence Regression Using Deep Learning.
Deeper LSTM Networks
You can make LSTM networks deeper by inserting extra LSTM layers with the output mode 'sequence'
before the LSTM layer. To prevent overfitting, you can insert dropout layers after the LSTM layers.
For sequencetolabel classification networks, the output mode of the last LSTM layer must be 'last'
.
numFeatures = 12; numHiddenUnits1 = 125; numHiddenUnits2 = 100; numClasses = 9; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits1,'OutputMode','sequence') dropoutLayer(0.2) lstmLayer(numHiddenUnits2,'OutputMode','last') dropoutLayer(0.2) fullyConnectedLayer(numClasses) softmaxLayer classificationLayer];
For sequencetosequence classification networks, the output mode of the last LSTM layer must be 'sequence'
.
numFeatures = 12; numHiddenUnits1 = 125; numHiddenUnits2 = 100; numClasses = 9; layers = [ ... sequenceInputLayer(numFeatures) lstmLayer(numHiddenUnits1,'OutputMode','sequence') dropoutLayer(0.2) lstmLayer(numHiddenUnits2,'OutputMode','sequence') dropoutLayer(0.2) fullyConnectedLayer(numClasses) softmaxLayer classificationLayer];
Algorithms
Long ShortTerm Memory Layer
An LSTM layer learns longterm dependencies between time steps in time series and sequence data.
The state of the layer consists of the hidden state (also known as the output state) and the cell state. The hidden state at time step t contains the output of the LSTM layer for this time step. The cell state contains information learned from the previous time steps. At each time step, the layer adds information to or removes information from the cell state. The layer controls these updates using gates.
The following components control the cell state and hidden state of the layer.
Component  Purpose 

Input gate (i)  Control level of cell state update 
Forget gate (f)  Control level of cell state reset (forget) 
Cell candidate (g)  Add information to cell state 
Output gate (o)  Control level of cell state added to hidden state 
This diagram illustrates the flow of data at time step t. The diagram highlights how the gates forget, update, and output the cell and hidden states.
The learnable weights of an LSTM layer are the input weights W
(InputWeights
), the recurrent weights R
(RecurrentWeights
), and the bias b
(Bias
). The matrices W, R,
and b are concatenations of the input weights, the recurrent weights, and
the bias of each component, respectively. These matrices are concatenated as follows:
$$W=\left[\begin{array}{c}{W}_{i}\\ {W}_{f}\\ {W}_{g}\\ {W}_{o}\end{array}\right],R=\left[\begin{array}{c}{R}_{i}\\ {R}_{f}\\ {R}_{g}\\ {R}_{o}\end{array}\right],b=\left[\begin{array}{c}{b}_{i}\\ {b}_{f}\\ {b}_{g}\\ {b}_{o}\end{array}\right],$$
where i, f, g, and o denote the input gate, forget gate, cell candidate, and output gate, respectively.
The cell state at time step t is given by
$${c}_{t}={f}_{t}\odot {c}_{t1}+{i}_{t}\odot {g}_{t},$$
where $$\odot $$ denotes the Hadamard product (elementwise multiplication of vectors).
The hidden state at time step t is given by
$${h}_{t}={o}_{t}\odot {\sigma}_{c}({c}_{t}),$$
where $${\sigma}_{c}$$ denotes the state activation function. The lstmLayer
function, by default, uses the hyperbolic tangent function (tanh) to compute the state
activation function.
The following formulas describe the components at time step t.
Component  Formula 

Input gate  $${i}_{t}={\sigma}_{g}({W}_{i}{x}_{t}+\text{}{\text{R}}_{i}{h}_{t1}+{b}_{i})$$ 
Forget gate  $${f}_{t}={\sigma}_{g}({W}_{f}{x}_{t}+\text{}{\text{R}}_{f}{h}_{t1}+{b}_{f})$$ 
Cell candidate  $${g}_{t}={\sigma}_{c}({W}_{g}{x}_{t}+\text{}{\text{R}}_{g}{h}_{t1}+{b}_{g})$$ 
Output gate  $${o}_{t}={\sigma}_{g}({W}_{o}{x}_{t}+\text{}{\text{R}}_{o}{h}_{t1}+{b}_{o})$$ 
In these calculations, $${\sigma}_{g}$$ denotes the gate activation function. The lstmLayer
function, by default, uses the sigmoid function given by $$\sigma (x)={(1+{e}^{x})}^{1}$$ to compute the gate activation function.
Layer Input and Output Formats
Layers in a layer array or layer graph pass data specified as formatted dlarray
objects.
You can interact with these dlarray
objects in automatic differentiation workflows such as when developing a custom layer, using a functionLayer
object, or using the forward
and predict
functions with dlnetwork
objects.
This table shows the supported input formats of a LSTMLayer
object and
the corresponding output format. If the output of the layer is passed to a custom layer that
does not inherit from the nnet.layer.Formattable
class, or a
FunctionLayer
object with the Formattable
option set
to false
, then the layer receives an unformatted dlarray
object with dimensions ordered corresponding to the formats outlined in this table.
Input Format  OutputMode  Output Format 

 "sequence" 

"last" 

In dlnetwork
objects, LSTMLayer
objects also support the following input and output format combinations.
Input Format  OutputMode  Output Format 

 "sequence" 

"last" 
 
 "sequence" 

"last" 
 
 "sequence" 

"last" 

To use these input formats in trainNetwork
workflows, first convert the data to "CBT"
(channel, batch, time) format using flattenLayer
.
If the HasStateInputs
property is 1
(true), then the
layer has two additional inputs with names 'hidden'
and
'cell'
, which correspond to the hidden state and cell state,
respectively. These additional inputs expect input format "CB"
(channel,
batch).
If the HasStateOutputs
property is 1
(true), then the
layer has two additional outputs with names 'hidden'
and
'cell'
, which correspond to the hidden state and cell state,
respectively. These additional outputs have output format "CB"
(channel,
batch).
Compatibility Considerations
Default input weights initialization is Glorot
Behavior changed in R2019a
Starting in R2019a, the software, by default, initializes the layer input weights of this layer using the Glorot initializer. This behavior helps stabilize training and usually reduces the training time of deep networks.
In previous releases, the software, by default, initializes the layer input weights using the
by sampling from a normal distribution with zero mean and variance 0.01. To reproduce this
behavior, set the 'InputWeightsInitializer'
option of the layer to
'narrownormal'
.
Default recurrent weights initialization is orthogonal
Behavior changed in R2019a
Starting in R2019a, the software, by default, initializes the layer recurrent weights of this layer with Q, the orthogonal matrix given by the QR decomposition of Z = QR for a random matrix Z sampled from a unit normal distribution. This behavior helps stabilize training and usually reduces the training time of deep networks.
In previous releases, the software, by default, initializes the layer recurrent weights using
the by sampling from a normal distribution with zero mean and variance 0.01. To reproduce
this behavior, set the 'RecurrentWeightsInitializer'
option of the layer
to 'narrownormal'
.
References
[1] M. Kudo, J. Toyama, and M. Shimbo. "Multidimensional Curve Classification Using PassingThrough Regions." Pattern Recognition Letters. Vol. 20, No. 11–13, pages 1103–1111.
[2] UCI Machine Learning Repository: Japanese Vowels Dataset. https://archive.ics.uci.edu/ml/datasets/Japanese+Vowels
[3] Hochreiter, S, and J. Schmidhuber, 1997. Long shortterm memory. Neural computation, 9(8), pp.1735–1780.
[4] Glorot, Xavier, and Yoshua Bengio. "Understanding the Difficulty of Training Deep Feedforward Neural Networks." In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, 249–356. Sardinia, Italy: AISTATS, 2010.
[5] He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. "Delving Deep into Rectifiers: Surpassing HumanLevel Performance on ImageNet Classification." In Proceedings of the 2015 IEEE International Conference on Computer Vision, 1026–1034. Washington, DC: IEEE Computer Vision Society, 2015.
[6] Saxe, Andrew M., James L. McClelland, and Surya Ganguli. "Exact solutions to the nonlinear dynamics of learning in deep linear neural networks." arXiv preprint arXiv:1312.6120 (2013).
Extended Capabilities
C/C++ Code Generation
Generate C and C++ code using MATLAB® Coder™.
Usage notes and limitations:
When generating code with Intel^{®} MKLDNN:
The
StateActivationFunction
property must be set to'tanh'
.The
GateActivationFunction
property must be set to'sigmoid'
.The
HasStateInputs
andHasStateOutputs
properties must be set to0
(false).
GPU Code Generation
Generate CUDA® code for NVIDIA® GPUs using GPU Coder™.
Usage notes and limitations:
For GPU code generation, the
StateActivationFunction
property must be set to'tanh'
.For GPU code generation, the
GateActivationFunction
property must be set to'sigmoid'
.The
HasStateInputs
andHasStateOutputs
properties must be set to0
(false).
See Also
trainingOptions
 trainNetwork
 sequenceInputLayer
 bilstmLayer
 gruLayer
 convolution1dLayer
 maxPooling1dLayer
 averagePooling1dLayer
 globalMaxPooling1dLayer
 globalAveragePooling1dLayer
 Deep Network
Designer
Topics
 Sequence Classification Using Deep Learning
 Sequence Classification Using 1D Convolutions
 Time Series Forecasting Using Deep Learning
 SequencetoSequence Classification Using Deep Learning
 SequencetoSequence Regression Using Deep Learning
 Classify Videos Using Deep Learning
 Long ShortTerm Memory Networks
 List of Deep Learning Layers
 Deep Learning Tips and Tricks
Open Example
You have a modified version of this example. Do you want to open this example with your edits?
MATLAB Command
You clicked a link that corresponds to this MATLAB command:
Run the command by entering it in the MATLAB Command Window. Web browsers do not support MATLAB commands.
Select a Web Site
Choose a web site to get translated content where available and see local events and offers. Based on your location, we recommend that you select: .
You can also select a web site from the following list:
How to Get Best Site Performance
Select the China site (in Chinese or English) for best site performance. Other MathWorks country sites are not optimized for visits from your location.
Americas
 América Latina (Español)
 Canada (English)
 United States (English)
Europe
 Belgium (English)
 Denmark (English)
 Deutschland (Deutsch)
 España (Español)
 Finland (English)
 France (Français)
 Ireland (English)
 Italia (Italiano)
 Luxembourg (English)
 Netherlands (English)
 Norway (English)
 Österreich (Deutsch)
 Portugal (English)
 Sweden (English)
 Switzerland
 United Kingdom (English)