Jump to content

User:Jeblad/Standard notation (neural net)

From Wikipedia, the free encyclopedia

Standard notation as it is used within deep learning, has changed a lot since the first published works. It is undergoing some standardization, but mostly at an informal level.

Notation

[edit]

Indexes

[edit]
training
Superscript like denotes the iᵗʰ training example in a trainingset
layer
Superscript like denotes the lᵗʰ layer in a set of layers
sequence
Superscript like denotes the tᵗʰ item in a sequence of items
1D node
Subscript like denotes the iᵗʰ node in a one-dimensional layer
2D node
Subscript or like or denotes the node at iᵗʰ row and jᵗʰ column in a two-dimensional layer[note 1]
1D weight
Subscript or like or denotes the weight between node iᵗʰ at previous layer and jᵗʰ at following layer[note 2]

Sizes

[edit]
number of samples
is the number of samples in the dataset
input size
is the (possibly multidimensional) size of input (or number of features)
output size
is the (possibly multidimensional) size of output (or number of classes)
hidden units
is the number of units in hidden layer
number of layers
is the number of layers in the network
input sequence size
is the size of the input sequence
output sequence size
is the size of the output sequence
input training sequence size
is the size of the input training sequence (each sample training sequence)
output training sequence size
is the size of the output training sequence (each sample training sequence)

Other

[edit]
cross entropy
elementwise sequence loss
and by using cross entropy that is the sum would be over for classification in and out of a single class

References

[edit]


Notes

[edit]
  1. ^ This can easily be confused with a weight index.
  2. ^ Michael Nielson defines as weight from kᵗʰ neuron to jᵗʰ, while Andrew Ng defines it in opposite direction.