TransformerEncoderLayer#
- class torch.nn.TransformerEncoderLayer(d_model, nhead, dim_feedforward=2048, dropout=0.1, activation=<function relu>, layer_norm_eps=1e-05, batch_first=False, norm_first=False, bias=True, device=None, dtype=None)[source]#
TransformerEncoderLayer is made up of self-attn and feedforward network.
This TransformerEncoderLayer implements the original architecture described in the Attention Is All You Need paper. The intent of this layer is as a reference implementation for foundational understanding and thus it contains only limited features relative to newer Transformer architectures. Given the fast pace of innovation in transformer-like architectures, we recommend exploring this tutorial to build efficient layers from building blocks in core or using higher level libraries from the PyTorch Ecosystem.
TransformerEncoderLayer can handle either traditional torch.tensor inputs, or Nested Tensor inputs. Derived classes are expected to similarly accept both input formats. (Not all combinations of inputs are currently supported by TransformerEncoderLayer while Nested Tensor is in prototype state.)
If you are implementing a custom layer, you may derive it either from the Module or TransformerEncoderLayer class. If your custom layer supports both torch.Tensors and Nested Tensors inputs, make its implementation a derived class of TransformerEncoderLayer. If your custom Layer supports only torch.Tensor inputs, derive its implementation from Module.
- Parameters
d_model (int) – the number of expected features in the input (required).
nhead (int) – the number of heads in the multiheadattention models (required).
dim_feedforward (int) – the dimension of the feedforward network model (default=2048).
dropout (float) – the dropout value (default=0.1).
activation (Union[str, Callable[[Tensor], Tensor]]) – the activation function of the intermediate layer, can be a string (“relu” or “gelu”) or a unary callable. Default: relu
layer_norm_eps (float) – the eps value in layer normalization components (default=1e-5).
batch_first (bool) – If
True, then the input and output tensors are provided as (batch, seq, feature). Default:False(seq, batch, feature).norm_first (bool) – if
True, layer norm is done prior to attention and feedforward operations, respectively. Otherwise it’s done after. Default:False(after).bias (bool) – If set to
False,LinearandLayerNormlayers will not learn an additive bias. Default:True.
Examples
>>> encoder_layer = nn.TransformerEncoderLayer(d_model=512, nhead=8) >>> src = torch.rand(10, 32, 512) >>> out = encoder_layer(src)
- Alternatively, when
batch_firstisTrue: >>> encoder_layer = nn.TransformerEncoderLayer( ... d_model=512, nhead=8, batch_first=True ... ) >>> src = torch.rand(32, 10, 512) >>> out = encoder_layer(src)
- Fast path:
forward() will use a special optimized implementation described in FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness if all of the following conditions are met:
Either autograd is disabled (using
torch.inference_modeortorch.no_grad) or no tensor argumentrequires_gradtraining is disabled (using
.eval())batch_first is
Trueand the input is batched (i.e.,src.dim() == 3)activation is one of:
"relu","gelu",torch.functional.relu, ortorch.functional.geluat most one of
src_maskandsrc_key_padding_maskis passedif src is a NestedTensor, neither
src_masknorsrc_key_padding_maskis passedthe two
LayerNorminstances have a consistentepsvalue (this will naturally be the case unless the caller has manually modified one without modifying the other)
If the optimized implementation is in use, a NestedTensor can be passed for
srcto represent padding more efficiently than using a padding mask. In this case, a NestedTensor will be returned, and an additional speedup proportional to the fraction of the input that is padding can be expected.
- forward(src, src_mask=None, src_key_padding_mask=None, is_causal=False)[source]#
Pass the input through the encoder layer.
- Parameters
src (Tensor) – the sequence to the encoder layer (required).
src_mask (Optional[Tensor]) – the mask for the src sequence (optional).
src_key_padding_mask (Optional[Tensor]) – the mask for the src keys per batch (optional).
is_causal (bool) – If specified, applies a causal mask as
src mask. Default:False. Warning:is_causalprovides a hint thatsrc_maskis the causal mask. Providing incorrect hints can result in incorrect execution, including forward and backward compatibility.
- Return type
- Shape:
see the docs in
Transformer.