TransformerEncoderLayer¶
- class
dragon.vm.torch.nn.
TransformerEncoderLayer
(
d_model,
nhead,
dim_feedforward=2048,
dropout=0.1,
activation='relu',
norm_first=False
)[source]¶ Layer for a standard transformer encoder . [Vaswani et.al, 2017].
Examples:
src = torch.ones(4, 2, 8) encoder_layer = torch.nn.TransformerEncoderLayer(d_model=8, nhead=2) out = encoder_layer(src)
__init__¶
TransformerEncoderLayer.
__init__
(
d_model,
nhead,
dim_feedforward=2048,
dropout=0.1,
activation='relu',
norm_first=False
)[source]¶Create a
TransformerEncoderLayer
.- Parameters:
- d_model (int) – The dimension of features.
- nhead (int) – The number of parallel heads.
- dim_feedforward (int, optional, default=2048) – The dimension of feedforward network.
- dropout (float, optional, default=0.1) – The dropout ratio.
- activation (str, optional, default='relu') – The activation function.
- norm_first (bool, optional, default=False) – Apply layer form before attention and feedforward.