TransformerLayer
Transformer Encoder layer.
Inherits from tensorflow.keras.layers.Layer.
Arguments
__init__
args:
depth
: (int) depth of the model (corresponds to embedding size).num_heads
: (int) number of attention heads.pwff_nodes
: (int) size of Dense ReLU layer in Pointwise FF block.rate
: (float) dropout probability. Defaults to 0.1 as in original paper.
call
args:
input_tensor
: (tf.tensor) input tensor (usually from PositionalEmbedding layer).
Returns
pwff_output
: (tf.tensor) Layer’s output.