View on GitHub

maximal

A TensorFlow-compatible Python library that provides models and layers to implement custom Transformer neural networks. Built on TensorFlow 2.

MultiHeadSelfAttention()

Multi Head Self-Attention layer, it is a concatenation of SelfAttention() layers. (tensorflow.keras.layers.Layer)

Inherits from tensorflow.keras.layers.Layer.

Arguments

__init__ arguments:

call arguments:

Returns