Concept
Multi-Head Attention 0
Multi-Head Attention is a mechanism that allows a model to focus on different parts of an input sequence simultaneously, enhancing its ability to capture diverse contextual relationships. By employing multiple attention heads, it enables the model to learn multiple representations of the input data, improving performance in tasks like translation and language modeling.
Relevant Degrees