Skip to content
Technical GlossaryDeep Learning

Cross-Attention

An attention mechanism that allows one representation set to draw context from another representation set.

Cross-attention is especially important in encoder-decoder architectures and multimodal systems. One side acts as the query while the other serves as the context source. This mechanism creates powerful matching structures for tasks such as translation, image captioning, question answering, and multimodal generation. It provides a learnable bridge across different information sources.