# Multi-Head Attention: N Parallel Heads, Concat + Projection, Grouped-Query Attention (GQA), Multi-Query Attention (MQA)

> Source: https://sukruyusufkaya.com/en/learn/llm-muhendisligi/multi-head-attention-gqa-mqa-llama-3
> Updated: 2026-05-13T13:00:27.562Z
> Category: LLM Mühendisliği
> Module: Module 8: Attention Mathematics — The Heart of Transformer
**TLDR:** Why we split single attention into N parallel heads: each head's capacity to learn different patterns (syntactic, semantic, positional). Concat + output projection architecture, head pruning empirical findings, Llama-3 grouped-query attention (GQA), Mistral multi-query attention (MQA), head visualization with Turkish examples.

