# Normalization Revolution: LayerNorm, RMSNorm and Pre-LN vs Post-LN — Cornerstone of Training Stability

> Source: https://sukruyusufkaya.com/en/learn/llm-muhendisligi/layernorm-rmsnorm-pre-ln-post-ln
> Updated: 2026-05-13T13:00:28.349Z
> Category: LLM Mühendisliği
> Module: Module 10: Transformer Block — Anatomy of the Block
**TLDR:** Mathematical and systems anatomy of transformer training stability: LayerNorm (Ba 2016) classical formula, RMSNorm (Zhang 2019) — Llama-3's choice, why gain parameter only, computational savings. Pre-LN (modern) vs Post-LN (original Vaswani) trade-off, gradient flow, deep transformer stability. Normalization concerns in Turkish model fine-tuning.

