# Birth of RLHF: A Seven-Year Journey from Christiano 2017 to ChatGPT — Historical and Philosophical Anatomy of Human Preference Alignment

> Source: https://sukruyusufkaya.com/en/learn/llm-muhendisligi/rlhf-dogusu-christiano-2017-chatgpt-tarih-felsefe
> Updated: 2026-05-13T13:00:29.665Z
> Category: LLM Mühendisliği
> Module: Module 15: Preference Alignment — RLHF, PPO, DPO, GRPO
**TLDR:** Historical and philosophical foundations of RLHF: a seven-year transformation starting from Christiano et al. 2017 'Deep RL from Human Preferences', through Stiennon 2020 summarization work, Ouyang 2022 InstructGPT, to December 2022 ChatGPT launch. Why SFT alone is insufficient, the tension of the 'helpful-harmless-honest' triangle, Goodhart's Law and the reward hacking problem. What alignment means with Turkish cultural context — sen/siz distinction, social sensitivity, KVKK boundaries. The most conceptually critical lesson of the curriculum.

