OpenAI GPT-5/5.1 Distillation Pipeline: Stored Completions + FT API Hybrid
OpenAI 'Stored Completions' feature (2024+): after GPT-5/5.1 inference, save completions → free dataset for distill. FT GPT-4o-mini on these completions → small-model-big-knowledge transfer. License matters (only completions you generated with your API key).
Şükrü Yusuf KAYA
22 min read
Advanced1. Distillation Workflow#
1. Production'da GPT-5 ile inference yap client.chat.completions.create(model="gpt-5", messages=[...], store=True) 2. OpenAI dashboard'tan stored completions'ı download (JSONL) 3. Bu completions'ı FT dataset olarak GPT-4o-mini'ye FT et 4. Inference'i GPT-4o-mini-ft'a yönlendir → maliyet %95 düşer, kalite GPT-5'in %80-90'ı
Cost saving: GPT-5 input 30/M. GPT-4o-mini-ft input 1.20. 30-50x cheaper for ~%85 kalite.
✅ Teslim
- Production traffic'inde 1000 GPT-5 completion sakla. 2) GPT-4o-mini'ye FT et. 3) A/B test (GPT-5 vs FT'lı mini). 4) Sonraki ders: 14.4 — Anthropic Claude (Bedrock).
Yorumlar & Soru-Cevap
(0)Yorum yazmak için giriş yap.
Yorumlar yükleniyor...
Related Content
Part 0 — Engineering Foundations
Welcome to the Fine-Tuning Cookbook: System, Stage Taxonomy, and the Reproducibility Contract
Start LearningPart 0 — Engineering Foundations
Reproducibility Stack: Seeds, cuDNN Flags, and Deterministic CUDA — End the 'Works on My Machine' Problem
Start LearningPart 0 — Engineering Foundations