# Reasoning Trace Dataset Generation: Teacher Distillation + Self-Bootstrapping

> Source: https://sukruyusufkaya.com/en/learn/fine-tuning-cookbook/ftc-reasoning-trace-dataset-generation
> Updated: 2026-05-14T14:42:58.875Z
> Category: Fine-Tuning Cookbook (Model-by-Model)
> Module: Part XII — Reasoning Model FT (R1-style)
**TLDR:** Trace data generation for reasoning SFT: (a) Teacher distillation — DeepSeek-R1 (MIT license!), Gemini-thinking, o3 API; (b) Self-bootstrapping — small model generates traces + verifiable filter keeps correct; (c) Hybrid. Llama 3.1 70B teacher local serve + 10K trace generation on RTX 4090 (~24h).

