# Lab: vLLM Llama-3.1-8B Caching Açık vs Kapalı

> Source: https://sukruyusufkaya.com/en/learn/prompt-caching-context-engineering/pcce-65-vllm-lab
> Updated: 2026-05-14T14:48:51.698Z
> Category: Prompt Caching & Context Engineering
> Module: 10. Self-Hosted Inference + Caching
**TLDR:** vLLM ile Llama-3.1-8B host edip, caching açık ve kapalı senaryolarda throughput + latency karşılaştırması.

