Model Support in a VLM Serving Stack Is Not a Checkbox - It Is a Six-Layer Systems Contract
Why real multimodal model support is a six-layer serving-stack contract, from API extraction to decoder reentry.
All the articles I've posted.
Why real multimodal model support is a six-layer serving-stack contract, from API extraction to decoder reentry.
How I designed, implemented, and hardened a cost-efficient RAG chatbot for my personal site with citations, streaming, and build-time indexing.
How gathering your information in one place transforms AI from a generic assistant into your personal superpower.
Why this feels like the shift from horsepower to a real vehicle—and why I am excited about the future.
Why vLLM requests can take 2-3 hours under heavy load - analyzing KV cache block exhaustion, queue starvation.
Continuing the Mini SGLang deep dive - covering request batching, overlap scheduling, and tensor parallelism.
Deep dive into Mini SGLang architecture - covering system design, engine initialization, KV cache, and single request lifecycle.