10.1 Self-Hosting LLMsOllama, vLLM, TGI, hardware sizing, and GPU vs CPU inference for self-hosted LLMs.