Skip to main content

Applied LLM Engineering — From Fundamentals to Production

AI-Generated Content

AI-generated content may contain errors. Always verify against official sources.

Applied LLM Engineering — From Fundamentals to Production

A practitioner's guide to building reliable LLM pipelines — from understanding how they work to shipping production systems.


What You Will Learn

By the end of this course you will be able to:

  • Explain how LLMs work at the token and sampling level
  • Call any LLM API (OpenAI, Anthropic, Google Gemini, open-source) confidently
  • Write production-quality prompts and evaluate them systematically
  • Build Retrieval-Augmented Generation (RAG) pipelines from scratch
  • Chain LLM calls using LangChain
  • Create autonomous agents with tool-calling
  • Fine-tune and deploy open-source models
  • Evaluate and monitor LLM systems in production

Course Map

#ChapterCore Skill
1LLM FoundationsHow models work, tokenization, model landscape
2Prompt EngineeringZero/few-shot, chain-of-thought, structured output
3LLM APIsOpenAI, Anthropic, Gemini, open-source via Ollama
4RAGEmbeddings, vector stores, retrieval pipelines
5ChainsLangChain LCEL, multi-step workflows
6AgentsTool-calling, ReAct, agent frameworks
7Fine-TuningLoRA, QLoRA, instruction tuning
8EvaluationRAGAS, LLM-as-judge, tracing
9DeploymentFastAPI, Docker, cost optimisation

Official Documentation References

Every chapter links to primary official sources:


Prerequisites

  • Python 3.10+ — comfortable with functions, classes, and pip
  • Basic understanding of REST APIs
  • No ML background required
Start Here

Head to 1.1 What is an LLM? to begin.