1.1 What is a Large Language Model?
Understand neural networks, transformers, and LLMs — parameters, training vs inference.
Understand neural networks, transformers, and LLMs — parameters, training vs inference.
Tokens, BPE encoding, context length limits, and why context explosion happens.
Autoregressive generation, temperature, top-p, and sampling strategies explained.
GPT, Claude, DeepSeek, Llama, Mistral — open vs closed models and size trade-offs.
Ollama local setup, first API call to OpenAI/Anthropic, and comparing outputs.