LLM (Large Language Model) — a transformer neural network with tens of billions to trillions of parameters, trained on a massive text corpus. Generates human-like output for any NLP task: chat, code, summarise, translate. 2026 leaders: GPT-5 / GPT-5 Pro (OpenAI), Claude Opus 4.7 / Sonnet 4.6 (Anthropic), Gemini 2.5 (Google), Llama 3 (Meta open), DeepSeek R1 (open). Parameters: 8B → 1.8T. Context window: 200k → 2M tokens.
Below: details, example, related terms, FAQ.
# OpenAI API call
curl https://api.openai.com/v1/chat/completions \
-H "Authorization: Bearer $OPENAI_API_KEY" \
-d '{
"model": "gpt-5",
"messages": [{"role": "user", "content": "Explain TCP vs UDP"}]
}'Llama 3 (Meta), Mistral Large, Qwen 2.5, DeepSeek R1 — free weights, MIT/Apache. Performance is approaching GPT-5.
70B model needs a server with 2× H100 (~$80k) or cloud GPU $5/h. ROI at > 10M tokens/day.
LLMs generate likely text, not facts. Mitigations: RAG (grounding) + fact-check output + low temperature (0.1-0.3 for facts).