Loading...
Loading...
The lifecycle of an LLM — pre-training, fine-tuning, alignment, and how inference works when you send a prompt
The model reads trillions of tokens and learns to predict the next word. Cost: $10-50M. Duration: Weeks.
The model learns to follow instructions from human-written conversations. Cost: $10-100K.
Humans rank outputs by quality. The model learns to be helpful, harmless, and honest. Cost: $5-50K.
When you send a prompt:
| Concept | Definition |
|---|---|
| Token | Unit of text (word/subword) |
| Context Window | Max tokens a model can process |
| Temperature | Controls randomness (0=deterministic) |
| Training vs Inference | Training = learning; Inference = using |
a) Answering questions b) Predicting next token c) Following instructions
a) Recurrent Language Hidden Framework
b) Reinforcement Learning from Human Feedback
c) Recursive Language Hierarchical Function