Inference
Also known as: Model Inference, Prediction
In one sentence
When a trained AI model processes new input and generates a prediction or response—the 'using' phase after training is done.
Explain like I'm 12
After you've trained AI by showing it examples, inference is when you actually use it—like asking a question and getting an answer.
In context
Every time you chat with ChatGPT, you're triggering inference. The model uses what it learned during training to generate responses to your prompts.
See also
Related Guides
Learn more about Inference in these guides:
Efficient Inference Optimization
AdvancedOptimize AI inference for speed and cost: batching, caching, model serving, KV cache, speculative decoding, and more.
8 min readMachine Learning Fundamentals: How Machines Learn from Data
BeginnerUnderstand the basics of machine learning. From training to inference—a practical introduction to how ML systems work without deep math or coding.
11 min readSupervised vs Unsupervised Learning: When to Use Which
BeginnerUnderstand the difference between supervised and unsupervised learning. Learn when to use each approach with practical examples and decision frameworks.
9 min read