AI & Machine Learning Category
INFERENCE
Using a trained model to generate outputs.
Definition
Inference is the stage where a trained model receives new input and produces a prediction, classification, or generated response.
Practical Example & Use Case
Latency planning for an AI feature often focuses on inference time because slower model responses quickly affect the user experience in chat and search flows.
Editorial review date: 2026-04-14