What is Inference?
The process of using a trained AI model to generate predictions or outputs from new input data.
Why It Matters
Inference is what happens every time you use an AI tool, and its speed and cost directly affect user experience.
Real-World Example
When you ask ChatGPT a question, the model runs inference to generate the answer.
“Understanding terms like Inference matters because it helps you have better conversations with developers and make smarter decisions about your software. You do not need to be technical. You just need to know enough to ask the right questions.”
Related Terms
Large Language Model (LLM)
An AI system trained on massive amounts of text that can understand and generate human language.
Training Data
The dataset used to teach an AI model patterns and knowledge during its initial training.
Quantisation
Reducing the precision of numbers in an AI model to make it smaller and faster without losing much accuracy.
Latency
The time delay between a request and its response
Learn More at buildDay Melbourne
Want to understand these concepts hands-on? Join our one-day workshop and build a real web application from scratch.
Related Terms
Latency
The time delay between a request and its response
Large Language Model (LLM)
An AI system trained on massive amounts of text that can understand and generate human language.
Quantisation
Reducing the precision of numbers in an AI model to make it smaller and faster without losing much accuracy.
Training Data
The dataset used to teach an AI model patterns and knowledge during its initial training.
Transformer
A type of AI architecture that processes text by paying attention to relationships between all words at once, rather...
Attention Mechanism
A technique that lets AI models focus on the most relevant parts of the input when generating output.