Now available: This Is Server CountryGet the book
Computing & AI

inference

/IN-fer-ens/

The process of using a trained AI model to generate predictions, responses, or outputs based on new input. Unlike training (which happens once), inference happens billions of times daily as users interact with AI systems. A single ChatGPT response requires inference across all 96+ transformer layers, consuming significant computational resources.

Referenced in the Book

Discussed in Chapter 1 of This Is Server Country

Related Terms

Back to Glossary View all "I" terms