Back to glossary

LLM (Large Language Model)

A neural network trained on massive text corpora that can generate, understand, and transform natural language for tasks like summarization, classification, and conversation.

Large Language Models like GPT-4, Claude, and Gemini are transformer-based neural networks with billions of parameters, trained on trillions of tokens of text data. They learn statistical patterns in language that enable remarkably flexible text generation and understanding.

For product teams, LLMs unlock features that were previously impossible or required months of custom ML work: conversational interfaces, content generation, text classification, sentiment analysis, summarization, and translation. The key shift is from building task-specific models to prompting general-purpose models — dramatically reducing time to ship AI features.

The practical challenge is making LLMs reliable in production. They hallucinate, they're expensive at scale, and their behavior is hard to predict. Successful product teams invest heavily in prompt engineering, output validation, caching, model routing (using cheaper models for simpler tasks), and evaluation pipelines. The goal isn't perfection — it's building systems where LLM failures are gracefully handled and continuously improved.

Related Terms

Further Reading