Support Vector Machine (SVM)
A classification algorithm that finds the optimal hyperplane separating different classes by maximizing the margin between the nearest data points of each class, called support vectors.
Support Vector Machines find the decision boundary that maximizes the distance (margin) to the closest training examples from each class. This maximum-margin principle provides strong theoretical guarantees about generalization and produces classifiers that are robust to overfitting, especially in high-dimensional spaces.
The kernel trick extends SVMs to nonlinear problems by implicitly mapping data into higher-dimensional spaces where linear separation is possible. The RBF (Radial Basis Function) kernel is most popular, effectively measuring similarity between points and enabling complex, curved decision boundaries without explicitly computing the high-dimensional mapping.
SVMs were dominant in machine learning before deep learning and remain relevant for specific use cases: small to medium datasets where they often outperform neural networks, high-dimensional data like text classification and genomics where their performance scales well, and situations requiring strong theoretical guarantees. For most production growth applications today, gradient boosting has largely replaced SVMs for tabular data, and transformers have replaced them for text, but SVMs remain a useful tool in the ML practitioner's toolkit for specific scenarios.
Related Terms
RAG (Retrieval-Augmented Generation)
A technique that grounds LLM responses in external data by retrieving relevant documents at query time and injecting them into the prompt context.
Embeddings
Dense vector representations of text, images, or other data that capture semantic meaning in a high-dimensional space, enabling similarity search and clustering.
Vector Database
A specialized database optimized for storing, indexing, and querying high-dimensional vector embeddings with sub-millisecond similarity search.
LLM (Large Language Model)
A neural network trained on massive text corpora that can generate, understand, and transform natural language for tasks like summarization, classification, and conversation.
Fine-Tuning
The process of further training a pre-trained LLM on a domain-specific dataset to specialize its behavior, style, or knowledge for a particular task.
Prompt Engineering
The practice of designing and iterating on LLM input instructions to reliably produce desired outputs for a specific task.