HTTP/2
A major revision of the HTTP protocol that improves performance through multiplexing, header compression, server push, and stream prioritization. HTTP/2 enables multiple concurrent requests over a single TCP connection, eliminating head-of-line blocking at the application layer.
HTTP/2 addresses the fundamental performance limitations of HTTP/1.1. Multiplexing allows multiple requests and responses to be interleaved on a single connection, eliminating the need for multiple TCP connections and the associated overhead. Header compression using HPACK reduces bandwidth usage for repeated headers. Server push allows the server to proactively send resources it knows the client will need.
For AI product teams, HTTP/2 improves the performance of applications that make multiple API calls to AI services. A page that fetches recommendations, search results, and personalization data can issue all requests simultaneously over a single connection, reducing total page load time. Growth teams benefit because faster page loads directly correlate with higher conversion rates and engagement. HTTP/2's server push capability can proactively deliver AI-generated content that the application will need, reducing perceived latency. Most modern web servers, CDNs, and browsers support HTTP/2, making it a low-effort, high-impact infrastructure improvement for any AI product serving web-based interfaces.
Related Terms
Content Delivery Network
A geographically distributed network of proxy servers that caches and delivers content from locations closest to end users. CDNs reduce latency, improve load times, and absorb traffic spikes by serving content from edge nodes rather than a single origin server.
Edge Computing
A distributed computing paradigm that processes data closer to the source of generation rather than in a centralized data center. Edge computing reduces latency, conserves bandwidth, and enables real-time processing for latency-sensitive applications.
Serverless Computing
A cloud execution model where the provider dynamically manages server allocation and scaling. Developers deploy functions or containers without provisioning infrastructure, paying only for actual compute time consumed rather than reserved capacity.
Function as a Service
A serverless computing category where developers deploy individual functions that execute in response to events. FaaS platforms like AWS Lambda, Google Cloud Functions, and Azure Functions handle all infrastructure management, scaling each function independently.
Platform as a Service
A cloud computing model that provides a complete development and deployment environment without managing underlying infrastructure. PaaS offerings like Heroku, Vercel, and Google App Engine handle servers, storage, networking, and runtime configuration.
Infrastructure as a Service
A cloud computing model that provides virtualized computing resources over the internet. IaaS offerings like AWS EC2, Google Compute Engine, and Azure Virtual Machines give teams full control over servers, storage, and networking without owning physical hardware.