The 7 RAG Pipeline Mistakes Everyone Makes
Most RAG implementations fail not because of the LLM, but because of bad chunking, wrong embedding models, or missing re-ranking. We break down the seven most common mistakes and how to fix each one.
Welcome to the first AI Growth Stack Weekly. This week we're covering the foundations: RAG pipelines that actually work, how growth loops compound with AI, and the embedding models worth your attention in 2026.
Most RAG implementations fail not because of the LLM, but because of bad chunking, wrong embedding models, or missing re-ranking. We break down the seven most common mistakes and how to fix each one.
AI-powered growth loops create self-reinforcing cycles where each user's activity generates inputs that attract more users. Learn the three loop archetypes that work best for AI-native products.
We benchmarked OpenAI, Cohere, Voyage, and BGE models on real-world retrieval tasks. The results surprised us — the most expensive model isn't always the best for your use case.
A comprehensive overview of the embedding landscape, from choosing the right model dimensions to evaluating multilingual performance and cost trade-offs.
A great discussion thread on the current state of RAG adoption in production, with insights from engineers at Notion, Linear, and Vercel on what's actually working.