My Awesome Newsletter

Archives
March 2, 2026

Edition #6: Fine-Tuning is Dead (Wait, Didn't I Just Say RAG Was Dead?)

Welcome back to Fine-Tuned. This week we are looking at the fine-tuning ecosystem.\n\n### 🔬 The Deep Dive: The Convergence of RAG and Fine-Tuning\n\nIf you look at Twitter, you’ll see a religious war between two camps:\nCamp A: \”You don’t need fine-tuning, just use better RAG.\”\nCamp B: \”RAG is too slow and loses context, just fine-tune an SLM.\”\n\nThe reality in 2026 is that the best teams are doing both.\n\n**The New Architecture: RAFT (Retrieval-Augmented Fine-Tuning)**\n\nInstead of choosing one or the other, we are seeing the rise of RAFT. Here is how it works:\n1. You take a massive, unstructured dataset (like your entire corporate knowledge base).\n2. You use a cheap model to generate synthetic Q&A pairs from that data. \”Given this paragraph about the Q3 budget, what are 5 questions a user might ask, and what are the exact answers?\”\n3. You fine-tune an SLM on these Q&A pairs so the model actually learns the domain knowledge and the required answering format.\n4. AT INFERENCE, you still use a vector database to pull in the exact numbers (since fine-tuning is bad at memorizing exact digits). \n\nThe fine-tuned model now knows how to answer and understands the jargon of your business, while the RAG system provides the exact, up-to-date facts.\n\n—-\n\n### 🗞️ The Roundup: 3 Big Updates This Week\n\n1. Synthetic Data is the Only Data Left: We have officially run out of high-quality human text on the internet to train models on. The biggest breakthrough this month was a paper proving that models trained on curated synthetic data actually outperform models trained on raw human internet data.\n2. The 1-Bit LLM: A new architecture just dropped that quantizes model weights down to a single bit (1 or 0) instead of 16-bit or 8-bit floating-point numbers. This means we might soon be running 70B parameter models natively on iPhones with zero latency.\n3. \”Small\” is Relative: A year ago, a 7B model was considered small. Today, \”small\” means 1B or 2B parameters. These ultra-tiny models are being embedded directly into specific features rather than running as a centralized brain for the whole app.\n\n—-\n\n### 🛠️ Tool of the Week: Axolotl\n\nIf you want to try fine-tuning but don’t want to deal with CUDA out-of-memory errors and complex PyTorch scripts, use Axolotl. It’s a declarative configuration framework for fine-tuning. You write a single YAML file describing your dataset and model, and Axolotl handles the entire training loop efficiently.\n\n—-\n\n*Keep building.*\n- Kyle Anderson

Don't miss what's next. Subscribe to My Awesome Newsletter:
Powered by Buttondown, the easiest way to start and grow your newsletter.