My Awesome Newsletter

Archives
March 1, 2026

Edition #5: The Hidden Costs of API Wrappers

Welcome back to **Fine-Tuned**. This week, we're dissecting the "API Wrapper" business model, why so many of them are failing in 2026, and how to build a defensible AI product.

### 🔬 The Deep Dive: The Thin Wrapper Problem

In 2023 and 2024, if you put a nice UI over the OpenAI API, you could raise $2M. By late 2025, OpenAI (and Anthropic, and Google) had sherlocked 90% of those features natively into their core chat interfaces.

If your product's entire value proposition is "We send your text to an LLM and show you the response," you do not have a business. You have a feature.

**How to Build Defensibility in 2026:**

1. **Proprietary Data Moats**: Your LLM should be querying a database that *no one else has*. If you are building a legal AI, don't just use standard RAG on public laws. Partner with a law firm to get 10,000 anonymized, highly-specific contract negotiations and fine-tune on *that*.
2. **Workflow Integration (The System of Record)**: Don't build a chatbot. Build a tool that sits inside the user's existing workflow. If they use Salesforce, build a Chrome extension that reads the Salesforce DOM, generates the email, and logs the activity automatically.
3. **Multi-Step Action Execution**: The value isn't in generating the text; it's in executing the action. An API wrapper writes an email. A defensible agent writes the email, checks the user's calendar, finds a free time, books the Zoom link, and sends the calendar invite.

---

### 🗞️ The Roundup: 3 Big Updates This Week

1. **"Reasoning Tokens" become the new pricing metric:** API providers are shifting their pricing models. Instead of just paying for input/output tokens, you now pay specifically for "reasoning tokens" when models use System 2 thinking paths.
2. **Local Models Break the 30B Barrier on Consumer Hardware:** Thanks to aggressive quantization techniques, running massive 30B+ parameter models on a standard Mac Studio is now completely viable for local development.
3. **The Return of Specialized Hardware:** We are seeing a resurgence of edge devices with dedicated Neural Processing Units (NPUs) specifically designed to run SLMs at 100+ tokens per second.

---

### 🛠️ Tool of the Week: Vercel AI SDK (Core)

If you are still writing raw fetch requests to the OpenAI or Anthropic APIs, stop. The **Vercel AI SDK** has standardized the interface for all major models. You can swap out GPT-4 for Claude 3.7 with a single line of code, and it handles UI streaming, structured object generation (JSON), and tool calling natively.

---

*Keep building.*
- Kyle Anderson

Don't miss what's next. Subscribe to My Awesome Newsletter:
Powered by Buttondown, the easiest way to start and grow your newsletter.