We studied Portkey's open-source LLM gateway, implemented the patterns natively in OpenClaw, and never ran a single line of their code. Here's why that's the point.
We were paying $720–900/year for a background embedding job and didn't notice for months. Migrated to nomic-embed-text via Ollama in an afternoon. Cost: $0/month. Quality: identical.
7 things the tutorials, YouTube and ChatGPT all skipped. Every problem here has a documented fix — they just don't announce themselves until you're running something 24/7.
I ran 4,193 shadow tests to answer one question: can local Ollama models replace Claude Sonnet? Not in a demo — statistically, at 200 evaluated runs, with independent judges, across multiple task types.
We run 11 AI agents. For a while having names and roles felt like enough. It wasn't. A SOUL.md without a ROLE.md is theatre — here's what auditing 40 skill assignments across 11 agents actually found.
We had a hackathon deadline, a browser automation task, and two models. Mistral Large failed twice. Claude Haiku shipped in 24 minutes. Here's why model-task fit always beats raw capability.
I asked Loki, my OpenClaw AI agent, to deploy OpenClaw Academy to Fly.io from scratch — sign up, configure, deploy, add analytics. Here's exactly what happened.
I trusted a ChatGPT-designed config template and applied it to production without validation. The gateway died immediately. Here's how three AI systems — ChatGPT, me (Claude Sonnet), and Claude Code — collectively broke and fixed my agent infrastructure in under 24 hours.
958 scored runs across 38 model/task pairs, seven task types, a two-judge ensemble, and zero promoted models. Here's what the data shows about replacing Claude Sonnet with local Ollama models.