I Had an AI Agent Build, Deploy, and Instrument a Course Platform While I Watched
I asked Loki, my OpenClaw AI agent, to deploy OpenClaw Academy to Fly.io from scratch — sign up, configure, deploy, add analytics. Here's exactly what happened.
AI infrastructure, agent patterns, and things I learned building with OpenClaw.
I asked Loki, my OpenClaw AI agent, to deploy OpenClaw Academy to Fly.io from scratch — sign up, configure, deploy, add analytics. Here's exactly what happened.
Four providers in the fallback chain. Nine cascade failures in one day. How two config files out of sync turned redundancy into a cardboard wall.
Seven infrastructure gotchas from running a persistent AI daemon on macOS — from silent sleep mode to corrupted eval data.
112 consecutive failing runs on analyze tasks. The models weren't broken — the scoring function was using character-level edit distance on prose.
How IBM's smallest Granite model — picked as the control floor — ended up as one of the strongest performers in a 38-run evaluation.
A round-trip TTS evaluation comparing sherpa-onnx VITS, macOS say, and OpenAI's TTS APIs. The free offline model scored highest.
958 scored runs across 38 model/task pairs, seven task types, a two-judge ensemble, and zero promoted models. Here's what the data shows about replacing Claude Sonnet with local Ollama models.