Hey, I’m Anthony.
Cloud engineering leader, tinkerer, and occasional writer. Field Notes is where I explore things in tech I find interesting — not always cloud, not always AI, just stuff worth digging into. Pull up a post.
Filed under: Cloud Engineering · AI Infrastructure · Local Lab You’ve Got APIs. Now You’ve Got AI APIs. Now What? Picture this: you grab an Ollama model, wire it into your app locally — done. Celebrate. But two months later? You’ve got five apps, a handful of models, and absolutely no visibility into what’s being called, how often, or what it’s costing you in compute. Sound familiar? Welcome to the reason LLM gateways exist. This is Part 1 of a two-part Field Notes series. Here we cover what an LLM gateway is, why you’d want one, and how to get Bifrost running locally on your Mac against Ollama with qwen3.5 — fully offline, fully free, fully yours. ...