Tobias Mann and Thomas Claburn
Our Take: I’m not necessarily encouraging you to rolll your own here, but it is worth being aware of this business model change – and the fact that from the get-go the definition of a token as a metric has been less than clear and open.
Your Takeaway: If you’re reliant on third party LLMs, remember to account for the risk of them changing their measurement metrics and charging – it’s been on the cards for a while.
Usage-based pricing for AI coding tools is becoming expensive and restrictive. This article shows how to run local AI coding agents like Claude Code, Pi Coding Agent, and Cline to avoid those costs. Local models work well for small projects but may need human approval to avoid mistakes.
Highlights
Over the past few weeks, we’ve seen Anthropic toy with dropping Claude Code from its most affordable plans while Microsoft has skipped testing the waters and moved GitHub Copilot to a purely usage-based model. The whole debacle got us thinking. Do we even need Anthropic or OpenAI’s top models, or can we get away with a smaller local model? Sure, it might be slower, less capable, and a little more frustrating to work with, but you can’t beat the price of free… Well, assuming you’ve already got the hardware that is.
