Thoughts

Token taxes are just cloud computing's final power grab

The per-token pricing model is designed to keep you dependent, not to reflect actual compute costs.

Every API call you make is a small tax on your product’s future. The token pricing model sold as “pay for what you use” is actually designed to make local inference look impossibly complex while keeping you hooked on metered compute.

The rental trap dressed as convenience

Token pricing feels reasonable when you’re prototyping. A few pence per thousand tokens, what’s the harm? But scale up and you’re paying enterprise software prices for what amounts to matrix multiplication. The major cloud providers have learned from Adobe’s playbook. Make the alternative seem technically daunting while your monthly bill creeps up.

We’ve seen this pattern before. AWS made running your own servers feel antiquated until you realised you were spending more on their convenience than buying hardware. The same thing is happening with inference. Companies are discovering that a decent GPU cluster costs less per month than their OpenAI bill.

Local inference isn’t the future, it’s the present

The technical barriers are crumbling faster than anyone expected. NVIDIA’s RTX cards can run surprisingly capable models. Apple’s M-series chips handle inference workloads beautifully. Even edge devices are getting powerful enough for real-time processing.

The dirty secret is that most production workloads don’t need the absolute latest model. They need consistent, fast, private inference. Local deployment gives you all three without the recurring fees.

The token tax era is ending. Not because the technology got better, but because the economics finally make sense.

Related