Introduction I've been seeing more developers say that Codex has become easier to use, more cost-effective, or simply a better fit for some workflows than it used to be. This is not a "Claude Code is bad, everyone should switch" article. I still use Claude Code at work, and if cost were less of a factor in my personal setup, I would probably be using both more actively. If you're already comfort
Hello everyone! I wanted to write this article to share my experience with agentic coding without Claude and Codex, I started dabbling with agentic coding a few months ago when Claude had decent limits on the 20$ plan, You prompt the agent: I want e2e tests, and it will study the codebase and implement them. When I've started hitting limits on Claude code, and this is not a secret that they reduc
Book: TypeScript in Production Also by me: The TypeScript Library — the 5-book collection My project: Hermes IDE | GitHub — an IDE for developers who ship with Claude Code and other AI coding tools Me: xgabriel.com | GitHub You have seen the shape of this incident before. A 500 lands in production. The frontend says "checkout failed". The Hono service that owns /checkout called the prici
Every observability vendor has bolted "AI" to their landing page. Half of those features are genuine improvements. The other half are autocomplete in a costume. After a few years of running these tools across enterprise estates, here is where AI-augmented SRE actually pays off, where it doesn't, and what we'd advise teams adopting it today. The single most defensible use case. A medium-sized estat
Iris v0.4.0 ships today. It's the release where protocol-native eval crosses from "deterministic rules" into "semantic scoring" — without giving up any of what made the deterministic layer work. Three headline features plus a lot of infrastructure work that quietly compounds. I'll go through each, why it matters, and how it fits the thesis. Heuristic rules catch a lot: length, keyword overlap, PII