The setup I was building an AI article generator. Four phases: Research — pull live SERP data, extract entities, score competitors Brief — Claude Sonnet writes a structured brief from the SERP Draft — section-by-section drafting in a fixed persona voice Polish — adversarial review pass that flags AI tells End-to-end: 6–8 minutes. Each phase: 60–180 seconds. Multiple LLM calls per phase. Token st
I needed to coordinate background scripts running across different machines. The obvious answer was Redis. Everyone uses Redis for this. The tutorials all use Redis. The Stack Overflow answers all say "just use Redis." So I looked at what deploying Redis would actually cost me: A running Redis server I had to maintain A broker to connect workers to it Celery or RQ on top of that Memory-based stora