Agentic Coding Is Not a Trap: I Answered the Viral HN Post With My Own Production Logs I made the exact mistake that viral post criticizes: I gave an agent an ambiguous task and went to make coffee. Came back 40 minutes later to 23 modified files, three broken tests, and a refactor nobody asked for. I'm not telling this to complain — I'm telling it because that day I started keeping logs of my a
PostgreSQL Query Rewriting Techniques The previous articles in this series covered performance problems you fix by adding indexes, restructuring joins, or tuning memory. This one is about the queries where the plan is "fine" — every node is doing something reasonable — but the query itself is asking the wrong question, producing unnecessarily large intermediate results or forcing the planner dow
Disclosure: I'm a senior backend tech lead and I run HostingGuru, where Telegram alerts ship as a built-in feature. This tutorial works on any platform — it's the manual version of what HostingGuru does for you. Useful even if you never become a customer. There's a hierarchy of where production alerts go, ranked by how likely you are to actually see them. Email → 14% open rate within an hour, less
This section is the map for the rest of the book. The five stages introduced in the 1.1 chapter overview (parse, analyze/rewrite, plan, portal, execute) are traced here through the actual code: which functions implement each stage, and in what order they get called. The mechanics of each of the five stages are unpacked in later chapters. Here, only the skeleton matters: how a backend starts up, ho
PostgreSQL Internals · Chapter 1 Query Processing Suppose a client sends SELECT * FROM users WHERE id = 1. The path that single line travels before coming back as a result row is longer than you might expect. Inside the PostgreSQL backend, that SQL goes through a five-stage pipeline. Backend entry and dispatch. The backend receives the message from the client and decides which processing path it s
很多团队的网络监控并不算差。 链路可用率有、接口带宽有、CPU 和内存有、异常告警也接进了企业微信、飞书和短信。但真正出了事,复盘时还是会出现同一句话:当时知道出问题了,但没有把现场留住。 这就是为什么越来越多团队开始关注网络回溯分析系统。 它解决的不是“能不能看到告警”这个初级问题,而是更关键的两个问题: 告警发生时,能不能快速还原到底是哪一段流量、哪一条路径、哪一种会话出了问题 事故结束后,能不能基于证据复盘,而不是靠聊天记录和印象拼凑过程 对云上和混合云场景来说,这件事尤其重要。因为链路更长、设备更多、路径更动态,很多故障不是“持续坏”,而是短时抖动、瞬时拥塞、路径切换、策略误命中。如果没有回溯能力,排障就很容易沦为赛后猜谜。 这篇文章不讲空洞概念,直接从一线运维视角拆清楚:云上网络回溯分析系统到底该怎么建,应该覆盖哪些能力,落地时最容易踩哪些坑。 先说结论: 传统监控擅长发现“异常
If you manage a remote team of 10+ people, laptop battery monitoring is one of those quiet problems you only notice when it's too late: a dev's MacBook dies on a client call, a sales rep's Dell shuts down mid-demo, or you suddenly need to replace 8 laptops in the same quarter because nobody saw it coming. This guide walks through how to track laptop battery health across a remote team — the metric
The Problem You install OpenClaw, configure it, and let it run in the background. But how do you actually know it's working? There's no built-in status page. No heartbeat alerts. No way to see if it's processing tasks or just sitting idle. I built a simple, self-hostable monitoring dashboard for OpenClaw agents: 🔗 OpenClaw Monitor on GitHub Tech Stack: Frontend: Vue 3 (Composition API) + Elemen