It started at midnight I had 24 hours, a free Replit subscription, and an idea: what if I could build something like Miro — but actually understand every line of code in it? The core problem I had to solve first Multiplayer sync sounds simple until you actually build it. The hard part isn't sending a canvas update — it's figuring out what to send. canvas.on('object:modified', (e) => { socket.emi
The previous two posts covered how events flow from the SDK to the UI. This post focuses on visualizing one specific type of event: tool calls. Tool invocations are the most frequent operations in an Agent application. A typical task might call tools twenty or thirty times—reading files, writing files, executing commands, searching code. If every tool call renders as the same gray block, it's hard
Post 1 covered how AgentBridge converts the SDK's AsyncStream<SDKMessage> into [AgentEvent]. This post looks at what [AgentEvent] becomes — how TimelineView renders 18 event types, handles scroll behavior, and stays smooth when the event count gets large. TimelineView is the main body of the workspace, filling all the space between the sidebar and the input box. Its view hierarchy is shallow: Time
I write a lot of READMEs. I ship faster than I document. I work with AI agents that write code in seconds and READMEs in minutes, and somewhere between the first commit and the third refactor, the README I wrote on Tuesday stops matching the code I wrote on Friday. The install command says npm start. The package.json defines start:prod. Anyone copying that command would have failed instantly. I'd