The previous two posts covered how events flow from the SDK to the UI. This post focuses on visualizing one specific type of event: tool calls. Tool invocations are the most frequent operations in an Agent application. A typical task might call tools twenty or thirty times—reading files, writing files, executing commands, searching code. If every tool call renders as the same gray block, it's hard
Post 1 covered how AgentBridge converts the SDK's AsyncStream<SDKMessage> into [AgentEvent]. This post looks at what [AgentEvent] becomes — how TimelineView renders 18 event types, handles scroll behavior, and stays smooth when the event count gets large. TimelineView is the main body of the workspace, filling all the space between the sidebar and the input box. Its view hierarchy is shallow: Time
選定理由 Paper: https://arxiv.org/abs/2512.01020 【社会課題】 【データの設計と従来技術の限界】 Issue Tree(法的論点ツリー)に変換し、葉ノードに対しルーブリック基準を適用可能にした。原告・被告・裁判所の主張をツリー構造で整理した約24,000インスタンスのデータセットを構築。評価軸は「論点カバレッジ」と「正確さ」の2次元。以下がサンプルである: 【原告の主張】被告は540万円を支払え └─【原告】保険金の支払い義務がある ├─【原告】死亡は突発的・偶発的な事故だった │ └─【原告】餅を食べて窒息死=外因による傷害 │ └─【被告】死因は既往症の可能性が高い └─【裁判所の結論】突発的事故と認定 ただし窒息死は証明不十分 この
Introduction To understand knowledge graphs, you first need to grasp three core concepts: entities, relations, and triples. Imagine a knowledge graph as a network that models the real world using nodes and connections. In this network, an entity is any distinct thing or object such as a person, city, or company. For example, “Sreeni”, “Plano”, and “Caterpillar” are all entities. A relation descr
This is my Day 2 of learning AI fundamentals where I will be covering the following concepts: Vector Embeddings How Tokenisation and Vector Embeddings relate to each other Vector embeddings is the process of turning each token id(generated during tokenisation) into high dimensional vector where semantic similarity results into geometric closeness. Think of it like this: dog is closer to puppy, al