What if managing your bookmarks felt like talking to a colleague? Not clicking through menus, not filling out forms, not dragging items between folders. Just saying what you want done. That's what prompt-based bookmark management looks like. LinkaGoGo connects to AI assistants like Claude, ChatGPT, and Cursor through the Model Context Protocol (MCP) — giving your AI direct access to your bookmark
Originally published at curatedmcp.com/blog/week-2026-19 The MCP marketplace continues to expand with four new servers landing this week, reflecting a clear market demand for data accessibility and content safety. Week 19 brings tools spanning SEO analytics, conversational UI, marketing data aggregation, and threat detection—a diverse mix that signals the ecosystem is moving beyond developer-centr
Why Figma MCP Isn’t Enough Why Figma MCP Alone Can’t Guarantee Production-Ready UI — and What Product Teams Must Do Instead Extraordinary results require an extraordinary team. I’m surrounded by people who treat design and development like a mission. They are warriors in the tech trenches, and this win belongs to them. No fluff. No filler. Just the facts on how we shattered our veloci
If you've watched a junior pen-tester spend a Monday morning typing the same amass enum -passive -d $TARGET, subfinder -d $TARGET -silent, pipe to httpx, pipe to naabu, feed surviving hosts into nuclei, dump JSON The work isn't hard. The glue is. Every team I've talked to has rebuilt this This post is about a different shape of the problem: what happens when you MCP tools that an AI agent can call
I spent the long weekend pushing Logic Apps MCP server capabilities further than I had before — and hit two bugs worth documenting. Both are filed. If you're building in this space, save yourself the debugging time. If you've been following along, the MCP server and BODMAS Agent are covered in the previous posts. This post is just about what broke when I wired them together. The Agent Loop fails w
Claude + Mobile via MCP: Giving the Model Hands on a Real Phone I plugged in a Pixel two months ago, ran one command in Claude Desktop, and watched it open Maps and start navigation to my home address from a single sentence prompt. It was the first time I'd ever seen a language model physically operate a phone. Latency was about two seconds per action; the part that surprised me was the third st
AI-Native Mobile Testing: What It Actually Means in 2026 The phrase "AI-native" has been thrown around in the testing space since 2019. Almost every tool calling itself that just bolts a language model on top of Appium and ships the same brittle XPath selectors with a new label. That's not AI-native testing. That's Appium with a chatbot. This post is about what AI-native actually has to mean to
The Missing Control Plane for Local AI Agents I sat with my Pixel for 20 minutes trying to get Claude Desktop to dictate a Slack message via accessibility. It was miserable. The model was capable. The transport wasn't. That gap — between an AI that can reason and an AI that can actually do — is what I've been working on with Drengr. This post is the version of the argument I'd give to anyone bui