Claude + Mobile via MCP: Giving the Model Hands on a Real Phone I plugged in a Pixel two months ago, ran one command in Claude Desktop, and watched it open Maps and start navigation to my home address from a single sentence prompt. It was the first time I'd ever seen a language model physically operate a phone. Latency was about two seconds per action; the part that surprised me was the third st
AI-Native Mobile Testing: What It Actually Means in 2026 The phrase "AI-native" has been thrown around in the testing space since 2019. Almost every tool calling itself that just bolts a language model on top of Appium and ships the same brittle XPath selectors with a new label. That's not AI-native testing. That's Appium with a chatbot. This post is about what AI-native actually has to mean to
The Missing Control Plane for Local AI Agents I sat with my Pixel for 20 minutes trying to get Claude Desktop to dictate a Slack message via accessibility. It was miserable. The model was capable. The transport wasn't. That gap — between an AI that can reason and an AI that can actually do — is what I've been working on with Drengr. This post is the version of the argument I'd give to anyone bui
In Q1 2026, our team audited 14 2FA libraries for Next.js 15 and found that migrating from Google Authenticator’s legacy TOTP implementation to Speakeasy 2 reduced average 2FA setup time per user from 42 seconds to 21 seconds — a 50% reduction verified across 12,000 production user onboarding flows. ⭐ vercel/next.js — 139,252 stars, 30,994 forks 📦 next — 155,273,313 downloads last month Data