A College Project That Planted a Seed Years ago I was on a university team trying to build a Go AI. We explored monte carlo simulation for lookahead search, basic neural networks for pattern recognition, and expert systems for encoding domain knowledge. None of them worked well enough on their own. Go's branching factor is enormous, so brute-force search fails quickly. Neural networks without th
If you’ve been building with AI recently, you’ve probably seen these terms everywhere: AI Gateway. And depending on where you read, they either sound like the same thing… or completely different systems. Some vendors use them interchangeably. Others define only one and ignore the rest. And if you try to piece it together yourself, you end up with a vague understanding that doesn’t really help when
The first article on this blog explained how it was built in 30 minutes with Claude Code. Naturally, a blog needs comments. Same constraints: no database, no external dependencies, no Disqus tracking visitors. Just PHP + JSON files. Built in one session with Claude Code — the interesting part wasn't the code, it was the security audit that followed. A comment system without a database seems trivia
When building applications with large language models (LLMs), one of the most overlooked costs is how structured data is represented. Most systems use JSON. And JSON is inefficient for LLM input. KODA (Knowledge-Oriented Data Abstraction) is a schema-first data format designed to reduce token usage when sending structured data to LLMs. It works by: Defining structure once (schema-first) Encoding v
In an era where data privacy is often the price we pay for convenience, medical information remains the most sensitive frontier. When you upload a patient's transcript or a personal health log to a centralized API, you're essentially trusting a third party with your most intimate data. But what if the "brain" lived entirely within your browser? Today, we are diving deep into the world of Edge AI a
Self-attention already helps a transformer understand relationships between words using Query, Key, and Value. But there’s a problem. One attention mechanism usually ends up focusing on a limited kind of relationship at a time. Language doesn’t work like that. A sentence can have structure, meaning, and long-range links all at once. That’s why transformers use multi-head attention. Instead of doin
I keep seeing the same argument about AI making us dumber. It's the same argument people had about search engines, and before that books. The usual response is to point at history and say "every generation panics, every generation was wrong, relax." I think that response is half right, and the wrong half is what bothers me. Tools change what we bother to remember. The people who'd trained their wh
We debate endlessly about whether AI will ever achieve consciousness, but we forget how consciousness actually compiled in the first place. It wasn’t spawned in a vacuum; it was forged by the brutal necessity of survival. For millions of iterations over millions of years, early cognition was nothing but pure instinct and bloodlust—refined only by the fight for the right to exist. Humanity is not