Loading...
Loading...
29 posts
I spent a session orchestrating three concurrent AI coding agents across four repos. They shipped 20+ PRs, wrote 100+ blog posts, removed an entire database dependency, and resolved merge conflicts. Here is what actually works.
How I built a full LittleBird clone with screen context reading, meeting recording, arena mode, and MCP tool support — from scratch to packaged .app in a single coding session.
Marcus built a memorial chatbot because staying close to loss felt safer than silence. The rest of us keep repeating the same prompt, hoping the ending ...
Five lines of Python and an API key produce a working demo. The gap between that demo and a production system contains failure modes the prototype...
Most enterprise AI transformations are solving problems that spreadsheets handle at 1/50th the cost. The misalignment is driven by career incentives,...
Most AI agent infrastructure is premature. The agents themselves barely work. The industry is selling Formula 1 equipment to people still learning to...
AI coding assistants output shell commands, not GUI instructions. That single fact is reversing a decade of developer tooling trends.
AI code reviewers moved from rules-based checking to reasoning-based analysis. The gap between what they catch and what humans catch is closing fast.
Async code generation turns development into specification and review. The coding happens in the background. This changes what it means to be a senior...
AI commoditized the pattern recognition and architectural intuition that made 10x developers valuable. The bottleneck moved from individual output to...
AI-generated code produces different bugs than human-written code. QA built for syntax checking is testing for the wrong failures.
I shipped a voice profile extractor at 60% accuracy. Simple pattern matching outperformed ML for writing voice replication.
Different LLMs have different strengths. Routing tasks to the right model -- like heterogeneous compute -- turns out to be more valuable than using one ...
I built a multi-AI content pipeline combining Gemini and Claude. The failures taught me more than the architecture.
I built a voice replication system by extracting patterns from my blog corpus. Here's what it captures, what it misses, and what that reveals about...
The best AI companies don't write prompts by hand. They generate them programmatically, test them systematically, and optimize them continuously.
Combining Semgrep, CodeQL, SonarQube, and Snyk gets you 44.7% vulnerability detection. Semantic SAST combines Tree-sitter with LLM reasoning to do better.
I added Gemini to a codebase that already had Claude embedded. The useful discovery was about shared context files, not model capabilities.
Every web design decision now must serve two audiences: humans who browse visually and AI agents that consume data programmatically. The architectural...
I asked Claude to analyze my writing style across my blog posts. The patterns it found -- and the ones I didn't know I had -- were genuinely surprising.
OCode: Why I Built My Own Claude Code (and Why You Might Too): A few nights ago, I opened my Anthropic invoice.
Autonomous agents need the same infrastructure primitives that web services got a decade ago: identity, policy, and secrets as first-class citizens.
A photographer friend posted a sunset photo after three hours of waiting for the perfect light. Within minutes: 'Obvious Midjourney.' 'Nice prompt, bro.'
AI doesn't make everyone equally skilled. It amplifies existing ability. That changes what technical interviews should test.
Deep dive into RAG architectures: chunking strategies, retrieval methods, embedding optimization, and production patterns with research-backed analysis.
Systematic experiments on temperature and top-p sampling parameters across 1000 real queries with empirical data on creativity, coherence, and...
OpenAI recently rolled back a GPT-4 update due to sycophantic behavior. The word itself--'sycophantic'--feels like a punchline from a _Black Mirror_...
Most AI products are designed to fail. Not because the technology is bad, but because product teams are building for the wrong expectations entirely.
I kept writing terrible JIRA titles during customer calls. So I built a Chrome extension to fix it.