Context Engineering Is Architecture: 650 Files Later
Context engineering for AI agents across a 650-file, seven-layer hierarchy. Three production failures, real token budgets, and the system that survived.
AI & TechnologyThoughts on design, development, AI infrastructure, and building products.
Context engineering for AI agents across a 650-file, seven-layer hierarchy. Three production failures, real token budgets, and the system that survived.
AI & TechnologyMulti-agent deliberation catches failures that single-agent systems miss. Here is the architecture, the dead ends, and what is actually worth building.
AI & TechnologyMy Claude Code agent inherited every sloppy human habit at machine speed. I built 3 philosophies, 150+ quality gates, and 95 hooks. Here's what worked.
AI & TechnologyMost codebases slow down as they grow. Mine accelerates. 95 hooks, 44 skills, and 14 configs make each feature cheaper than the last.
AI & TechnologyFive prompt failures taught me that structured observation beats clever wording. Boyd's OODA loop maps directly to my daily prompt engineering workflow.
AI & TechnologyI built an autonomous agent system with stop hooks, spawn budgets, and filesystem memory. Here are the failures and what actually ships code.
AI & TechnologyMcKinsey found 90% of companies claim AI adoption but only 23% scale beyond pilots. I've witnessed three flavors of AI theater and practiced one myself.
AI & TechnologyI use Claude Code daily with 86 hooks and a full quality gate system. Here's where I vibe code, where I engineer, and why the boundary matters.
AI & TechnologyTranslating my site into 6 languages revealed that Korean costs 2.8x more tokens than English for identical content. An interactive visualizer shows why.
AI & TechnologyI built a 3,500-line agent system with 86 hooks and consensus validation. Here's what I learned about RAG, fine-tuning, and agent orchestration.
AI & TechnologyHow a personal portfolio site went from a 76 mobile Lighthouse performance score with 0.493 CLS to a perfect 100/100/100/100 across all categories.
AI & TechnologyTechnical writing at Introl
Comprehensive hardware recommendations and cost analysis for running large language models locally.
GPU selection guide comparing NVIDIA's latest datacenter accelerators for different AI workloads.
Deep technical dive into Google's Tensor Processing Unit evolution from TPUv1 to TPUv5.
Resource sharing strategies for GPU clusters in containerized environments.
Guide to building and managing distributed AI computing with Ray framework.
Analysis of open source LLM economics and DeepSeek's competitive positioning.
Future datacenter power requirements and NVIDIA's next-generation GPU roadmap.
Small modular reactor solutions for powering next-generation AI infrastructure.
Technical analysis of DeepSeek's Multi-Head Compression architecture innovations.