From “It Works” to “It Thinks”: Building a Real Agentic AI System at Home
Or: why context windows, KV cache, and humility matter more than bigger models. My last blog post ended on a high note: I had a functional, fully local LLM stack running in my homelab. Models were answering questions, Open-WebUI was humming along, and Ollama had made the early experimentation delightfully easy. Then I tried to…