I was using Sudolang to craft prompts, and having the AI modify my prompts. The more it modified them, the more they looked like math equations to me. I decided to skip to math equations directly and tried about 200 different constants and equations in my tests to come up with that 3 line prompt. There are many variations on it. Details in my git repository.
Hmm. The explanation seems short enough to have written by hand easily. But I suppose that the natural style of AI output has the upside that it demonstrates the Markdown rendering well.
I connect local models to MCPs with LM Studio and I'm blown away at how good they are. But the issues creep up when you hit longer context like you said.
OpenAI and Anthropic's real moat is hardware. For local LLMs, context length and hardware performance are the limiting factors. Qwen3 4B with a 32,768 context window is great. Until it begins filling up and performance drops quickly.
I use local models when possible. MCPs work well, but their large context injection makes switching to an online provider the no-brainer.
reply