From 54344fc75e5929fd00cf7ebc4a085865bbe8631b Mon Sep 17 00:00:00 2001 From: Firehose Bot Date: Wed, 29 Apr 2026 16:02:08 +0100 Subject: [PATCH] now really the spelling changes --- .../2026/04-24-smaller-open-llms-now-work-for-open-agents.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/app/priv/blog/engineering/2026/04-24-smaller-open-llms-now-work-for-open-agents.md b/app/priv/blog/engineering/2026/04-24-smaller-open-llms-now-work-for-open-agents.md index a230adf..044d102 100644 --- a/app/priv/blog/engineering/2026/04-24-smaller-open-llms-now-work-for-open-agents.md +++ b/app/priv/blog/engineering/2026/04-24-smaller-open-llms-now-work-for-open-agents.md @@ -25,10 +25,10 @@ I have been using Claude Code for about a year. I noticed last week I started ta That also gave me some anxiety. What if my account got pulled for whatever reason? It was clear from the start of the monthly subscriptions (to me and many others. not everyone apparently) that this was not sustainable, and that at some point they would have to raise prices. Anthropic is now putting more and more of their stuff behind per-token pricing. And their models are expensive. -If you have to prompt precisily, you can just as well use a smaller model +If you have to prompt precisely, you can just as well use a smaller model --- -At the same time, Opus 4.7 requires more explicit prompting. Probably to compete with OpenAIs models. "Creativity" is nice when you are doing an architectural spike, but less so when you try to do some precise work in a larger codebase. Small open weight models (at least until last week) also require(d?) more precise prompting. Why would you pay for an expensive frontier model + closed source harness when you have to be precise in your prompts?You might as well do the same with a local model. +At the same time, Opus 4.7 requires more explicit prompting. Probably to compete with OpenAIs models. "Creativity" is nice when you are doing an architectural spike, but less so when you try to do some precise work in a larger codebase. Small open weight models (at least until last week) also require(d?) more precise prompting. Why would you pay for an expensive frontier model + closed source harness when you have to be precise in your prompts? You might as well do the same with a local model. I ran a small experiment last month with a model that is a lot smaller than what I usually use (4Billion and 9B parameters) and ran an auto-improvement loop on the prompt. If you know what the outcome should be, A larger model can iteratively generate a more detailed prompt that works for smaller models. Since everything runs on my own laptops, and I have cheap green energy at night, I can use this to become more independent of frontier labs, and help others. Note that I did that once, because the larger models now need much less RAM than they did a month ago. So I can keep one running in the background and whenever I have a question or something do build I will just prompt it from my writing environment (editor) or my coding agent.