He mentions Max as another place that they didn't properly predict plan and pricing relative to usage. I'd bet the farm that it's the next to be 'A/B' tested.
I keep telling them and they still want to spend money on tokens at the Anthropic casino, even though they are egregiously price gouging and applying upper limits so you spend more on tokens.
Sometimes you can't help gamblers who want to gamble on tokens to hit the jackpot on fixing a typical issue which can be done by local models or even reading the documentation.
I’m sorry but that’s just dumb. An LLM is a tool. Your brain is not a substitute for an LLM in the same way your fingers are not a substitute for a wrench.
The year is 2026 and if you are using your brain on chore work like one-off scripts, refactoring, boilerplate test code, then you are wasting time and money and I don’t want to work with you.
Local models are fine for this and can do it in a fraction of the time your brain will take to even get bootstrapped
The year is 2026 the average RAM for the most common type of developer’s (web) machine is 16GB. 8 will be the lower end. Tell me which model can one run on this machine locally?
He mentions Max as another place that they didn't properly predict plan and pricing relative to usage. I'd bet the farm that it's the next to be 'A/B' tested.
This should be a warning to those who feel that it's ok to offload your creativity to a subscription service. Always need a local model in some form.
You could judge the costs of the AI products you're using by the standard API pricing, not promotional subscription offers.
Not even that way, given that the price is still highly subsidized by investors and circular deals.
There is very little vendor lock. We can keep using subsidized model until it’s not. Then switch to next subsidized model.
It's like chairs!
Are there local models that are anywhere near as good at coding as opus 4.6?
People will insist otherwise, but I haven't seen anything close to sonnet 4.6 that can be run locally.
I don't think anyone can honestly say a huge frontier model is actually going to be matched by something running on 64gb locally?
I have read many comments saying Qwen3.5 various ~30B models, Gemma 4 ~30B models and now Qwen3.6 "better than sonnet".
I don't know how large sonnet and opus are but the rumor is 1T and 5T respectively.
I keep telling them and they still want to spend money on tokens at the Anthropic casino, even though they are egregiously price gouging and applying upper limits so you spend more on tokens.
Sometimes you can't help gamblers who want to gamble on tokens to hit the jackpot on fixing a typical issue which can be done by local models or even reading the documentation.
The 'local model' is called your brain.
I’m sorry but that’s just dumb. An LLM is a tool. Your brain is not a substitute for an LLM in the same way your fingers are not a substitute for a wrench.
The year is 2026 and if you are using your brain on chore work like one-off scripts, refactoring, boilerplate test code, then you are wasting time and money and I don’t want to work with you.
Local models are fine for this and can do it in a fraction of the time your brain will take to even get bootstrapped
The year is 2026 the average RAM for the most common type of developer’s (web) machine is 16GB. 8 will be the lower end. Tell me which model can one run on this machine locally?
Dup: https://news.ycombinator.com/item?id=47854477
Maybe those already on $20 a month plans won't be nerfed much more?
It's yet another austerity move, pretty much in line with the recent ones.