> I spent around $400 on o3 last month because I was banging my head against a wall with some really tricky code. When I hit truly difficult problems, throwing o3 at them for a few hours beats getting stuck in debugging rabbit holes for days.
I wish I could get some examples of this. I remember asking seniors for help on problems and they could help in minutes what would take me hours. And, likewise, I’ve had people ask me for a minute of help to stop them being blocked for days.
It surprised me that people would block for days and I wonder what that looked like. Do they just stare at the screen, walk around, play candy crush?
I’m trying to figure out if the author is a junior, average, or senior programmer and what they get stuck on would be something that a good programmer learns how to debug around. So you can finally get a good ROI on being a good programmer by comparing the time to what an AI would cost.
Yea I’d love to know what these look like. I’ve been using Claude and Claude code a bit for things I’m too lazy to do myself but know it’ll be solid at, for example a parser for iso dates in Janet. I set a $10 budget in February and have used $4 of it. I genuinely don’t know how to use these tools to spend $20 a month but at the same time they've been very helpful.
I similarly get confused about normal blocking. I genuinely don’t understand it. I get procrastination. I go for runs during work. But rarely is something genuinely so much that it’ll take hours of programming.
Edit: of course some tasks take multiple days. But visible progress is almost always achieved on the scale of hours not days.
Tried Claude yesterday to help me extract rows from a financial statement PDf. Let's automate boring stuff !!
After multiple failures , I did it myself
I sat down at a Piano yesterday and tried to play a beautiful song even though I never really used a piano before. Sounded horrible, must be the pianos fault.
One reason to use LLMs is to understand when and how to use them properly. They're like an instrument you have to learn. Once you do that, they're like a bicycle for your mind.
Currently my biggest gripe is asking AI to make one modification to file, it then decides to do 4 other "extra" changes I didn't ask for. AI please stay focused!
As with any other tool, half the art of using LLMs productively is the ability to recognize when they're a good fit and when they're not. So long as AGI doesn't come around, there will always be cases where they fall on their face and as far as I'm concerned, we're pretty far from AGI.
Most people don't realize that LLMs by design were not made for document processing, data extraction etc. For that, you would have to use a dedicated tool like Klippa DocHorizon, which built its own AI OCR from scratch. It also provides an API that you can use to send your documents and receive formatted data. It's less popular than, say, Textract or Tesseract, but it's far more accurate, especially if you're dealing with sensitive data that you wouldn't want an LLM to hallucinate.
> Let’s be conservative and say $800/day (though I’d assume many of you charge more). The AI subscription math is a no-brainer. One afternoon saved per month = $200 in billable time. Claude Max pays for itself in 5 saved hours. Cursor pays for itself in 45 minutes.
Is the argument that by using these AI subscriptions, you have free time that didn't have before and now you work less hours? Or that the extra productivity you get from that AI subscription allows you to charge more per hour? or maybe that you can do more projects simultaneously, and therefore get more $ per day?
Otherwise I don't get how the AI subscription "pays for itself".
If a task that normally takes X hours without AI now takes Y hours with AI and you charge $Z/hr, then the value of AI is X - Y * Z for this single task. If the value across all tasks is greater than the amount you are paying for AI, then it has “paid for itself”.
I've never done contractor or $/hr work and I have no idea how these things work in reality, but:
If a task takes you five hours to do without AI and 1 hour with AI charged at 100$, in the without-AI case you're making 500$, in the with-AI case you're making 100$ - price_of_AI, right?
Otherwise your example assumes you're charging someone 5 hours of work when in reality it took you 1 hour and then you spent an additional 4 hours watching TV.
In any case this thinking exercise made me realize that maybe it's more about staying competitive against other peers than about "AI paying for itself". If you're really charging for hours of work, then it is really a competitive advantage against people not using AI.
Assuming an AI-enhanced contractor can do the same amount of work than a non-AI-enhanced contractor in fewer hours, then I'd assume they would get more contracts, because the overall project is cheaper for whoever is hiring them. Does that really lead to you making more money, though? No idea honestly. Probably not? I just can't see how using AI "pays for itself". At best you're making now less money than before, because you're required to pay for the AI subscription if you want to stay competitive.
It sort of seems like the author is billing for the time saved and having a coffee or something. Seems a bit off.
Paying programmers by the hour always seemed weird for me because of the productivity spikes. I prefer paying by the job or just buying a weekly block that assumes someone works full time on my project.
I would be annoyed with a consultant who billed me for 4 hours when it was 30 minutes of work and an estimated savings from using AI.
I suggest to try Claude Code out and see for yourself. Solving problems is now cost fraction of the time. What took hours/days before, now might take MINUTES., it's insane.
That's why anyone suggesting AI won't take your jobs is delusional. We will need significantly less developer time.
Or, as has typically been the case in IT, the pie will get bigger.
Maybe I'll program my oven to cook the pizza at a slightly lower temperature, but crisp the living carbon out of it for the final 90 seconds. Just like I'm doing manually now, but less prone to errors because I ignored the pizza while browsing HN and can smell it bur.......
I think developers will just do more. Already in a team of 10 programmers there’s a few who don’t do very much. I think they will be gone. And a 1x programmer becomes a 10x programmer using AI.
I think there will be more work for developers, but it will be harder to slack off.
Or get a 3090 24 GB off eBay for 700€ and run mistral, devstral, qwen3 and others indefinitely.
I'm a big fan of two-tier systems where the 90% case is taken on by something cheap, and renting the premium tier. Works for cars, server hosting, dining ...
I did the math, it just doesn't check out, especially if you take into account time to generate and electricity.
Even assuming you generate 100 tok/second, 24 hr, that's only 8 mil tokens. About $5 for Gemini 2.5 Flash or $50 for 2.5 Pro. But you will not be able to run it 24/7 locally - sleep, thinking between prompts, etc
Even at continuous usage (24 hr/day), it's still cheaper to just pay Google for Gemini 2.5 Flash tokens for half a year.
And this is NOT taking into account that mistral/devstral/qwen3 are vastly inferior models.
How can Google be so cheap? TPUs, batching multiple requests, economy of scale, loss-leader, ...
Running stuff locally makes sense for fun, as backup for loss of connectivity, if you want to try uncensored models, but it doesn't make sense economically.
Would be cheaper in the long run maybe but heat, noise, electricity cost and the fact that the models are not as good as sonnet 4 and opus 4 are some things to consider.
I tried running those locally via Ollama on my 3080 (latency doesn't matter) and the performance is abysmal. They're almost random token generators, literally unusable. My hallucination rate was near 100% for basic tasks like data extraction from markdown.
This is probably the easiest problem to solve, isn't it? Most tools offer a n API endpoint setup that you can use locally and also expose on a local network easily.
I haven't paid for the nicer models — now wondering if you could use less capable models to get some code up and running, then pay for better models just to "polish up" the code, get it to "final".
These articles seem to justify spending money without considering alternatives It's like saying "cold meds let me get back to work, I am well justified paying $100/day for NyQuil".
I would appreciate a less "just take my money" and more "here are features various tools offer for particular price, I chose x over y cos z". Would sound more informed.
Would also like to see a reason on not using open source tools and locking yourself out of various further ai-integration opportunities because $200/mo service doesn't support em.
Regarding the o3 "trick", if I understand it correctly, I'm trying to do something similar with the "MCP Super Assistant" paired with the DesktopCommander MCP.
I still haven't managed to make the mcp proxy server reliable enough, but looks promising. If it works, the model would have pretty direct access to the codebase, although any tool call is requires a new chat box.
I guess aider in copy-paste mode would be another solution for cheapskates like myself (not a dev and I barely do any programming, but I like to tinker).
I don’t want to pay to use LLM tools, just like I don’t pay for Linux, Git, or Postgres. Sadly, open source LLMs are way behind private LLMs (whereas Linux/Git/Postgres are top notch software better than private ones)
No they're not. I prefer closed source LLMs because I'm not interested in having a small hardware park, but DeepSeek 0528 or Qwen 3 235 are really good models. Not as good as Claude 4, but absolutely good enough for a lot of development use cases. You just have to spend money on the right hardware.
I wish I could get some examples of this. I remember asking seniors for help on problems and they could help in minutes what would take me hours. And, likewise, I’ve had people ask me for a minute of help to stop them being blocked for days.
It surprised me that people would block for days and I wonder what that looked like. Do they just stare at the screen, walk around, play candy crush?
I’m trying to figure out if the author is a junior, average, or senior programmer and what they get stuck on would be something that a good programmer learns how to debug around. So you can finally get a good ROI on being a good programmer by comparing the time to what an AI would cost.
I similarly get confused about normal blocking. I genuinely don’t understand it. I get procrastination. I go for runs during work. But rarely is something genuinely so much that it’ll take hours of programming.
Edit: of course some tasks take multiple days. But visible progress is almost always achieved on the scale of hours not days.
One reason to use LLMs is to understand when and how to use them properly. They're like an instrument you have to learn. Once you do that, they're like a bicycle for your mind.
Here ya go, fixed your analogy
Deleted Comment
If you want it to be human readable you can use iXBRL which is embedded in HTML.
Is the argument that by using these AI subscriptions, you have free time that didn't have before and now you work less hours? Or that the extra productivity you get from that AI subscription allows you to charge more per hour? or maybe that you can do more projects simultaneously, and therefore get more $ per day?
Otherwise I don't get how the AI subscription "pays for itself".
If a task takes you five hours to do without AI and 1 hour with AI charged at 100$, in the without-AI case you're making 500$, in the with-AI case you're making 100$ - price_of_AI, right?
Otherwise your example assumes you're charging someone 5 hours of work when in reality it took you 1 hour and then you spent an additional 4 hours watching TV.
In any case this thinking exercise made me realize that maybe it's more about staying competitive against other peers than about "AI paying for itself". If you're really charging for hours of work, then it is really a competitive advantage against people not using AI.
Assuming an AI-enhanced contractor can do the same amount of work than a non-AI-enhanced contractor in fewer hours, then I'd assume they would get more contracts, because the overall project is cheaper for whoever is hiring them. Does that really lead to you making more money, though? No idea honestly. Probably not? I just can't see how using AI "pays for itself". At best you're making now less money than before, because you're required to pay for the AI subscription if you want to stay competitive.
Paying programmers by the hour always seemed weird for me because of the productivity spikes. I prefer paying by the job or just buying a weekly block that assumes someone works full time on my project.
I would be annoyed with a consultant who billed me for 4 hours when it was 30 minutes of work and an estimated savings from using AI.
That's why anyone suggesting AI won't take your jobs is delusional. We will need significantly less developer time.
Maybe I'll program my oven to cook the pizza at a slightly lower temperature, but crisp the living carbon out of it for the final 90 seconds. Just like I'm doing manually now, but less prone to errors because I ignored the pizza while browsing HN and can smell it bur.......
My pizza!
I think there will be more work for developers, but it will be harder to slack off.
What I hate is that I will not pay more than 20-50$ per month for side projects and that gap annoys me.
I'm a big fan of two-tier systems where the 90% case is taken on by something cheap, and renting the premium tier. Works for cars, server hosting, dining ...
Even assuming you generate 100 tok/second, 24 hr, that's only 8 mil tokens. About $5 for Gemini 2.5 Flash or $50 for 2.5 Pro. But you will not be able to run it 24/7 locally - sleep, thinking between prompts, etc
Even at continuous usage (24 hr/day), it's still cheaper to just pay Google for Gemini 2.5 Flash tokens for half a year.
And this is NOT taking into account that mistral/devstral/qwen3 are vastly inferior models.
How can Google be so cheap? TPUs, batching multiple requests, economy of scale, loss-leader, ...
Running stuff locally makes sense for fun, as backup for loss of connectivity, if you want to try uncensored models, but it doesn't make sense economically.
For investigating concepts, can AI for this at all? Subscriptions will be a better approach.
Local models are underutilized in comparison to the promise they have. Imagine browser-agnostic universal AdBlock and that's just the surface.
If I worked on a codebase I couldn’t send to the hosted tools I’d go this route in a heartbeat.
I think if you’re comparing it to Claude max it’s ok, but the payback period on GitHub copilot for example is almost 7 years.
This is probably the easiest problem to solve, isn't it? Most tools offer a n API endpoint setup that you can use locally and also expose on a local network easily.
Deleted Comment
I would appreciate a less "just take my money" and more "here are features various tools offer for particular price, I chose x over y cos z". Would sound more informed.
Would also like to see a reason on not using open source tools and locking yourself out of various further ai-integration opportunities because $200/mo service doesn't support em.
I still haven't managed to make the mcp proxy server reliable enough, but looks promising. If it works, the model would have pretty direct access to the codebase, although any tool call is requires a new chat box.
I guess aider in copy-paste mode would be another solution for cheapskates like myself (not a dev and I barely do any programming, but I like to tinker).