Some semi-unhinged musings on where LLMs fit into my life—and how I'll keep using them.
See full article...
See full article...
Yup your definition was what I was looking for. Friends of mine who have vibe coded apps for just themselves reported that they can spend ~$100 for basic desktop apps, web scrapers, APIs, etc. I was curious to see how your costs compared to theirs.It fit within the $20/month I'm paying for Claude Pro, so the cost to me was $20. For other definitions/applications of "cost," I don't think data are available to do more than wild-ass-guess the numbers, which @LetterRip took a stab at doing right here.
Over the weekend, I used Claude Code to build a personal finance web app to replace my old system. It's a React app and Postgresql db and deployed with Docker. It took around 9 hours to build, split over Saturday and part of Sunday using Opus and the Design and Superpowers plugins. Cost was my Pro sub and $60 in Extra usage charges.Yup your definition was what I was looking for. Friends of mine who have vibe coded apps for just themselves reported that they can spend ~$100 for basic desktop apps, web scrapers, APIs, etc. I was curious to see how your costs compared to theirs.
LLMs can be fantastic if you’re using them to do something that you mostly understand. If you’re familiar enough with a problem space to understand the common approaches used to solve it, and you know the subject area well enough to spot the inevitable LLM hallucinations and confabulations, and you understand the task at hand well enough to steer the LLM away from dead-ends and to stop it from re-inventing the wheel, and you have the means to confirm the LLM’s output, then these tools are, frankly, kind of amazing.
Thanks for pointing that out, d'oh. Yes, I kept the mu-plugin that adds no-cache headers active. I looked at ways to tweak the Apple News plugin to alter its behavior, and also at poking deeper into Wordpress' guts to see if screwing around with how the post publication event works would be the right call, and in the end I decided to stick with what I know—and I know how http headers work. It seemed the safest, sanest way forward.
I appreciate the write up and found it very interesting!No, no one told me to write this. I've been a reader since 1998 and I generally find that my own personal interests align with the audience's, because I am them. Further, I've been employed here since 2012. I'm a senior editor with direct reports and I sit on the Ars editorial board. I am the "on high" at this point.
I had a solid experience over the Christmas break and I wanted to write it up. If I have further solid experiences, I'll write those up, too!
Cool your jets Vin Diesel. I'm referring to coding specific tools/models, not wider AI tools that are doing and integrating into more complex workflows. This article is about coding tools and I'm talking about coding tools.It's amazing how you've managed to learn about AI but avoided to hear any of the advertisement around AI tossed by Anthropic, OpenAI, Microsoft, etc. Maybe go check what those companies are saying AI can do. You might learn something new about the people "missing the point". I would love to go back to the time when LLMs were just decent tools for text analysis and the like. But we're not back then. We are here now.
I can't speak to your efforts personally but I continue to see the same tit for tats in almost every AI discussion thread, this one included. The interactions barely ever move to "this could be a practical policy" or "a law like this might prevent mass redundancy while still allowing AI to assist in X sector" or "AI can be useful in this instance but we need tax funded income safety nets for such and such." It's almost always "AI is an abomination" vs "AI is the technology of our time and you just need to suck it up and get used to it."To ask those questions…
I use the $100 a month Claude Max and almost never hit the limits. This is with just hours and hours of typing into terminal a day sometimes.I appreciate the write up and found it very interesting!
Question: I'd love to know more about the costs to run these code helper LLMs.
You mentioned additional costs being very common, and that you ran out of credits.
What would the cost per prompt or hour work out to be?
Would it cost $100 a day if you used this 6 or 8 hours a day? $800 a day?
Does it start to approach the cost of hiring an actual person?
Yeah like all those "extremists" who keep saying we have a massive climate crises, and between crypto currencies and LLMs, we've wiped out what little energy savings we've made for the last 30 years. Bringing that up is such a bummer and has no "nuance".But the comments are always filled with people who have the nuance of a Knight Templar. How is any productive conversation going to happen when there are so many extremists torpedo-ing any and all discussions about LLMs… genuinely flummoxed.
That was never a thing. Why do people keep acting as if the dotcom bubble was about the internet/web itself? It was simply called the dot com bubble. It was never an internet bubble. Very specific services/websites got overvalued, and a LOT of money was involved, but it was never about the whole internet, or even involving a majority of it.The dotcom bubble burst and a lot of idiots went broke, but "the internet" didn't go away. It really was an important new technology - it just wasn't the magical panacea that the startups were selling.
I remeber what a revolution GNU software was and how it managed to democratize computers with open source. Watching people throw that all way to chase the next shiny thing while absolutely demolishing what remains of the decentralized, open internet was one of the saddest parts of 2025. I got a notification today on a different forum that someone had returned after a multi-year absence and had just noticed one of my open source projects. Of course I was greeted by that Cloudflare time waster because the forum is buckling under the strain of AI bots. And for what? So you don't have to think? So you can buy a subscription to avoid learning? Thanks.I use the $100 a month Claude Max and almost never hit the limits. This is with just hours and hours of typing into terminal a day sometimes.
Over the weekend, I used Claude Code to build a personal finance web app to replace my old system. It's a React app and Postgresql db and deployed with Docker. It took around 9 hours to build, split over Saturday and part of Sunday using Opus and the Design and Superpowers plugins. Cost was my Pro sub and $60 in Extra usage charges.
On the other hand, over the course of about 2 weeks, I built a Sonos dashboard web app that can track the SiriusXM songs that play and keeps statistics, favorites, and lets me update alarms. The web app is only about 30% as complex as the finance app but I've also built an MCP server for the Sonos app and use it as a PWA. This one was built up over time and I never used more than the 5 hour quota limit so cost was limited to the $20 sub.
The cost differs from tool to tool, provider to provider. MS GitHub Copilot is considered to be the cheapest option. In enterprise version, it gives a user 300 premium requests and unlimited number of regular requests (smaller and older LLMs) per month for $18. For the best LLM (Claude Opus 4.5) they charge 3x. It's impossible to have a simple comparison of the costs between using a human and LLM. It depends on the human, LLM, project type etc. But there is no doubt that LLM gives the human a huge boost. And, the more proficient (in coding and domain expertise) the human is, the greater leverage LLM provides. Keep in mind that with proper use, the best LLMs can produce more code in 6 hours than human would in a month (or more).I appreciate the write up and found it very interesting!
Question: I'd love to know more about the costs to run these code helper LLMs.
You mentioned additional costs being very common, and that you ran out of credits.
What would the cost per prompt or hour work out to be?
Would it cost $100 a day if you used this 6 or 8 hours a day? $800 a day?
Does it start to approach the cost of hiring an actual person?
I'm curious how you think fail2ban would have helped in this specific situation—are you suggesting i have f2b watching https traffic?You should have something like fail2ban on any hosting server so any rogue bot doesn't crash your system. This is amateur stuff.
Noted above, but while there are options to pay for extra usage (or to do straight API-based billing), everything I've ever done with Claude Code (including this project) was done with a regular $20/month pro subscription. I've never paid for overages or instant credit refills; sometimes, for projects burning a lot of tokens, this means I have to wait a few hours for the token count to reset.I appreciate the write up and found it very interesting!
Question: I'd love to know more about the costs to run these code helper LLMs.
You mentioned additional costs being very common, and that you ran out of credits.
What would the cost per prompt or hour work out to be?
Would it cost $100 a day if you used this 6 or 8 hours a day? $800 a day?
Does it start to approach the cost of hiring an actual person?
You just make a rule for 403. After three fails it bans that source for 5 minutes.I'm curious how you think fail2ban would have helped in this specific situation—are you suggesting i have f2b watching https traffic?
IME, after extensive on-and-off usage over years, fail2ban is a terrible tool for most applications. It's gross and heavy and bloated, and the amount of CPU time it steals from the host doing live matching against your logs grows unacceptably large as the traffic scales. For SCW, fail2ban just watching ssh traffic racks up more cpu time than my redis and mariadb processes combined. If you think fail2ban is a good idea versus better tools and methods, you should re-examine your requirements and what you think it's accomplishing for you. Perhaps you fall into the extremely narrow band where the tool gives you some real utility? If not, repeating context-free advice like "lol just use fail2ban" is cargo-cult system administration, and that's amateur stuff.
I'd use plain ol' nftables rate limiting for any given port or service before I reached for fail2ban.
I'm happy to discuss my sites' security postures & strategies in detail if you'd like, and answer any questions you have. That way, you won't have to make silly assumptions and maybe we'll all learn something!![]()
Again, why would I want a whole-ass extra application to screw with my layer 7 traffic like that, when nginx is perfectly fine generating 403s without all the extra logging and tracking? (Edited to add: the "but what if a rogue host takes down my web server by generating 403s?" concern is, frankly, silly.)You just make a rule for 403. After three fails it bans that source for 5 minutes.
Edit: I just want to explain that you can't look at the application layer 403 with nfttables. You also don't need a database file, unless you want to keep long-term bans after a reboot, just don't use one at all. And you should be using pyinotify, so it doesn't have read your entire logs again, even if you are lazy and don't rotate your logs. I hope this helps.
Ahh, the original sinIt suggested Python for our log colorizer because of the language’s mature regex support—and to keep the code somewhat readable for poor, dumb me.
This is a bad argument. If I create art inspired by other works of art, it's a new thing. It's filtered through my human experience. I create things that are the sum of my experience and skills.The "stolen" stuff is a murky one and enormously charged with emotion. If you go to a bunch of museums to observe a heap of paintings, then read a stack of books, then create your own work inspired by these learnings is this stealing? You wouldn't really call it this. With AI, it's similar in the sense that the technology has trained on material. It hasn't stolen it from anyone. Arguably.
I remeber what a revolution GNU software was and how it managed to democratize computers with open source. Watching people throw that all way to chase the next shiny thing while absolutely demolishing what remains of the decentralized, open internet was one of the saddest parts of 2025.
I appreciate the write up and found it very interesting!
Question: I'd love to know more about the costs to run these code helper LLMs.
You mentioned additional costs being very common, and that you ran out of credits.
What would the cost per prompt or hour work out to be?
Would it cost $100 a day if you used this 6 or 8 hours a day? $800 a day?
Does it start to approach the cost of hiring an actual person?
Modern python has a built in library for ipaddresses - ipaddress
If I were to write your app I would have probably used existing libraries - apache-log-parser for the parsing, and rich for colorization and display, argparse for arguement parsing.
ipaddress.ip_address(val).version and it will return either 4 or 6 if it's a valid address and raise a ValueError if it isn't.I'll do you one better, I restarted the Copilot app several times today in the hopes of resolving UI issues that started creeping up (failure to render code blocks) and getting back to vibe coding.Come on, I'm a power-user using Windows. That's never, never, never ever happened to me.
And please don't ask me why I was repeatedly re-booting my system at 7 am this morning.
Google search has indeed been declining for a while before AI came along, but that resulted from a conscious choice by Google to worsen their product for the purposes of increased revenue/profit through enshitification.You must have been asleep for the past 10 years or so... The "decentralized, open Internet" was already dying long before LLMs appeared on the landscape.
BTW same with people who complain about AI ruining Google search: It was getting harder and harder to find anything but SEO fodder and webshops with Google since a long time. Real, decentralized content in good old home pages, blogs and forums was vaporizing since ages. What we got instead was social media filled with engagement bait. And this did NOT only happen in 2025.
Sounds like you might be doing something wrong there. Are you setting enough context when you load the model? You certainly shouldn't need to use a RAG solution for what you're describing, unless you're really pushing some very odd edge cases.
Many of us would say around mid-last year, with the releases of GPT-OSS, GLM-4.5, Qwen 3 (for open weights) and Gemini 2.5 for closed. But YMMV.
When the developers of "AI" products include relevant disclaimers about the limitations and true costs of the tools they sell and for which they avidly, rapaiously, and relentlessly seek highly speculative investments, it might then be appropriate to think discussion of the pros and cons of their efforts should be more balanced.I can't speak to your efforts personally but I continue to see the same tit for tats in almost every AI discussion thread, this one included. The interactions barely ever move to "this could be a practical policy" or "a law like this might prevent mass redundancy while still allowing AI to assist in X sector" or "AI can be useful in this instance but we need tax funded income safety nets for such and such." It's almost always "AI is an abomination" vs "AI is the technology of our time and you just need to suck it up and get used to it."
It's fighting. It's not constructive. It feels like a lot of wasted time and energy.
This isn’t about IF LLMs can be useful, it’s about control.
We already gave up control of our social graphs to Meta. We gave up control of our PCs to Microsoft. We gave up control of video ownership to Netflix. Now you want to give up control of programming, of thinking and problem solving?
No, I'm not asking for skepticism to be set aside. I'm suggesting more efforts be directed towards how we might steer AI towards a future we want given it's here to stay in some form. Be that specific regulation, taxation, policy, incentives, etc. It's in our interests to assume AI becomes as powerful as some are suggesting so that we can do the thinking needed while there's still time. And I think part of this role falls on the tech community like us.Are you asking skepticism be set aside? We should stop resisting the deluge of fuzzy investor pitches? Now's your chance to show us who you are.
The problem with making that assumption is that it serves the interests of the very people who are pushing the fantastical narratives about what it can/will be able to do.No, I'm not asking for skepticism to be set aside. I'm suggesting more efforts be directed towards how we might steer AI towards a future we want given it's here to stay in some form. Be that specific regulation, taxation, policy, incentives, etc. It's in our interests to assume AI becomes as powerful as some are suggesting so that we can do the thinking needed while there's still time. And I think part of this role falls on the tech community like us.
While you argue about the best route, I'm going to continue to suggest rather than drive at full speed in a random direction, we slow down, figure out where this bus is headed, and look at getting drivers who aren't either fucking lying liars who lie or starry-eyed evangelists.No, I'm not asking for skepticism to be set aside. I'm suggesting more efforts be directed towards how we might steer AI towards a future we want given it's here to stay in some form. Be that specific regulation, taxation, policy, incentives, etc. It's in our interests to assume AI becomes as powerful as some are suggesting so that we can do the thinking needed while there's still time. And I think part of this role falls on the tech community like us.
Sure but that is the part that arguably can't be controlled. The speed of development is being driven by global competitive forces. It's effectively the genie out of the bottle stuff that's been touched on.While you argue about the best route, I'm going to continue to suggest rather than drive at full speed in a random direction, we slow down, figure out where this bus is headed, and look at getting drivers who aren't either fucking lying liars who lie or starry-eyed evangelists.
Okay. I'll accept your assurances and continue in Full Mock the Lemmings mode.Sure but that is the part that arguably can't be controlled. The speed of development is being driven by global competitive forces. It's effectively the genie out of the bottle stuff that's been touched on.
What can be controlled is how government and industry implement these tools via various policy and legislative levers.