Hey, I'm nearly 80 years old. I haven't written a line of code in over 10 years. But I'm coding now, with the help of Claude & Gemini, and having a great time. Each block of Python or Applescript that they generate for me is a much better learning tool than a book - I'm going through the code line by line and researching everything. And I'm also learning how to deal with LLMs and their strengths & weaknesses. Correcting them from time to time when they screw up. Lots of fun.
> Each block of Python or Applescript that they generate for me is a much better learning tool than a book - I'm going through the code line by line and researching everything.
I have been doing something similar. In my case, I prefer reading reference documentation (more to the point, more accurate), but I can never figure out where to start. These LLMs allow me to dive in and direct my own learning, by guiding my readings of that documentation (i.e. the authoritative source).
I think there has been too much emphasis (from both the hypesters and doomsayers) on AI doing the work, rather than looking at how we can use it as a learning tool.
You are an inspiration. I will remember this when I grow older. Just wanted to say this, I am 1/2 your age, and I am sure there are 1/3 or even 1/4 people here. ;)
It's cool to rediscover Applescript for me (I'm late 40's) but it's a funny thing where I can like smell the NeXT in it almost nostalgically but it's quite handy in this new era of hijacking mac mini's (OpenClaw obviously is one way to do it, but why not just straight to the core).
I personally think coders get better with age, like lounge singers.
Learning for what? That day when you write it yourself, that will never come ...
There is only so much you can learn by reading; it requires doing.
The good thing about traditional sources like books, tutorials and other people's code bases is that they give you something, but don't write your project for you.
Now you can be making a project, yet be indefinitely procrastinating the learn-by-doing part.
While I have never developed software professionally, in the four decades I have been using computers I have often written scripts and done other simple programming for my own purposes. When I was in my thirties and forties especially, I would often get enjoyably immersed in my little projects.
These days, I am feeling a new rush of drive and energy using Claude Code. At first, though, the feeling would come and go. I would come up with fun projects (in-browser synthesizers, multi-LLM translation engines) and get a brief thrill from being able to create them so quickly, but the fever would fade after a while. I started paying for the Max plan last June, but there were weeks at a time when I barely used it. I was thinking of downgrading to Pro when Opus 4.5 came along, I saw that it could handle more sophisticated tasks, and I got an idea for a big project that I wanted to do.
I have now spent the last two months having Claude write and build something I really wanted forty years ago, when I was learning Japanese and starting out as a Japanese-to-English translator: a dictionary that explains the meanings, nuances, and usages of Japanese words in English in a way accessible to an intermediate or advanced learner. Here is where it stands now:
It will take a few more months before the dictionary is more or less finished, but it has already reached a stage where it should be useful for some learners. I am releasing all of the content into the public domain, so people can use and adapt it however they like.
Same here - it's like programming with a couple of buddies. Occasionally they goof off and wreck everything, but we put it back together and end up with a finished project. I'm literally going through my backlog of projects from the early 80s! There are parts of each of these projects that were black holes for me - just didn't know enough to get a toe hold. With Karl (that's my agent) he explains everything I don't understand, does stuff, breaks stuff, and so on. It's really a blast.
As a principal engineer I feel completely let down. I've spent decades building up and accumulating expert knowledge and now that has been massively devalued. Any idiot can now prompt their way to the same software. I feel depressed and very unmotivated and expect to retire soon. Talk about a rug pull!
My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
> My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
I've been a tech lead for years and have written business critical code many times. I don't ever want to go back to writing code. I am feeling supremely empowered to go 100x faster. My contribution is still judgement, taste, architecture, etc. And the models will keep getting better. And as a result, I'll want to (and be able to) do even more.
I also absolutely LOVE that non-programmers have access to this stuff now too. I am always in favor of tools that democratize abilities.
Any "idiot" can build their own software tailored to how their brains think, without having to assemble gobs of money to hire expensive software people. Most of them were never going to hire a programmer anyway. Those ideas would've died in their heads.
> I also absolutely LOVE that non-programmers have access to this stuff now too. I am always in favor of tools that democratize abilities.
Here's the other edge of that sword. A couple back-end devs in my department vibe-coded up a standard AI-tailwind front-end of their vision of revamping our entire platform at once, which is completely at odds with the modular approach that most of the team wants to take, and would involve building out a whole system based around one concrete app and 4 vaporware future maybe apps.
And of course the higher-ups are like “But this is halfway done! With AI we can build things in 2 weeks that used to six months! Let’s just build everything now!” Nevermind that we don’t even have the requirements now, and nailing those down is the hardest part of the whole project. But the higher-ups never live through that grind.
If AI completely erases the profession of software developer, I'll find something else to do. Like I can't in good faith ever oppose a technology just because it's going to make my job redundant, that would be insane.
Any given system will still need people around to steer the AI and ensure the thing gets built and maintained responsibly. I'm working on a small team of in-house devs at a financial company, and not worried about my future at all. As an IC I'm providing more value than ever, and the backlog of potential projects is still basically endless- why would anyone want to fire me?
Just sold a house/moved out after being laid off in mid-January from a govt IT contractor(there for 8 great years and mostly remote). I started my UX Research, Design and Front End Web Design coding career in 2009, but now I think it's almost a stupid go nowhere vanishing career, thanks to AI.
I think much like you that AI is and will just continue to destroy the economy! At least I got to sell a house and make a profit--stash it away for when the big AI market crash happens (hopefully not a 2030 great depression tho). As then it's a down market and buying stocks, bitcoin and houses is always cheaper.
I don't have an answer for this, and won't pretend to.
But my take on this is that accountability will still be a purely human factor. It still is. I recently let go of a contractor who was hired to run our projects as a Scrum/PM, and his tickets were so bad (there were tickets with 3 words in them, one ticket was in the current sprint, that was blocked by a ticket deep in the backlog, basic stuff). When I confronted him about them, he said the AI generated them.
So I told him that:
1. That's not an excuse, his job is to verify what it generated and ensure it's still good.
2. That actually makes it look WORSE, that not only did he do nearly 0 work, that he didn't even check the most basic outputs. And I'm not anti-AI, I expressly said that we should absolutely use AI tools to accelerate our work. But that's not what happened here.
So you won't get to say (at least I think for another few years) "my AI was at fault" – you are ultimately responsible, not your tools. So people will still want to delegate those things down the chain. But ultimately they'll have to delegate to fewer people.
The models will not keep betting better. We have pased "peak LLM" already, by my estimate. Some of the parlour tricks that are wrapped around the models will make some incremental improvements, but the underlying models are done. More data, more parameters, are no longer doing to do anything.
> Any idiot can now prompt their way to the same software.
I must say I find this idea, and this wording, elitist in a negative way.
I don't see any fundamental problem with democratization of abilities and removal of gatekeeping.
Chances are, you were able to accumulate your expert knowledge only because:
- book writing and authorship was democratized away from the church and academia
- web content publication and production were democratized away from academia and corporations
- OSes/software/software libraries were all democratized away from corporations through open-source projects
- computer hardware was democratized away from corporations and universities
Each of the above must have cost some gatekeepers some revenue and opportunities. You were not really an idiot just because you benefited from any of them. Analogously, when someone else benefits at some cost to you, that doesn't make them an idiot either.
This is really interesting to me; I have the opposite belief.
My worry is that any idiot can prompt themselves to _bad_ software, and the differentiator is in having the right experience to prompt to _good_ software (which I believe is also possible!). As a very seasoned engineer, I don't feel personally rugpulled by LLM generated code in any way; I feel that it's a huge force multiplier for me.
Where my concern about LLM generated software comes in is much more existential: how do we train people who know the difference between bad software and good software in the future? What I've seen is a pattern where experienced engineers are excellent at steering AI to make themselves multiples more effective, and junior engineers are replacing their previous sloppy output with ten times their previous sloppy output.
For short-sighted management, this is all desirable since the sloppy output looks nice in the short term, and overall, many organizations strategically think they are pointed in the right direction doing this and are happy to downsize blaming "AI." And, for places where this never really mattered (like "make my small business landing page,") this is an complete upheaval, without a doubt.
My concern is basically: what will we do long term to get people from one end to another without the organic learning process that comes from having sloppy output curated and improved with a human touch by more senior engineers, and without an economic structure which allows "junior" engineers to subsidize themselves with low-end work while they learn? I worry greatly that in 5-10 years many organizations will end up with 10x larger balls of "legacy" garbage and 10x fewer knowledgeable people to fix it. For an experienced engineer I actually think this is a great career outlook and I can't understand the rug pull take at all; I think that today's strong and experienced engineer will be command a high amount of money and prestige in five years as the bottom drops out of software. From a "global outcomes" perspective this seems terrible, though, and I'm not quite sure what the solution is.
2. We'll come up with better techinques to make guardrails to help
Making up examples:
* right now, lots of people code with no tests. LLMs do better with tests. So, training LLMs to make new and better tests.
* right now, many things are left untested because it's work to build the infrastructure to test them. Now we have LLMs to help us build that infrustructure so we can use it make better tests for LLMs.
> Any idiot can now prompt their way to the same software.
It may look the same, but it isn't the same.
In fact if you took the time to truly learn how to do pure agentic coding (not vibe coding) you would realize as a principal engineer you have an advantage over engineers with less experience.
The more war stories, the more generalist experience, the more you can help shape the llm to make really good code and while retaining control of every line.
This is an unprecedented opportunity for experienced devs to use their hard won experience to level themselves up to the equivalence of a full team of google devs.
What I want when I'm coding, especially on open source side projects, is to retain copyright licensing over every line (cleanly, without lying about anything).
Hmm. TIL: The real exposure isn't Anthropic, OpenIA claiming your code, it's you unknowingly distributing someone else's GPL code because the model silently reproduced it, with essentially zero recourse for the model owner.
> My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
My experience is the opposite. Those with a passion for the field and the ability to dig deeply into systems are really excited right now (literally all that power just waiting to be guided to do good...and oh does it need guidance!). Those who were just going through the motions and punching a clock are pretty unmotivated and getting ready to exit.
Sometimes I dream about being laid off from my FAANG job so I have some time to use this power in more interesting than I'm doing at work (although I already get to use it in fairly interesting ways in my job).
I echo another reply here, if anything my experience coding feels even more valuable now.
It was never about writing the code—anyone can do that, students in college, junior engineers…
Experience is being able to recognize crap code when you see it, recognizing blind alleys long before days or weeks are invested heading down them. Creating an elegant API, a well structured (and well-organized) framework… Keeping it as simple as possible that just gets the job done. Designing the code-base in a way that anticipates expansion…
I've never felt the least bit threatened by LLMs.
Now if management sees it differently and experienced engineers are losing their jobs to LLMs, that's a tragedy. (Myself, I just retired a few years ago so I confess to no longer having a dog I this race.)
how would you suggest someone who just started their career moves ahead to build that “taste” for lean and elegant solutions? I am onboarding fresh grads onto my team and I see a tendency towards blindly implementing LLM generated code. I always tell people they are responsible for the code they push, so they should always research every line of code, their imported frameworks and generated solutions. They should be able to explain their choices (or the LLM’s). But I still fail to see how I can help people become this “new” brand of developer. Would be very happy to hear your thoughts or how other people are planning to tackle this. Thanks!
Sorry for the dumb question but how could you feel threatened by LLMs if you retired just a few years ago? Considering the hype started somewhere in 2022-2023.
I consider myself very good at writing software. I built and shipped many projects. I built systems from zero. Embedded, distributed, SaaS- you name it.
I'm having a lot of fun with AI. Any idiot can't prompt their way to the same software I can write. Not yet anyways.
I grew up without a mentor and my understanding of software stalled at certain points. When I couldn’t get a particular os API to work, in Google and stack overflow didn’t exist, and I had no one around me to ask. I wrote programs for years by just working around it.
After decades writing software I have done my best to be a mentor to those new to the field. My specialty is the ability to help people understand the technology they’re using, I’ve helped juniors understand and fix linker errors, engineers understand ARP poisoning, high school kids debug their robots. I’ve really enjoyed giving back.
But today, pretty much anyone except for a middle schooler could type their problems into a ChatGPT and get a more direct answer that I would be able to give. No one particularly needs mentorship as long as they know how to use an LLM correctly.
I have this feeling as well. At one point I thought when I got older it might be nice to teach - Steve Wozniak apparently does. But, it doesn't feel like I can really add much. Students have infinite teachers on youtube, and now they have Gemini/Claude/ChatGPT which are amazing. Sure, today, maybe, I could see myself as mostly a chaperone in some class to once in a while help a student out with some issue but that possibility seems like it will be gone in 1 to 2 years.
Today every single software engineer has an extremely smart and experienced mentor available to them 24/7. They don't have to meet them for coffee once a month to ask basic questions.
That said, I still feel strongly about mentorship though. It's just that you can spend your quality time with the busy person on higher-level things, like relationship building, rather than more basic questions.
How would this affect future generations of ... well anyone, when they have 24/7 access to extremely smart mentor who will find solution to pretty much any problem they might face?
Can't just offload all the hard things to the AI and let your brain waste away. There's a reason brain is equated to a muscle - you have to actively use it to grow it (not physically in size, obviously).
I agree with you about using our brains. I honestly have no idea.
But I can tell you that, just like with most things in life, this is yet another area where we are increasingly getting to do just the things we WANT to do (like think about code or features and have it appear, pixel pushing, smoothing out the actual UX, porting to faster languages) and not have to do things most people don't want to do, like drudgery (writing tests, formatting code, refactoring manually, updating documentation, manually moving tickets around like a caveman). Or to use a non tech example, having to spend hours fixing word document formatting.
So we're getting more spoiled. For example, kids have never waited for a table at a restaurant for more than 20 mins (which most people used to do all the time before abundant food delivery or reservation systems). Not that we ever enjoyed it, but learning to be bored, learning to not just get instant gratification is something that's happening all over in life.
Now it's happening even with work. So I honestly don't know how it'll affect society.
"No one particularly needs mentorship as long as they know how to use an LLM correctly."
The "as long as they know how..." is doing a lot of work there.
I expect developers with mentors who help give them the grounding they need to ask questions will get there a whole lot faster than developers without.
As a Principal SWE, who has done his fair share of big stuff.
I'm excited to work with AI. Why? Because it magnifies the thing I do well: Make technical decisions. Coding is ONE place I do that, but architecture, debugging etc. All use that same skill. Making good technical decisions.
And if you can make good choices, AI is a MEGA force multiplier. You just have to be willing to let go of the reins a hair.
As a self teaching beginner* this is where I find AI a bit limiting. When I ask ChatGPT questions about code it is always about to offer up a solution, but it often provides inappropriate responses that don't take into account the full context of a project/task. While it understands what good structure and architecture are, it's missing the awareness of good design and architecture and applying to the questions I have, and I don't have have the experience or skill set to ask those questions. It often suggests solutions (I tend to ask it for suggestions rather than full code, so I can work it out myself) that may have drawbacks that I only discover down the line.
Any suggestions to overcome this deficit in design experience? My best guess is to read some texts on code design or alternatively get a job at a place to learn design in practice. Mainly learning javascript and web app development at the moment.
*Who has had a career in a previous field, and doesn't necessarily think that learning programming with lead to another career (and is okay with that).
Same here, although hopefully won't be retiring soon.
What's missing from this is that iconic phrase that all the AI fans love to use: "I'm just having fun!"
This AI craze reminds me of a friend. He was always artistic but because of the way life goes he never really had opportunity to actively pursue art and drawing skills. When AI first came out, and specifically MidJourney he was super excited about it, used it a lot to make tons and tons of pictures for everything that his mind could think of. However, after awhile this excitement waned and he realized that he didn't actually learn anything at all. At that point he decided to find some time and spend more time practicing drawing to be able to make things by himself with his own skills, not by some chip on the other side of the world and he greatly improved in the past couple of years.
So, AI can certainly help create all the "fun!!!" projects for people who just want to see the end result, but in the end would they actually learn anything?
I mean. Sounds like the guy had existing long term goals, needed to overcome an activation threshold, and used AI as a catalyst to just get started. Seems like, behaviorally, AI was pivotal for him to learn things, even if the things he learned came from elsewhere / his own effort.
I suppose, yes, AI was like a kickstart. But the point is - he didn't just stick to AI, he realized that in terms of skill and fulfillment it's a no-go direction. Because you neither learn anything, nor create anything yourself.
IMHO any idiot can create a piece of crap.
It takes experience to create good software.
Use your experience Luke! Now you have a team of programmers to create what ever you fancy! Its been great for me, but I have only been programming C++ for 36 years.
I think it’s important for you to understand that there were always way more people who loved programming than were able to work professionally as high-level coders. Sure, if you spent most of your working life writing code, you’d be very proficient. But for many, many others, they haven’t been able to spend the time developing those muscles. Modern LLMs really are a joyful experience for people who enjoy software creation but haven’t had the 10,000 hours.
> As a principal engineer I feel completely let down. I've spent decades building up and accumulating expert knowledge and now that has been massively devalued. Any idiot can now prompt their way to the same software. I feel depressed and very unmotivated and expect to retire soon. Talk about a rug pull!
Really?
The vibe coders are running into a dark forest with a bunch of lobsters (OpenClaw) getting lost and confused in their own tech debt and you're saying they can prompt their way to the same software?
Someone just ended up wiping their entire production database with Claude and you believe that your experience is for nothing, towards companies that need stable infrastructure and predictability.
Cognitive debt is a real thing and being unable to read / write code that is broken is going to be an increasing problem which experienced engineers can solve.
Problem is, it's the people in higher positions who should be aware of that, except they don't care. All they would see is how much more profit company can make if it reduces workforce.
Plenty of engineers do realize that AI is not some magical solution to everything - but the money and hype tends to overshadow cooler heads on HN.
Completely the opposite experience here! I am a tech lead with decades of experience with various programming languages.
When it comes to producing code with an llm, most noobs get stuck producing spaghetti and rolling over. It is so bad that I have to go prompt-fix their randomly generated architecture, de-duplicate, vectorize and simplify.
If they lack domain knowledge on top of being a noob it is a complete disaster. I saw llm code pick a bad default (0) for a denominator and then "fix" that by replacing with epsilon.
It isn't the end, it is a new beginning. And I'm excited.
No offense but you sound more like a “principle coder”, not a principle engineer. At least in many domains and orgs, Most principal engineers are already spending most their time not coding. But -engineering- still take sip much or most of their time.
I felt what you describe feeling. But it lasted like a week in December. Otherwise there’s still tons of stuff to build and my teams need me to design the systems and review their designs. And their prompt machine is not replacing my good sense. There’s plenty of engineering to do, even if the coding writes itself.
I urge you to actually try these tools. You will very quickly realize you have nothing to worry about.
In the hands of a knowledgeable engineer these tools can save a lot of drudge work because you have the experience to spot when they’re going off the rails.
Now imagine someone who doesn’t have the experience, and is not able to correct where necessary. Do you really think that’s going to end well?
Really? I love LLMs because I can't stand the process of taking the model in my brain and putting it in a file. Flow State is so hard for me to hit these days.
So now I spec it out, feed it to an LLM, and monitor it while having a cup of tea. If it goes off the rails (it usually does) I redirect it. Way better than banging it out by hand.
What I keep hearing is that the people who weren't very good at writing software are the ones reluctant to embrace LLMs because they are too emotionally attached to "coding" as a discipline rather than design and architecture, which are where the interesting and actually difficult work is done.
Really? To me it seems that quite the opposite is true - people who were never very good at writing code are excited about LLMs because suddenly they can pretend to be architects without understanding what's happening in the codebase.
Same as with AI-art, where people without much drawing skills were excited about being able to make "art".
This comment about the OpenClaw guy hits a little too close to home:
“Peter Steinberger is a great example of how AI is catnip very specifically for middle-aged tech guys. they spend their 20s and 30s writing code, burn out or do management stuff for a decade, then come back in their late 40s/50s and want to try to throw that fastball again. Claude Code makes them feel like they still got it.”
Same but for me it's 25 years of accumulated personal backlog that I'm finally burning through. Like I've been a project hoarder and now I have a house elf to tidy up and do all that widget fobbering business. I just need to figure out what the rules of the house are.
This describes me nearly perfectly. Though I didn’t exactly burn out of coding, I accidentally stumbled upon being an EM while I was coding well and enjoying. But being EM stuck so I got into managing team(s) at biggish companies which means doing everything except one that I enjoy the most which is coding.
However now that I run my own startup I’m back to enjoying coding immensely because Claude takes care of grunt work of writing code while allowing me to focus on architecture, orchestration etc. Immense fun.
Me too, only I'm "only" 42! Got my first job as a programmer at 18 and (in retrospect) burnt out at some point and thought going into managment was the fix.
I spent the last 2 days primarily using Claude instead of coding things myself at work. I felt the exact opposite way. It was so unfulfilling. I’d equate it to the feeling of getting an A on a test, knowing I cheated. I didn’t accomplish anything. I didn’t learn anything. I got the end result with none of the satisfaction and learned nothing in the process.
I’m probably going to go back and redo everything with my own code.
That's interesting. I have been thinking about how the vastly different reactions people seem to have to agentic coding could be influenced by what they value about coding. To me it seems like there are three joys in coding:
1. Creating something
2. Solving puzzles
3. Learning new things
If you are primarily motivated by seeing a finished product of some sort, then I think agentic coding is transcendent. You can get an output so much quicker.
If your enjoyment comes from solving hard puzzles, digging into algorithms, how hardware works, weird machine quirks, language internals etc... then you're going to lose nearly all of that fun.
And learning new things is somewhere in the middle. I do think that you can use agentic coding to learn new technologies. I have found llms to be a phenomenal tool for teaching me things, exploring new concepts, and showing me where to go to read more from human authors. But I have to concede that the best way to learn is by doing so you will probably lose out on some depth and stickiness if you're not the one implementing something in a new technology.
Of course most people find joy in some mix of all three. And exactly what they're looking for might change from project to project. I'm curious if you were leaning more towards 2 and 3 in your recent project and that's why you were so unsatisfied with Claude Code.
I'll add "craftsmanship". It isn't just delivering "A" finished product, you want to deliver a "good", if not "the best", finished product.
I guess if you're in an iterative MVP mindset then this matters less, but that model has always made me a little queasy. I like testing and verifying the crap out of my stuff so that when I hand it off I know it's the best effort I could possibly give.
Relying on AI code denies me the deep knowledge I need to feel that level of pride and confidence. And if I'm going to take the time to read, test and verify the AI code to that level, then I might as well write most of it unless it's really repetitive.
I don't think AI coding means you stop being a craftsman. It is just a different tool. Manual coding is a hand tool, AI coding is a power tool. You still retain all of the knowledge and as much control over the codebase as you want, same with any tool.
It's a different conversation when we talk about people learning to code now though. I'd probably not recommend going for the power tool until you have a solid understanding of the manual tools.
It can be a power tool if used in a limited capacity, but I'd describe vibe-coding as sending a junior construction worker out to finish a piece of framing on his own.
Will he remember to use pressure treated lumber? Will he use the right nails? Will he space them correctly? Will the gaps be acceptable? Did he snort some bath salts and build a sandcastle in a corner for some reason?
All unknowns and you have to over-specify and play inspector. Maybe that's still faster than doing it yourself for some tasks, but I doubt most vibe-coders are doing that. And I guess it doesn't matter for toy programs that aren't meant for production, but I'm not wired to enjoy it. My challenge is restraining myself from overengineering my work and wasting time on micro-optimizations.
That's a really good point. And I agree that kind of confidence in craftsmanship is something that's missing from agentic coding today... it does make slop if you're not careful with it. Even though I've learned how to guide agents, I still have some uneasiness about missing something sloppy they have done.
But then it makes me ask if the agents will get so good that craftsmanship is a given? Then that concern goes away. When I use Go I don't worry too much about craftsmanship of the language because it was written by a lot of smart people and has proven itself to be good in production for thousands of orgs. Is there a point at which agents prove themselves capable enough that we start trusting in their craftsmanship? There's a long way to go, but I don't think that's impossible.
I find there are still opportunities to solve puzzles. Claude Code might build something in an unsatisfying or inelegant way, and you can suggest a better approach. You can absolutely write core components — the fun parts you crave — of the code and give it to an LLM to flesh out the rest.
One of the recent joys I’ve had is having CC knit together separate notebooks I’d been updating for a couple of years into a unified app. It can be a fulfilling experience.
I can see where this idea is coming from, but I don't agree with the conclusion at all. As someone who loves solving puzzles and learning new things, AI has been a godsend. I also very much like creating things, but even more than that, I like doing all three at once.
I think of AI like a microdose of Speed Force. Having super speed doesn't mean you don't like running; it just means you can run further and more often. That in turn justifies a greater amount of time spent running.
Without the Speed Force, most of the time you were reliant on vehicles (i.e. paying for third-party solutions) to get where you needed to go. With the Speed Force, not only can you suddenly meet a lot more of your transportation needs by foot, you're able to run to entirely new destinations that you'd never before considered. Eventually, you may find yourself planning trips to yet unexplored faraway harsh terrains.
If your joy in running came from attempting to push your biological physical limits, maybe you hate the Speed Force. If you enjoy spending time running and navigating unfamiliar territory, the Speed Force can give you more of that.
Sure, there are also oddballs who don't know how to run, yet insist on using the Speed Force to awkwardly jump somewhere vaguely in the vicinity of their destination. No one's saying they don't exist, but that's a completely different crowd from experienced speedsters.
> If you are primarily motivated by seeing a finished product of some sort, then I think agentic coding is transcendent
As someone who enjoys technology, and using it, and can just barely sort-of code but really not, agentic coding must be wonderful. I have barely scratched the surface with a couple of scripts. But simply translating "here's what I want, and how I would have done it the last time I used Linux 20 years ago, show me how to do it with systemd" is so much easier than digging through years of forum posts and trying to make sure they haven't all been obsoleted.
None of it is new. None of it is fancy. I do regret that people aren't getting credit for their work, but "automount this SMB share from my NAS" isn't going to make anyone's reputation. It's just going to make my day easier. I really did learn enough to set up a NAT system to share a DSL connection with an office in the late 1990s on OpenBSD. It took a long time, and I don't have that kind of free time anymore. I will never git gud. It's this, or just be another luser who goes without.
I think I'd add a #4 to this list, and that's helping people. I like making things that people can use to make their life easier. That's probably my number one.
The "creating something" idea... That's more complex. With agentic coding something can be created, but did I create it? Using agentic coding feels like hiring someone to do the work for me. For example, I just had all the windows in my house replaced. A crew came out at did it. The job is done, but I didn't do anything and felt no pride or sense of accomplishment in having these new windows. It just happened. Contrast that to a slow drain I had in my bathroom. I took the pipes apart, found the blockage, cleared it out, and reassembled the drain. When I next used the sink and the water effortlessly flowed away, I felt like I accomplished something, because I did it, not some plumber I hired.
So it isn't even about learning or solving puzzles, it's about being the person who actually did the work and seeing the result of that effort.
Yes! Good points! I think what I meant for point 1 was more "outputting something" vs "creating something". In my mind that encompasses materializing something into the world to achieve whatever you wanted, whether you were aiming to help others, solve a problem you alone have, or scratch some other sort of itch. It's about achieving some end. And helping somebody can be achieved indirectly and still be satisfying.
The inherent value of creating is something I was missing. Solving puzzles might be part of that, but not all. It's the classic Platonic question about how we value actions: for their own sake, for their results, or for both.
I think we agree that coding can be both, and it sounds like you feel the value for its own sake is lackluster in agentic coding -- It's just too easy. And I think that's the core sliding scale: Do you value creation more for its own sake or for its results? Where you land on that spectrum probably influences how people feel about agentic coding.
That being said, I also think that agentic coding can give enough of a challenge to scratch the itch of intrinsic value of creating. To a certain degree I think it's about moving up the abstraction chain to work more on architecture and product design. Those things can be fun and rewarding too. But fundamentally it's a preference.
It's kind of a weird thing. I spent 2 days working one some code, which in a way was the process of working out the requirements and functionality that was required. I then told Claude to look at it in and refactor it.
I did put in 2 days of work to come up with what Claude used to ultimately do what it did... but when I look at the resulting code, I feel nothing. Having the idea isn't the same as being the one who actually did the thing. I plan to delete the branch next week. I don't want to maintain what it did, and think it should be less complex than it made it.
I'm squarely into #1, but it usually requires #2 (at a high level) and has #3 as a side effect. But there's also #0 which kicks it all off: the triggering problem/question.
Like just yesterday I started to notice the increasing pressure of an increasingly hard-to-navigate number of Claude chats. So I went searching for something to organize them. I did find an extension, but it's for Chrome, and I'm a Firefox person, so I had Claude look at it with the initial idea of porting to Firefox. Then in the analysis, Claude mentioned creating an extension from scratch, and that's what I went for.
I've never really used JavaScript, let alone created a Firefox extension before, but in a few minutes I was iterating on one, figuring out how I wanted it to work with Claude, and now I have a very nice and featureful chats organizer. And I haven't even peeked at the code. I also now have a firm idea of this general spec of how I want arbitrary list-organizing UI to look+behave going forward.
I think your comment really captures some of the reasons behind the differences between people’s reactions to Claude pretty well.
I will add though, on 2 and 3, during most of the coding I do in my day job as a staff engineer, it’s pretty rare for me to encounter deeply interesting puzzles and really interesting things to learn. It’s not like I’m writing a compiler or and OS kernel or something; this is web dev and infra at a mid sized company. For 95% of coding tasks I do I’ve seen some variation already before and they are boring. It’s nice to have Claude power through them.
On system design and architecture, the problems still tend to be a bit more novel. I still learn things there. Claude is helpful, but not as helpful as it is for the code.
I do get the sense that some folks enjoy solving variations of familiar programming puzzles over and over again, and Claude kills that for them. That’s not me at all. I like novelty and I hate solving the same thing twice. Different tastes, I guess.
The creator of OpenClaw had a great line about this:
"If your identity is tied to you being an iOS developer, you are going to have a rough time. But if your identity is 'I'm a builder!' it is a very exciting time to be alive."
Plus, there is no rule that says you can't keep coding if it's faster for you and/or it's quicker in general. e.g I can write a Perl one liner much faster than Claude can. Heck, even if it's not faster and you enjoy coding, just keep coding.
When it comes to writing code, I can almost tell before writing code that whether this particular piece of code will be intellectually stimulating to me. If so, I write it myself without thinking about whether Claude might have done it faster. If not, I let Claude write it. Currently I'd estimate maybe 70% of the code falls in the first category, and the remaining 30% is something I would've used a lot of my own willpower to get started anyways.
Also, when I write code myself, I still ask Claude to review it. It's faster than asking a human colleague to review it, so you can have Claude review often. Just today after a five-minute review Claude said a piece of code I wrote had four bugs, three of which were hallucinations and one was a real bug. I honestly do think it would have taken me a bit more than five minutes to find that one real bug.
I'm a few years younger than the OP, but I remember the early Internet days. I started with Perl CGI scripts, ASP, even some early server side JS, in the form of Netscape Livewire.
Over the past couple months, I've created several applications with Claude Code. Personal projects that would've taken me weeks, months, or possibly forever, since I generally get distracted and move on to something else. I write pretty decent specs, break things into phases, and make sure each phase is solid before moving on to the next.
I have Claude build things in frameworks I would've never tried myself, just because it can. I do actually look at the code. Some of it is slop. In a few cases, it looks like it works, but it'll be a totally naive or insecure implementation. If I really don't like how it did something, I'll revert and give it another attempt. I also have other AIs review it and make suggestions.
It's fun, but I ultimately gain little intellectual satisfaction from it. It's not like the old days at all. I don't feel like I'm growing my skill set. Yes, I learned "something", but it's more about the capabilities of AI, not the end result.
Still, I'm convinced this is the future. Experienced developers are in the best position to work with AI. We also may not have a choice.
This past week I found and fixed a bug that happens once in 40,000 transactions working with Claude Code - Opus 4.6. Our legacy app was designed around 2008 and has had zillions of band aids added since then. Nobody (~700 person company) has been able to reliably reproduce this issue to confidently claim that they know what the cause is and how to definitively fix it. That all changed yesterday. I spent today writing up summaries that were shared far and wide. My wizard status is yet again renewed.
I disagree. I need to be able to support what I ship and answer to the details of what it does and why it does it. I can only truly do that if I write it myself.
There should also be a symbiotic relationship at a job. Yes, they get something from me, but I should also get something… learning and some amount of satisfaction… in addition to the paycheck. I can get a paycheck anywhere.
It’s not the “new norm” unless employees accept it as the new normal. I don’t know why anyone would accept a completely one-sided situation like that.
> I need to be able to support what I ship and answer to the details of what it does and why it does it. I can only truly do that if I write it myself.
How do you function on a team, where you have to maintain code others have written?
We talk to each other. If someone wrote something I don't understand, I defer to them. If someone wrote something who is no longer with the company, we trying to make sense of it, and in some cases end up re-writing some things.
There are only 3 or 4 of us working on most of the code I touch. 3 of us have worked together in some form or another for close to 20 years.
You're paid by a company to create software, so they can use it to drive business value and make a profit. You did so effortlessly. But it didn't make you feel personally fulfilled. So you're going to go back and re-do it, so you feel better?
How do you think your company's CEO is going to feel when you tell them you could be finishing the software much faster, but you'd rather not, because it feels better to do it by hand?
I'm over 50 now and feel like this as well. Haven't used Claude yet but used Codex a bunch, and it's been SO MUCH fun going over all the old perl & shell scripting stuff that I used to do years ago before I got into healthcare time and morphed to a hobby sysadmin.
Staying up and re-learning what I used to love long ago has given me a new found passion as well. Even if I do vibe code some scripts, at least I have the background now to go through them and make sure they make sense. They're things I'm using in my own homelab and not something that I'm trying to spin up a Github repo for. I'm not shipping anything. I'm refreshing my old skills and trying to bring some of them up to date.
An unfortunate reality is that my healthcare career is going to be limited due to multiple injuries along the way, and I need to try to be as current as I can in case something happens. My safety net is limited.
A real-life scene that made me chuckle last weekend…
“Oh shit, Hey Babe did you close my laptop?”
My not-very-technical friend as we returned home from a Sunday afternoon trip to the park with the kids to find his Claude Code session had been thwarted.
So excited to be getting to my backlog of apps that I've wanted but couldn't take the time to develop on my own. I'm 66 and have been in the software field in various capacities (but programming mostly as a hobby). Here's a partial list of apps I've completed in the last few months:
- Media Watch app to keep a list of movies and shows my wife and I want to watch
- Grocery List with some tracking of frequent purchases
- Health Log for medical history, doc appointments and past visits
- Habits Tracker with trends I’m interested
- Daily Wisdom Reader instead of having multiple ebooks to keep track of where I'm at
- A task manager similar to the old LifeBalance app
- A Home Inventory app so that I can track what I have, warranty, and maintenance
- An ios watch app to see when I'm asleep so that it can turn off my music or audiobook
- An ios watch chess tactics trainer app
- some games
Many of these are similar to paid offerings, but those didn't check off all the features I really wanted, so I vibe-coded my own. They all do what I want, the way I want it to.
This is the reason. I have just been vibe-coding my way for a few months now, got almost all the tools (except Browser and Mail) that I use daily, designed by me (with the help of LLM).
I tried to execute a project in 1986 and was told it was impossible. Every few years as tech has improved I tried again, but it was still impossible. CD-ROM, CD-I, Web, Wiki, even AI a few years ago...
But 2 weeks ago I taught myself to vibe code, and I built it.
40 years of planning and a few days of work.
I'm freakin' thrilled.
Maybe the internet has made me too cynical, and I'm glad people seem to be having a good time, but at time of posting I can't help but notice that almost every comment here is suspiciously vague as to what, exactly, is being coded. Still better than the breathless announcements of the death of software engineering, but quite similar in tone.
The other week I used Copilot to write a program that scans all our Amazon accounts and regions, collects services and versions, and finds the ones going EOL within a year. The data on EOL dates is scraped from several sources and kept in JSON. There's about 16 different AWS API calls used. It generates reports in markdown, json, and csv, so humans can read the markdown (flags major things, explains stuff), and the csv can be used to triage, prioritize, track work over time. The result is deduplicated, sorted, consolidated (similar entries), and classified. I can automatically send reports to teams based on a regex of names or tags. This is more data than I get from AWS Health Dashboard, and can put it into any format I want, across any number of accounts/regions.
Afaik there are no open source projects that do this. AWS has a behemoth of a distributed system you can deploy in order to do something similar. But I made a Python script that does it in an afternoon with a couple of prompts.
I am currently using a Claude skill that I have been building out over the last few days that runs through my Amazon PPC campaigns and does a full audit. Suggestions of bid adjustments, new search terms and products to advertise against and adjustment to campaign structures. It goes through all of the analytics Amazon provides, which are surprisingly extensive, to find every search term where my product shows up, gets added to cart and purchased.
It's the kind of thing that would be hours of tedious work, then even more time to actually make all the changes to the account. Instead I just say "yeah do all of that" and it is done. Magic stuff. Thousands of lines of Python to hit the Amazon APIs that I've never even looked at.
It's thousands of lines of variation on my own hand-tooling, run through tests I designed, automated by the sort of onboarding docs I should have been writing years ago.
Do you trust the assembly your compiler puts out? The machine code your assembler puts out? The virtual machine it runs on? Thousands of lines of code you've never looked at...
We agree then that you can verify, test, and trust the deterministic code an LLM produces without ever looking at it.
> That's one reason we test
That's one way we can trust and verify code produced by an LLM. You can't stop doing all the other things that aren't coding.
I get there's a difference. Shitty code can be produced by LLMs or humans. LLMs really can pump out the shitty code. I just think the argument that you cant trust code you haven't viewed is not a good argument. I very much trust a lot of code I've never seen, and yes I've been bitten by it too.
Not trying to be an ass, more trying to figure out how im going to deal for the next decade before retirement age. Uts going to be a lot of testing and verification I guess
- A slimmed-down phpBB 2 "remake" in Bun.js/TypeScript
- An experimental SQLite extension for defining incremental materialized views
...And many more that are either too tiny, too idiosyncratic, or too day-job to name here. Some of these are one-off utilities, some are toys I'll never touch again, some are part of much bigger projects that I've been struggling to get any work done on, and so on.
I don't blame you for your cynicism, and I'm not blind to all of the criticism of LLMs and LLM code. I've had many times where I feel upset, skeptical, discouraged, and alienated because of these new developments. But also... it's a lot of fun and I can't stop coming up with ideas.
> I’m ready to retire. ... Fast forward decades and Claude Code is giving me that same energy and drive. I love it. It feels like it did back then. I’m chasing the midnight hour and not getting any sleep.
Of course you love it, you don't have to worry about retirement anymore.
Give me your 401k, then tell you feel about Claude Code.
Yes! Although 60 is still a decade away, I've spent a fair few evenings vibe-coding a FOSS dependency-free raw git repo browser.[1] Never would have even started such a project without LLMs because:
* Implementing a raw Git reader is daunting.
* Codifying syntax highlighting rules is laborious.
* Developing a nice UI/UX is not super enjoyable for me.
* Hardening with latest security measures would be tricky.
* Crafting a templating language is time-consuming.
Being able to orchestrate and design the high-level architecture while letting the LLM take care of the details is extremely rewarding. Moving all my repositories away from GitLab, GitHub, and BitBucket to a single repo under my own control is priceless.
This resonates. The emotional side of returning to coding is real.
With Claude Code specifically, I've noticed that the longer it runs autonomously, the more cost anxiety creeps in. You stop thinking about the problem and start watching the token counter.
What finally let me stop worrying and just build again was building a hard budget limit outside the app — not just alerts, but an actual kill switch.
I introduced my dad to claude code. He doesn’t even code, but now it’s a more welcoming and rewarding experience from the get-go. He’s happy, became more comfortable with linux.
Occasionally I remote in to help fix something, but the coding agent really takes a load off my back, and he can start learning without knowing where the endpoints are.
I remember before style sheets existed. Webites were all nested tables and font tags. I built a video website before YouTube be even existed. Claude code and AI is definitely an exciting time.
Opposite here. I was excited by writing code and worked on open source side projects consistently. Somehow, I've not done anything since around August 2025.
I have a sense that AI could have something to do with it.
AI is degrading the status of our profession; its perception in the public eye.
At the same time, it is stealing our work and letting cretins pretend to be software engineers.
Just checked out MoveOMeter.com Great idea - and I get how pitching to "an old coot" like my parents would get a laugh out of them before an insulting hurtful pass. Very clever positioning - I'd lean in on that. Your audience is there and waiting - which is tricky since your customer is actually the sales person and you need to give them the training up front to close the deal with their elder. Nice work!
I've always dabbled in electronics, as a hobbyist. I've never had any formal courseware or training in it.
But I have been haranguing Claude/Gemini to help me on an analog computer project for some months now that has sent me on a deep dive into op-amps and other electronics esoterica that I had previously only dabbled a bit in.
Along the way I've learned about relaxation oscillators, using PWM to multiply two voltages, integrating, voltage-following…
I could lean on electronics.stackexchange (where my Google searches often lead) but 1) I first have to know what I am even searching for and 2) even the EEs disagree on how to solve a problem (as you might expect) so I am still with no clear answer. Might as well trust a sometimes hallucinating LLM?
I guess I like the first point above the best—when the LLM just out of the blue (seemingly) suggests a PWM multiplier when I was thinking log/anti-log was the only way to multiply voltages. So I get to learn a new topology.
Or I'm focused on user-adjustable pots for setting machine voltages and the LLM suggests a chip with its own internal 2.45V reference that you can use to get specific voltages without burdening the user to dial it in, own a multimeter. So I get to learn about a chip I was unfamiliar with.
It just goes on an on.
(And, Mr. Eater, I only let the magic smoke out once so far, ha ha.)
The promise/potential of ever-refining skills and agents drives this compulsion for me. "NEXT time it will be even better. And NOW it's set up to avoid the pitfalls I faced last time." You can feel the exponential engine-building.
I'm not a SWE. I'm a mechanical engineer who spends his life in excel. So when I first made my own node editor app and then asked Claude to read that for my workflow in my second project.... I felt like God herself.
From what I've seen, and of course the models get better everyday, if you have very simple grunt work that needs to be done. Coding agents are basically magic. The moment something gets either difficult or subjective, coding agents love to add completely incorrect solutions.
Try to tell Claude Code to refactor some code and see if it doesn't just delete the entire file and rewrite it. Sure that's cute, but it's absolutely not okay in a real software environment.
I do find this stuff great for hobbyist projects. I don't know if I'd be willing to put money on the line yet
As a parent to two young kids and in more of a leadership position at work, Claude allows me to grind through my backlog of ideas in minutes between other tasks, and see which ones take flight.
As a solo dev, using LLMs for coding has made me a better programmer for sure!
I can ask an LLM for specific help with my codebase and it can explain things in context and provide actual concrete relevant examples that make sense to me.
Then I can ask again for explanations about idiomatic code patterns that aren't familiar for me.
Working on my own, I don't get that feedback and code review loop.
Working with new languages and techniques, or diving into someone else's legacy code base is no longer as daunting with an LLM to ask for help!
When it was just asking ChatGPT questions it was fine, I was having fun, I was able to unblock myself when I got non-trivial errors much quicker, and I still felt like I was learning stuff.
With Codex or Claude Code, it feels like I'm stuck LARPing as a middle manager instead of actually solving problems. Sometimes I literally just copy stuff from my assigned ticket into Claude and tell it to do that, I awkwardly wait for a bit, test it out to see if it's good enough, and make my pull request. It's honestly kind of demoralizing.
I suppose this is just the cost of progress; I'm sure there were people that loved raising and breeding horses but that's not an excuse to stop building cars.
I loved being able to figure out interesting solutions to software problems and hacking on them until something worked, and my willingness to do the math beforehand would occasionally give me an edge. Instead, now all I do is sit and wait while I'm cuckolded out of my work, and questioning why I bothered finishing my masters degree if the expectation now is to ship slop code lazily written by AI in a few minutes.
It was a good ride while it lasted; I got almost fifteen years of being paid to do my favorite thing. I should count my blessings that it lasted that long, though I'm a little jealous of people born fifteen years earlier who would be retiring now with their Silicon Valley shares. Instead, I get to sit here contemplating whether or not I can even salvage my career for the next five years (or if I need to make a radical pivot).
No, I'm in my mid 30's. Unless I win the lottery (which seems unlikely considering I don't buy lottery tickets), or I managed to get some obscenely lucky with shares at a startup, I realistically will need to work for at least twenty more years before retiring.
Same! After years in engineering management I'm building so many small side projects thanks to Claude Code. I'm creating at a breakneck pace. Claude Code has mostly raised the level of abstraction so I can focus much more on the creative aspect of building which has been so much fun.
There are definitely a lot of limitations with Claude Code, but it's fun to work through the issues, figure out Claude's behavior, and create guardrails and workarounds. I do think that a lot of the poor behavior that agents exhibit can be fixed with more guardrails and scaffolding... so I'm looking forward to the future.
Getting claude to build mathematical models for me and running simulations really got me back into doing sciency things too. It's the model that's important, not the boilerplate each time!
I'm 38 years old, and as a manager, it's gradually become difficult to find joy in coding. Claude Code has helped me rediscover that pleasure. Now, all I want to do is code every day and use up my quota.
I've also been loving the speed Claude has enabled me to move at, and now agree that the coding part of SWE has become LLM-wrangling instead. I now see interacting with an LLM, to build all parts of software, as the new "frontend".
Following this idea, what do people think "backend" work will involve? Building and tweaking models, and the infra around them? Obviously everyone will shift more into architecture and strategy, but in terms of hands-on technical work I'm interested in where people see this going.
The English language has the ability to be ambiguous, but I bet AI use will change the way we use the English language colloquially, to say more specifically what we mean. I worked as a home inspector for a while. Writing for an LLM is very similar to writing a home inspection report or legal brief (or talking to a group of teenagers). Navigate the minefield with very specific intention.
I’m a 13 year lurker, first time commenter (Not sure why this post compelled me). I don’t think this is a genuine take. I don’t see how a 60 year old has any kind of joy for actual software creation suddenly from llms. It might be a joy in seeing software automatically be created but it’s definitely not doing the work. (I may be biased, I left the field 5 years ago) I doubt he’s spending any time fixing the software to make it near usable for anyone besides himself and the semi-working state the llm gave him. Meaning he’s going to have 10 or more half-finished projects again.
I agree. This seems more like an excitement or joy after getting a new toy more that actual process of creating something. Particularly when person uses LLM in a pure vibe code approach where they have no idea what's happening in the code.
Same, early 50s and this is like the heyday of coding where you could rapidly iterate on things and actively make leaps and bounds of progress. Super fun.
Another +1 from me at 62 years. My problem is this has led to me feeling like I am tech lead for a team of a dozen excellent developers, but I have no task for them!
I'll be 38 next month. I always wonder what I'm do in 30 more years and I cannot see myself NOT coding. Happy to see that spark is alive and well within you.
Glad to see this. I was tired of seeing posts that are on the extremes - "death of software by AI" vs "AI can't do this and that".
I took a break from software, and over the last few years, it just felt repetitive, like I was solving or attempting to solve the same kinds of problems in different ways every 6 months. The feeling of "not a for loop again", "not a tree search again", "not a singleton again". There's an exciting new framework or a language that solves a problem - you learn it - and then there are new problems with the language - and there is a new language to solve that language's problem. And it is necessary, and the engineer in me does understand the why of it, but over time, it just starts to feel insane and like an endless loop. Then you come to an agreement: "Just build something with what I know," but you know so much that you sometimes get stuck in analysis paralysis, and then a shiny new thing catches your engineer or programmer brain. And before you get maintainable traction, I would have spent a lot of time, sometimes quitting even before starting, because it was logistically too much.
Claude Code does make it feel like I am in my early twenties. (I am middle-aged, not in 60s)
I see a lot of comments wondering what is being built -
Think about it like this, and you can try it in a day.
Take an idea of yours, and better if it is yours - not somebody else's - and definitely not AI's. And scope it and ground it first. It should not be like "If I sway my wand, an apple should appear". If you have been in software for long, you would have heard those things. Don't be that vague. You have to have some clarity - "wand sway detection with computer vision", "auto order with X if you want a real apple", etc.. AI is a catalyst and an amplifier, not a cheat code. You can't tell it, "build me code where I have tariffs replacing taxes, and it generates prosperity". You can brainstorm, maybe find solutions, but you can't break math with AI without a rigorous theory. And if you force AI without your own reasoning, it will start throwing BS at you.
There is this idea in your mind, discuss it with ChatGPT, Gemini, or Claude. See the flaws in the idea - discover better ideas. Discuss suggestions for frameworks, accept or argue with AI. In a few minutes, you ask it to provide a Markdown spec. Give it to Claude Code. Start building - not perfect, just start. Focus on the output. Does it look good enough for now? Does it look usable? Does it make sense? Is the output (not code) something you wanted? That is the MVP to yourself.
There's a saying - customers don't care about your code, but that doesn't mean you shouldn't. In this case, make yourself the customer first - care about the code later (which in an AI era is like maybe a 30min to an hour later)
And at this point, bring in your engineer brain. Typically, at this point, the initial friction is gone, you have code and something that is working for you in real - not just on a paper or whiteboard. Take a pause. Review, ask it to refactor - make it better or make it align with your way, ask why it made the decisions it made. I always ask AI to write unit tests extensively - most of which I do not even review. The unit tests are there just to keep it predictable when I get involved, or if I ask AI to fix something. Even if you want to remove a file from the project, don't do it yourself - acclimatize to prompting and being vague sometimes. And use git so that you can revert when AI breaks things. From idea to a working thing, within an hour, and maybe 3-4 more hours once you start reviews, refactors, and engineering stuff.
I also use it for iterative trading research. It is just an experiment for now, but it's quite interesting what it can do. I give it a custom backtesting engine to use, and then give it constraints and libraries like technical indicators and custom data indicators it can use (or you could call it skills) - I ask it to program a strategy (not just parameter optimize) - run, test, log, define the next iteration itself, repeat. And I also give it an exact time for when it should stop researching, so it does not eat up all my tokens. It just frees up so much time, where you can just watch the traffic from the window or think about a direction where you want AI to go.
I wanted to incorporate astrological features into some machine learning models. An old idea that I had, but I always got crapped out because of the mythological parts and sometimes mystical parts that didn't make sense. With AI, I could ask it to strip out those unwanted parts, explain them in a physics-first or logic-first way, and get deeper into the "why did they do this calculation", "why they reached this constant", and then AI obviously helps with the code and helps explain how it matches and how it works - helps me pin point the code and the theories. Just a few weeks ago, I implemented/ported an astronomy library in Go (github.com/anupshinde/goeph) to speed up my research - and what do I really know about astronomy! But the outputs are well verified and tested.
But, in my own examples, will I ever let AI unilaterally change the custom backtesting engine code? Never. A single mistake, a single oversight, can cost a lot of real money and wasted time in weeks or months. So the engine code is protected like a fortress. You should be very careful with AI modifying critical parts of your production systems - the bug double-counting in the ledger is not the same as a "notification not shown". I think managers who are blanket-forcing AI on their employees are soon going to realize the importance of the engineering aspect in software
Just like you don't trust just any car manufacturer or just any investment fund, you should not blindly trust the AI-generated code - otherwise, you are setting yourself up to get scammed.
btw how good are any of these tools for embedded programming? we need a new era for hardware enthusiasts. my dad made plenty of fun things in the 80s but it was at the tail end of the newess that came from radiokits and other gadgets that flooded the market due to the uchip
all the insane and/or speculative projects that i never did because they would require heavy lift but with vague outcomes are now in progress. it's glorious.
try asking claude to write in VB6. Make some Active Server Pages. Use COM components. Why not? We can do things "better" now, but what does that matter when you can do the same things as before, but better?
I'm so excited to be able to continue build things when I'm living on the streets. I'm glad to know that drive to create will always be with me and keep me warm during winters.
You can't speak this kind of truth on hacker news!
But, uh, yeah... I've been noticing a growing divide between people like OP who are either already retired or are wealthy enough that they could if they wanted to who absolutely love the new world of LLMs, and people who aren't currently financially secure and realize that LLMs are going to snatch their career away. Maybe not this year, but not too far out either.
I'm enjoying the new era of agentic-coding all your ideas, but it's been obvious to me for a while that jobs are going to tend towards ones where you're liked by the decisionmaker or capital owner and kept around to be the middleman decider-delegator to others/AI/robots.
What I think is lost on ones like OP, is that yes, they are financially secure in the current climate. But if the future that everyone seems to be ushering in does come true, even ones like OP will be in a different state of security.
How does the saying go again? "It takes a village to reach financially secure retirement"
I expect to have at least 15 more years in the workforce and I hate that I have to live through this "revolution". I worry about what will be final balance of lives improved vs lives worsened.
I’m on a field trip chaperoning my kid. I get a couple slack messages asking for some tweaks to a UI. I type a couple words into a Github AI Agent Session while riding the bus. Fixes are deployed to our staging env in 10 minutes.
I don't play games anymore. I just work on whacky ideas with LLMs. I even nuked my gaming PC and installed ollama+rocm to play with local models, run openclaw there to experiment with that too. It's a lot of fun. I feel like agents are particularly useful for people who are ADD and want to work on 10 things at once.
I have this idea that probably violates some law of computing but I am really stubborn to make it happen somehow.
I want a game that generates its own mechanics on the fly using AI. Generates itself live.
Infinite game with infinite content. Not like no mans sky where everything is painfully predictable and schematic to a fault. No. Something that generates a whole method of generating. Some kind of ultra flexible communication protocol between engine and AI generator that is trained to program that protocol.
Develop it into a framework.
Use that framework to create one game. A dwarf fortress adventure mode 2.0
I have no other desires, I have no other goals, I don’t care. I or better yet - someone else, must do it.
It sounds doable. An AI can be made to keep modifying a game's codebase. I imagine it'd be easiest to separate out a scripting layer for game mechanics & behavior that AI can iterate quickly on, although of course it could more riskily modify the engine itself.
Then you could open voting up to a community for a weekly mechanics-change vote (similar to that recent repo where public voting decided what the AI would do next), and AI will implement it with whatever changes it sees fit.
Honestly, without some dedicated human guidance and taste, it would probably be more of a novelty that eventually lost its shine.
My main worry is: what is the license on the code produced by Claude (or any other coding agent)? It seems like, if it was trained using open-source software, then the resulting code needs to be open-source as well and it should be compatible with the original source. Artwork produced by an AI cannot be copyrighted, but apparently code can be?
If the software produced is for internal use, the point is probably moot. But if it isn't, this seems like a question that needs to be answered ASAP.
Claude Code and it's parallels have extinguished multiple ones.
I was able to steer clear of the Bitcoin/NFT/Passport bros but it turns out they infiltrated the profession and their starry puppy delusional eyes are trying to tell me that iteration X of product Y released yesterday evening is "going to change everything".
They have started redefining what "I have build this" actually means, and they have outjerked the executives by slinging outrageous value creation narratives.
> I’m chasing the midnight hour and not getting any sleep.
You are 60; go spend some time with your grand-kids, smell a flower, touch grass forget chasing anything at this age cause a Tuesday like the others things are gonna wrap up.
I see many comments here about Claude and I get the same feeling I get when I see comments about MacOS: it's nice that you're content with it, but I don't trust Apple/Anthropic for a fraction of an angstrom.
Wake me when we have ethically trained, open source models that run locally. Preferably high-quality ones.
I have bipolar disorder. The more frustrating aspects of coding have historically affected me tenfold (sometimes to the point of severe mania). Using Claude Code has been more like an accessibility tool in that regard. I no longer have to do the frustrating bits. Or at the very least, that aspect of the job is thoroughly diminished. And yes - coding is "fun again".
I think coding can be an endurance sport sometimes. There are a lot of points at which you have to bang your head against a wall for hours or days to figure out the smallest issue. Having an agent do that frustrating part definitely lowers the endurance needed to stay productive on a project.
Hey, I'm nearly 80 years old. I haven't written a line of code in over 10 years. But I'm coding now, with the help of Claude & Gemini, and having a great time. Each block of Python or Applescript that they generate for me is a much better learning tool than a book - I'm going through the code line by line and researching everything. And I'm also learning how to deal with LLMs and their strengths & weaknesses. Correcting them from time to time when they screw up. Lots of fun.
> Each block of Python or Applescript that they generate for me is a much better learning tool than a book - I'm going through the code line by line and researching everything.
I have been doing something similar. In my case, I prefer reading reference documentation (more to the point, more accurate), but I can never figure out where to start. These LLMs allow me to dive in and direct my own learning, by guiding my readings of that documentation (i.e. the authoritative source).
I think there has been too much emphasis (from both the hypesters and doomsayers) on AI doing the work, rather than looking at how we can use it as a learning tool.
>>>>Hey, I'm nearly 80 years old.
You are an inspiration. I will remember this when I grow older. Just wanted to say this, I am 1/2 your age, and I am sure there are 1/3 or even 1/4 people here. ;)
I'm very happy for you and hope when I'm nearing 80 I get to be doing something similar.
It's cool to rediscover Applescript for me (I'm late 40's) but it's a funny thing where I can like smell the NeXT in it almost nostalgically but it's quite handy in this new era of hijacking mac mini's (OpenClaw obviously is one way to do it, but why not just straight to the core).
I personally think coders get better with age, like lounge singers.
AppleScript doesn’t have any NeXT heritage, it comes entirely from classic MacOS (debuted in System 7.1)
Sure, but you can feel some emergent philosophies that are starting to converge and there are recognizable aesthetics.
Good for you. Learning is a life long thing!
> better learning tool than a book
Learning for what? That day when you write it yourself, that will never come ...
There is only so much you can learn by reading; it requires doing.
The good thing about traditional sources like books, tutorials and other people's code bases is that they give you something, but don't write your project for you.
Now you can be making a project, yet be indefinitely procrastinating the learn-by-doing part.
Very similar here. I am 68.
While I have never developed software professionally, in the four decades I have been using computers I have often written scripts and done other simple programming for my own purposes. When I was in my thirties and forties especially, I would often get enjoyably immersed in my little projects.
These days, I am feeling a new rush of drive and energy using Claude Code. At first, though, the feeling would come and go. I would come up with fun projects (in-browser synthesizers, multi-LLM translation engines) and get a brief thrill from being able to create them so quickly, but the fever would fade after a while. I started paying for the Max plan last June, but there were weeks at a time when I barely used it. I was thinking of downgrading to Pro when Opus 4.5 came along, I saw that it could handle more sophisticated tasks, and I got an idea for a big project that I wanted to do.
I have now spent the last two months having Claude write and build something I really wanted forty years ago, when I was learning Japanese and starting out as a Japanese-to-English translator: a dictionary that explains the meanings, nuances, and usages of Japanese words in English in a way accessible to an intermediate or advanced learner. Here is where it stands now:
https://www.tkgje.jp/
https://github.com/tkgally/je-dict-1
It will take a few more months before the dictionary is more or less finished, but it has already reached a stage where it should be useful for some learners. I am releasing all of the content into the public domain, so people can use and adapt it however they like.
This is neat that you had fun making this.
What are some good examples of where your app excels? I've currently got https://jisho.org bookmarked.
Same here - it's like programming with a couple of buddies. Occasionally they goof off and wreck everything, but we put it back together and end up with a finished project. I'm literally going through my backlog of projects from the early 80s! There are parts of each of these projects that were black holes for me - just didn't know enough to get a toe hold. With Karl (that's my agent) he explains everything I don't understand, does stuff, breaks stuff, and so on. It's really a blast.
> it's like programming with a couple of buddies. Occasionally they goof off and wreck everything,
Nailed it :)
As a principal engineer I feel completely let down. I've spent decades building up and accumulating expert knowledge and now that has been massively devalued. Any idiot can now prompt their way to the same software. I feel depressed and very unmotivated and expect to retire soon. Talk about a rug pull!
My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
> My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
I've been a tech lead for years and have written business critical code many times. I don't ever want to go back to writing code. I am feeling supremely empowered to go 100x faster. My contribution is still judgement, taste, architecture, etc. And the models will keep getting better. And as a result, I'll want to (and be able to) do even more.
I also absolutely LOVE that non-programmers have access to this stuff now too. I am always in favor of tools that democratize abilities.
Any "idiot" can build their own software tailored to how their brains think, without having to assemble gobs of money to hire expensive software people. Most of them were never going to hire a programmer anyway. Those ideas would've died in their heads.
> I also absolutely LOVE that non-programmers have access to this stuff now too. I am always in favor of tools that democratize abilities.
Here's the other edge of that sword. A couple back-end devs in my department vibe-coded up a standard AI-tailwind front-end of their vision of revamping our entire platform at once, which is completely at odds with the modular approach that most of the team wants to take, and would involve building out a whole system based around one concrete app and 4 vaporware future maybe apps.
And of course the higher-ups are like “But this is halfway done! With AI we can build things in 2 weeks that used to six months! Let’s just build everything now!” Nevermind that we don’t even have the requirements now, and nailing those down is the hardest part of the whole project. But the higher-ups never live through that grind.
What you bring to the table night be fine, but how long do you think you'll find emoloyers willing to still pay for this?
One thing is for sure LLMs will bring down down the cost of software per some unit and increase the volume.
But..cost = revenue. What is a cost to one party is a revenue to another party. The revenue is what pays salaries.
So when software costs go down the revenues will go down too. When revenues go down lay offs will happen, salary cuts will happen.
This is not fictional. Markets already reacted to this and many software service companies took a hit.
If AI completely erases the profession of software developer, I'll find something else to do. Like I can't in good faith ever oppose a technology just because it's going to make my job redundant, that would be insane.
>What you bring to the table night be fine, but how long do you think you'll find emoloyers willing to still pay for this?
I'm assuming that the software factory of the future is going to need Millwrights https://en.wikipedia.org/wiki/Millwright
But, builders are builders. These tools turn ideas into things, a builders dream.
Any given system will still need people around to steer the AI and ensure the thing gets built and maintained responsibly. I'm working on a small team of in-house devs at a financial company, and not worried about my future at all. As an IC I'm providing more value than ever, and the backlog of potential projects is still basically endless- why would anyone want to fire me?
Just sold a house/moved out after being laid off in mid-January from a govt IT contractor(there for 8 great years and mostly remote). I started my UX Research, Design and Front End Web Design coding career in 2009, but now I think it's almost a stupid go nowhere vanishing career, thanks to AI.
I think much like you that AI is and will just continue to destroy the economy! At least I got to sell a house and make a profit--stash it away for when the big AI market crash happens (hopefully not a 2030 great depression tho). As then it's a down market and buying stocks, bitcoin and houses is always cheaper.
I don't have an answer for this, and won't pretend to.
But my take on this is that accountability will still be a purely human factor. It still is. I recently let go of a contractor who was hired to run our projects as a Scrum/PM, and his tickets were so bad (there were tickets with 3 words in them, one ticket was in the current sprint, that was blocked by a ticket deep in the backlog, basic stuff). When I confronted him about them, he said the AI generated them.
So I told him that:
1. That's not an excuse, his job is to verify what it generated and ensure it's still good.
2. That actually makes it look WORSE, that not only did he do nearly 0 work, that he didn't even check the most basic outputs. And I'm not anti-AI, I expressly said that we should absolutely use AI tools to accelerate our work. But that's not what happened here.
So you won't get to say (at least I think for another few years) "my AI was at fault" – you are ultimately responsible, not your tools. So people will still want to delegate those things down the chain. But ultimately they'll have to delegate to fewer people.
The models will not keep betting better. We have pased "peak LLM" already, by my estimate. Some of the parlour tricks that are wrapped around the models will make some incremental improvements, but the underlying models are done. More data, more parameters, are no longer doing to do anything.
AI will have to take a different direction.
> Any idiot can now prompt their way to the same software.
I must say I find this idea, and this wording, elitist in a negative way.
I don't see any fundamental problem with democratization of abilities and removal of gatekeeping.
Chances are, you were able to accumulate your expert knowledge only because:
- book writing and authorship was democratized away from the church and academia
- web content publication and production were democratized away from academia and corporations
- OSes/software/software libraries were all democratized away from corporations through open-source projects
- computer hardware was democratized away from corporations and universities
Each of the above must have cost some gatekeepers some revenue and opportunities. You were not really an idiot just because you benefited from any of them. Analogously, when someone else benefits at some cost to you, that doesn't make them an idiot either.
This is a good response. Progress has always been resisted by incumbents
This is really interesting to me; I have the opposite belief.
My worry is that any idiot can prompt themselves to _bad_ software, and the differentiator is in having the right experience to prompt to _good_ software (which I believe is also possible!). As a very seasoned engineer, I don't feel personally rugpulled by LLM generated code in any way; I feel that it's a huge force multiplier for me.
Where my concern about LLM generated software comes in is much more existential: how do we train people who know the difference between bad software and good software in the future? What I've seen is a pattern where experienced engineers are excellent at steering AI to make themselves multiples more effective, and junior engineers are replacing their previous sloppy output with ten times their previous sloppy output.
For short-sighted management, this is all desirable since the sloppy output looks nice in the short term, and overall, many organizations strategically think they are pointed in the right direction doing this and are happy to downsize blaming "AI." And, for places where this never really mattered (like "make my small business landing page,") this is an complete upheaval, without a doubt.
My concern is basically: what will we do long term to get people from one end to another without the organic learning process that comes from having sloppy output curated and improved with a human touch by more senior engineers, and without an economic structure which allows "junior" engineers to subsidize themselves with low-end work while they learn? I worry greatly that in 5-10 years many organizations will end up with 10x larger balls of "legacy" garbage and 10x fewer knowledgeable people to fix it. For an experienced engineer I actually think this is a great career outlook and I can't understand the rug pull take at all; I think that today's strong and experienced engineer will be command a high amount of money and prestige in five years as the bottom drops out of software. From a "global outcomes" perspective this seems terrible, though, and I'm not quite sure what the solution is.
My guesses are
1. We'll train the LLMs not to make sloppy code.
2. We'll come up with better techinques to make guardrails to help
Making up examples:
* right now, lots of people code with no tests. LLMs do better with tests. So, training LLMs to make new and better tests.
* right now, many things are left untested because it's work to build the infrastructure to test them. Now we have LLMs to help us build that infrustructure so we can use it make better tests for LLMs.
* ...?
> Any idiot can now prompt their way to the same software.
It may look the same, but it isn't the same.
In fact if you took the time to truly learn how to do pure agentic coding (not vibe coding) you would realize as a principal engineer you have an advantage over engineers with less experience.
The more war stories, the more generalist experience, the more you can help shape the llm to make really good code and while retaining control of every line.
This is an unprecedented opportunity for experienced devs to use their hard won experience to level themselves up to the equivalence of a full team of google devs.
> while retaining control of every line
What I want when I'm coding, especially on open source side projects, is to retain copyright licensing over every line (cleanly, without lying about anything).
Whoops!
Hmm. TIL: The real exposure isn't Anthropic, OpenIA claiming your code, it's you unknowingly distributing someone else's GPL code because the model silently reproduced it, with essentially zero recourse for the model owner.
> My experience is that people who weren't very good at writing software are the ones now "most excited" to "create" with a LLM.
My experience is the opposite. Those with a passion for the field and the ability to dig deeply into systems are really excited right now (literally all that power just waiting to be guided to do good...and oh does it need guidance!). Those who were just going through the motions and punching a clock are pretty unmotivated and getting ready to exit.
Sometimes I dream about being laid off from my FAANG job so I have some time to use this power in more interesting than I'm doing at work (although I already get to use it in fairly interesting ways in my job).
I echo another reply here, if anything my experience coding feels even more valuable now.
It was never about writing the code—anyone can do that, students in college, junior engineers…
Experience is being able to recognize crap code when you see it, recognizing blind alleys long before days or weeks are invested heading down them. Creating an elegant API, a well structured (and well-organized) framework… Keeping it as simple as possible that just gets the job done. Designing the code-base in a way that anticipates expansion…
I've never felt the least bit threatened by LLMs.
Now if management sees it differently and experienced engineers are losing their jobs to LLMs, that's a tragedy. (Myself, I just retired a few years ago so I confess to no longer having a dog I this race.)
how would you suggest someone who just started their career moves ahead to build that “taste” for lean and elegant solutions? I am onboarding fresh grads onto my team and I see a tendency towards blindly implementing LLM generated code. I always tell people they are responsible for the code they push, so they should always research every line of code, their imported frameworks and generated solutions. They should be able to explain their choices (or the LLM’s). But I still fail to see how I can help people become this “new” brand of developer. Would be very happy to hear your thoughts or how other people are planning to tackle this. Thanks!
Sorry for the dumb question but how could you feel threatened by LLMs if you retired just a few years ago? Considering the hype started somewhere in 2022-2023.
and they only got really good like last December.
I consider myself very good at writing software. I built and shipped many projects. I built systems from zero. Embedded, distributed, SaaS- you name it.
I'm having a lot of fun with AI. Any idiot can't prompt their way to the same software I can write. Not yet anyways.
I’m with you here.
I grew up without a mentor and my understanding of software stalled at certain points. When I couldn’t get a particular os API to work, in Google and stack overflow didn’t exist, and I had no one around me to ask. I wrote programs for years by just working around it.
After decades writing software I have done my best to be a mentor to those new to the field. My specialty is the ability to help people understand the technology they’re using, I’ve helped juniors understand and fix linker errors, engineers understand ARP poisoning, high school kids debug their robots. I’ve really enjoyed giving back.
But today, pretty much anyone except for a middle schooler could type their problems into a ChatGPT and get a more direct answer that I would be able to give. No one particularly needs mentorship as long as they know how to use an LLM correctly.
I have this feeling as well. At one point I thought when I got older it might be nice to teach - Steve Wozniak apparently does. But, it doesn't feel like I can really add much. Students have infinite teachers on youtube, and now they have Gemini/Claude/ChatGPT which are amazing. Sure, today, maybe, I could see myself as mostly a chaperone in some class to once in a while help a student out with some issue but that possibility seems like it will be gone in 1 to 2 years.
Today every single software engineer has an extremely smart and experienced mentor available to them 24/7. They don't have to meet them for coffee once a month to ask basic questions.
That said, I still feel strongly about mentorship though. It's just that you can spend your quality time with the busy person on higher-level things, like relationship building, rather than more basic questions.
How would this affect future generations of ... well anyone, when they have 24/7 access to extremely smart mentor who will find solution to pretty much any problem they might face?
Can't just offload all the hard things to the AI and let your brain waste away. There's a reason brain is equated to a muscle - you have to actively use it to grow it (not physically in size, obviously).
I agree with you about using our brains. I honestly have no idea.
But I can tell you that, just like with most things in life, this is yet another area where we are increasingly getting to do just the things we WANT to do (like think about code or features and have it appear, pixel pushing, smoothing out the actual UX, porting to faster languages) and not have to do things most people don't want to do, like drudgery (writing tests, formatting code, refactoring manually, updating documentation, manually moving tickets around like a caveman). Or to use a non tech example, having to spend hours fixing word document formatting.
So we're getting more spoiled. For example, kids have never waited for a table at a restaurant for more than 20 mins (which most people used to do all the time before abundant food delivery or reservation systems). Not that we ever enjoyed it, but learning to be bored, learning to not just get instant gratification is something that's happening all over in life.
Now it's happening even with work. So I honestly don't know how it'll affect society.
"No one particularly needs mentorship as long as they know how to use an LLM correctly."
The "as long as they know how..." is doing a lot of work there.
I expect developers with mentors who help give them the grounding they need to ask questions will get there a whole lot faster than developers without.
As a Principal SWE, who has done his fair share of big stuff.
I'm excited to work with AI. Why? Because it magnifies the thing I do well: Make technical decisions. Coding is ONE place I do that, but architecture, debugging etc. All use that same skill. Making good technical decisions.
And if you can make good choices, AI is a MEGA force multiplier. You just have to be willing to let go of the reins a hair.
As a self teaching beginner* this is where I find AI a bit limiting. When I ask ChatGPT questions about code it is always about to offer up a solution, but it often provides inappropriate responses that don't take into account the full context of a project/task. While it understands what good structure and architecture are, it's missing the awareness of good design and architecture and applying to the questions I have, and I don't have have the experience or skill set to ask those questions. It often suggests solutions (I tend to ask it for suggestions rather than full code, so I can work it out myself) that may have drawbacks that I only discover down the line.
Any suggestions to overcome this deficit in design experience? My best guess is to read some texts on code design or alternatively get a job at a place to learn design in practice. Mainly learning javascript and web app development at the moment.
*Who has had a career in a previous field, and doesn't necessarily think that learning programming with lead to another career (and is okay with that).
Same here, although hopefully won't be retiring soon.
What's missing from this is that iconic phrase that all the AI fans love to use: "I'm just having fun!"
This AI craze reminds me of a friend. He was always artistic but because of the way life goes he never really had opportunity to actively pursue art and drawing skills. When AI first came out, and specifically MidJourney he was super excited about it, used it a lot to make tons and tons of pictures for everything that his mind could think of. However, after awhile this excitement waned and he realized that he didn't actually learn anything at all. At that point he decided to find some time and spend more time practicing drawing to be able to make things by himself with his own skills, not by some chip on the other side of the world and he greatly improved in the past couple of years.
So, AI can certainly help create all the "fun!!!" projects for people who just want to see the end result, but in the end would they actually learn anything?
I mean. Sounds like the guy had existing long term goals, needed to overcome an activation threshold, and used AI as a catalyst to just get started. Seems like, behaviorally, AI was pivotal for him to learn things, even if the things he learned came from elsewhere / his own effort.
I suppose, yes, AI was like a kickstart. But the point is - he didn't just stick to AI, he realized that in terms of skill and fulfillment it's a no-go direction. Because you neither learn anything, nor create anything yourself.
IMHO any idiot can create a piece of crap. It takes experience to create good software. Use your experience Luke! Now you have a team of programmers to create what ever you fancy! Its been great for me, but I have only been programming C++ for 36 years.
I think it’s important for you to understand that there were always way more people who loved programming than were able to work professionally as high-level coders. Sure, if you spent most of your working life writing code, you’d be very proficient. But for many, many others, they haven’t been able to spend the time developing those muscles. Modern LLMs really are a joyful experience for people who enjoy software creation but haven’t had the 10,000 hours.
> As a principal engineer I feel completely let down. I've spent decades building up and accumulating expert knowledge and now that has been massively devalued. Any idiot can now prompt their way to the same software. I feel depressed and very unmotivated and expect to retire soon. Talk about a rug pull!
Really?
The vibe coders are running into a dark forest with a bunch of lobsters (OpenClaw) getting lost and confused in their own tech debt and you're saying they can prompt their way to the same software?
Someone just ended up wiping their entire production database with Claude and you believe that your experience is for nothing, towards companies that need stable infrastructure and predictability.
Cognitive debt is a real thing and being unable to read / write code that is broken is going to be an increasing problem which experienced engineers can solve.
Do not fall for the AI agent hype.
> Do not fall for the AI agent hype.
Problem is, it's the people in higher positions who should be aware of that, except they don't care. All they would see is how much more profit company can make if it reduces workforce.
Plenty of engineers do realize that AI is not some magical solution to everything - but the money and hype tends to overshadow cooler heads on HN.
Completely the opposite experience here! I am a tech lead with decades of experience with various programming languages.
When it comes to producing code with an llm, most noobs get stuck producing spaghetti and rolling over. It is so bad that I have to go prompt-fix their randomly generated architecture, de-duplicate, vectorize and simplify.
If they lack domain knowledge on top of being a noob it is a complete disaster. I saw llm code pick a bad default (0) for a denominator and then "fix" that by replacing with epsilon.
It isn't the end, it is a new beginning. And I'm excited.
I thought this was parody until the last sentence.
On the bright side, working in tech between 2006 and 2026 means you should be extremely wealthy and able to retire comfortably.
Uh if you worked for a top company or something. Most tech workers have made relatively ordinary salaries the last 20 years.
No offense but you sound more like a “principle coder”, not a principle engineer. At least in many domains and orgs, Most principal engineers are already spending most their time not coding. But -engineering- still take sip much or most of their time.
I felt what you describe feeling. But it lasted like a week in December. Otherwise there’s still tons of stuff to build and my teams need me to design the systems and review their designs. And their prompt machine is not replacing my good sense. There’s plenty of engineering to do, even if the coding writes itself.
I make documentation and diagrams for myself rather than writing code much of the time
I urge you to actually try these tools. You will very quickly realize you have nothing to worry about.
In the hands of a knowledgeable engineer these tools can save a lot of drudge work because you have the experience to spot when they’re going off the rails.
Now imagine someone who doesn’t have the experience, and is not able to correct where necessary. Do you really think that’s going to end well?
Really? I love LLMs because I can't stand the process of taking the model in my brain and putting it in a file. Flow State is so hard for me to hit these days.
So now I spec it out, feed it to an LLM, and monitor it while having a cup of tea. If it goes off the rails (it usually does) I redirect it. Way better than banging it out by hand.
Why did you leave this as a comment on someone talking about how happy they were about their own experience?
What I keep hearing is that the people who weren't very good at writing software are the ones reluctant to embrace LLMs because they are too emotionally attached to "coding" as a discipline rather than design and architecture, which are where the interesting and actually difficult work is done.
Really? To me it seems that quite the opposite is true - people who were never very good at writing code are excited about LLMs because suddenly they can pretend to be architects without understanding what's happening in the codebase.
Same as with AI-art, where people without much drawing skills were excited about being able to make "art".
Perhaps you are both right. People who see coding as a means to an end enjoy LLMs while people who saw it as the most enjoyable part don’t.
This comment about the OpenClaw guy hits a little too close to home:
“Peter Steinberger is a great example of how AI is catnip very specifically for middle-aged tech guys. they spend their 20s and 30s writing code, burn out or do management stuff for a decade, then come back in their late 40s/50s and want to try to throw that fastball again. Claude Code makes them feel like they still got it.”
And why would they not? do they have to feel they ain’t got it anymore because age?
Same but for me it's 25 years of accumulated personal backlog that I'm finally burning through. Like I've been a project hoarder and now I have a house elf to tidy up and do all that widget fobbering business. I just need to figure out what the rules of the house are.
> late 40s
This describes me nearly perfectly. Though I didn’t exactly burn out of coding, I accidentally stumbled upon being an EM while I was coding well and enjoying. But being EM stuck so I got into managing team(s) at biggish companies which means doing everything except one that I enjoy the most which is coding.
However now that I run my own startup I’m back to enjoying coding immensely because Claude takes care of grunt work of writing code while allowing me to focus on architecture, orchestration etc. Immense fun.
Me too, only I'm "only" 42! Got my first job as a programmer at 18 and (in retrospect) burnt out at some point and thought going into managment was the fix.
If you don’t mind sharing, what does your startup do?
What is an "EM"?
Engineering Manager (as opposed to people who stick to programming, called Individual Contributor.)
Oh, how I hate these horrible job descriptions.
But thanks for the info.
I spent the last 2 days primarily using Claude instead of coding things myself at work. I felt the exact opposite way. It was so unfulfilling. I’d equate it to the feeling of getting an A on a test, knowing I cheated. I didn’t accomplish anything. I didn’t learn anything. I got the end result with none of the satisfaction and learned nothing in the process.
I’m probably going to go back and redo everything with my own code.
That's interesting. I have been thinking about how the vastly different reactions people seem to have to agentic coding could be influenced by what they value about coding. To me it seems like there are three joys in coding:
1. Creating something
2. Solving puzzles
3. Learning new things
If you are primarily motivated by seeing a finished product of some sort, then I think agentic coding is transcendent. You can get an output so much quicker.
If your enjoyment comes from solving hard puzzles, digging into algorithms, how hardware works, weird machine quirks, language internals etc... then you're going to lose nearly all of that fun.
And learning new things is somewhere in the middle. I do think that you can use agentic coding to learn new technologies. I have found llms to be a phenomenal tool for teaching me things, exploring new concepts, and showing me where to go to read more from human authors. But I have to concede that the best way to learn is by doing so you will probably lose out on some depth and stickiness if you're not the one implementing something in a new technology.
Of course most people find joy in some mix of all three. And exactly what they're looking for might change from project to project. I'm curious if you were leaning more towards 2 and 3 in your recent project and that's why you were so unsatisfied with Claude Code.
I'll add "craftsmanship". It isn't just delivering "A" finished product, you want to deliver a "good", if not "the best", finished product.
I guess if you're in an iterative MVP mindset then this matters less, but that model has always made me a little queasy. I like testing and verifying the crap out of my stuff so that when I hand it off I know it's the best effort I could possibly give.
Relying on AI code denies me the deep knowledge I need to feel that level of pride and confidence. And if I'm going to take the time to read, test and verify the AI code to that level, then I might as well write most of it unless it's really repetitive.
I don't think AI coding means you stop being a craftsman. It is just a different tool. Manual coding is a hand tool, AI coding is a power tool. You still retain all of the knowledge and as much control over the codebase as you want, same with any tool.
It's a different conversation when we talk about people learning to code now though. I'd probably not recommend going for the power tool until you have a solid understanding of the manual tools.
It can be a power tool if used in a limited capacity, but I'd describe vibe-coding as sending a junior construction worker out to finish a piece of framing on his own.
Will he remember to use pressure treated lumber? Will he use the right nails? Will he space them correctly? Will the gaps be acceptable? Did he snort some bath salts and build a sandcastle in a corner for some reason?
All unknowns and you have to over-specify and play inspector. Maybe that's still faster than doing it yourself for some tasks, but I doubt most vibe-coders are doing that. And I guess it doesn't matter for toy programs that aren't meant for production, but I'm not wired to enjoy it. My challenge is restraining myself from overengineering my work and wasting time on micro-optimizations.
That's a really good point. And I agree that kind of confidence in craftsmanship is something that's missing from agentic coding today... it does make slop if you're not careful with it. Even though I've learned how to guide agents, I still have some uneasiness about missing something sloppy they have done.
But then it makes me ask if the agents will get so good that craftsmanship is a given? Then that concern goes away. When I use Go I don't worry too much about craftsmanship of the language because it was written by a lot of smart people and has proven itself to be good in production for thousands of orgs. Is there a point at which agents prove themselves capable enough that we start trusting in their craftsmanship? There's a long way to go, but I don't think that's impossible.
I find there are still opportunities to solve puzzles. Claude Code might build something in an unsatisfying or inelegant way, and you can suggest a better approach. You can absolutely write core components — the fun parts you crave — of the code and give it to an LLM to flesh out the rest.
One of the recent joys I’ve had is having CC knit together separate notebooks I’d been updating for a couple of years into a unified app. It can be a fulfilling experience.
I can see where this idea is coming from, but I don't agree with the conclusion at all. As someone who loves solving puzzles and learning new things, AI has been a godsend. I also very much like creating things, but even more than that, I like doing all three at once.
I think of AI like a microdose of Speed Force. Having super speed doesn't mean you don't like running; it just means you can run further and more often. That in turn justifies a greater amount of time spent running.
Without the Speed Force, most of the time you were reliant on vehicles (i.e. paying for third-party solutions) to get where you needed to go. With the Speed Force, not only can you suddenly meet a lot more of your transportation needs by foot, you're able to run to entirely new destinations that you'd never before considered. Eventually, you may find yourself planning trips to yet unexplored faraway harsh terrains.
If your joy in running came from attempting to push your biological physical limits, maybe you hate the Speed Force. If you enjoy spending time running and navigating unfamiliar territory, the Speed Force can give you more of that.
Sure, there are also oddballs who don't know how to run, yet insist on using the Speed Force to awkwardly jump somewhere vaguely in the vicinity of their destination. No one's saying they don't exist, but that's a completely different crowd from experienced speedsters.
> If you are primarily motivated by seeing a finished product of some sort, then I think agentic coding is transcendent
As someone who enjoys technology, and using it, and can just barely sort-of code but really not, agentic coding must be wonderful. I have barely scratched the surface with a couple of scripts. But simply translating "here's what I want, and how I would have done it the last time I used Linux 20 years ago, show me how to do it with systemd" is so much easier than digging through years of forum posts and trying to make sure they haven't all been obsoleted.
None of it is new. None of it is fancy. I do regret that people aren't getting credit for their work, but "automount this SMB share from my NAS" isn't going to make anyone's reputation. It's just going to make my day easier. I really did learn enough to set up a NAT system to share a DSL connection with an office in the late 1990s on OpenBSD. It took a long time, and I don't have that kind of free time anymore. I will never git gud. It's this, or just be another luser who goes without.
I think I'd add a #4 to this list, and that's helping people. I like making things that people can use to make their life easier. That's probably my number one.
The "creating something" idea... That's more complex. With agentic coding something can be created, but did I create it? Using agentic coding feels like hiring someone to do the work for me. For example, I just had all the windows in my house replaced. A crew came out at did it. The job is done, but I didn't do anything and felt no pride or sense of accomplishment in having these new windows. It just happened. Contrast that to a slow drain I had in my bathroom. I took the pipes apart, found the blockage, cleared it out, and reassembled the drain. When I next used the sink and the water effortlessly flowed away, I felt like I accomplished something, because I did it, not some plumber I hired.
So it isn't even about learning or solving puzzles, it's about being the person who actually did the work and seeing the result of that effort.
Yes! Good points! I think what I meant for point 1 was more "outputting something" vs "creating something". In my mind that encompasses materializing something into the world to achieve whatever you wanted, whether you were aiming to help others, solve a problem you alone have, or scratch some other sort of itch. It's about achieving some end. And helping somebody can be achieved indirectly and still be satisfying.
The inherent value of creating is something I was missing. Solving puzzles might be part of that, but not all. It's the classic Platonic question about how we value actions: for their own sake, for their results, or for both.
I think we agree that coding can be both, and it sounds like you feel the value for its own sake is lackluster in agentic coding -- It's just too easy. And I think that's the core sliding scale: Do you value creation more for its own sake or for its results? Where you land on that spectrum probably influences how people feel about agentic coding.
That being said, I also think that agentic coding can give enough of a challenge to scratch the itch of intrinsic value of creating. To a certain degree I think it's about moving up the abstraction chain to work more on architecture and product design. Those things can be fun and rewarding too. But fundamentally it's a preference.
It's kind of a weird thing. I spent 2 days working one some code, which in a way was the process of working out the requirements and functionality that was required. I then told Claude to look at it in and refactor it.
I did put in 2 days of work to come up with what Claude used to ultimately do what it did... but when I look at the resulting code, I feel nothing. Having the idea isn't the same as being the one who actually did the thing. I plan to delete the branch next week. I don't want to maintain what it did, and think it should be less complex than it made it.
I'm squarely into #1, but it usually requires #2 (at a high level) and has #3 as a side effect. But there's also #0 which kicks it all off: the triggering problem/question.
Like just yesterday I started to notice the increasing pressure of an increasingly hard-to-navigate number of Claude chats. So I went searching for something to organize them. I did find an extension, but it's for Chrome, and I'm a Firefox person, so I had Claude look at it with the initial idea of porting to Firefox. Then in the analysis, Claude mentioned creating an extension from scratch, and that's what I went for.
I've never really used JavaScript, let alone created a Firefox extension before, but in a few minutes I was iterating on one, figuring out how I wanted it to work with Claude, and now I have a very nice and featureful chats organizer. And I haven't even peeked at the code. I also now have a firm idea of this general spec of how I want arbitrary list-organizing UI to look+behave going forward.
I think your comment really captures some of the reasons behind the differences between people’s reactions to Claude pretty well.
I will add though, on 2 and 3, during most of the coding I do in my day job as a staff engineer, it’s pretty rare for me to encounter deeply interesting puzzles and really interesting things to learn. It’s not like I’m writing a compiler or and OS kernel or something; this is web dev and infra at a mid sized company. For 95% of coding tasks I do I’ve seen some variation already before and they are boring. It’s nice to have Claude power through them.
On system design and architecture, the problems still tend to be a bit more novel. I still learn things there. Claude is helpful, but not as helpful as it is for the code.
I do get the sense that some folks enjoy solving variations of familiar programming puzzles over and over again, and Claude kills that for them. That’s not me at all. I like novelty and I hate solving the same thing twice. Different tastes, I guess.
You're forgetting that (1) brings a sense of pride. "I built this". That's not true in many ways if you ask something else to do it
The creator of OpenClaw had a great line about this:
"If your identity is tied to you being an iOS developer, you are going to have a rough time. But if your identity is 'I'm a builder!' it is a very exciting time to be alive."
Plus, there is no rule that says you can't keep coding if it's faster for you and/or it's quicker in general. e.g I can write a Perl one liner much faster than Claude can. Heck, even if it's not faster and you enjoy coding, just keep coding.
When it comes to writing code, I can almost tell before writing code that whether this particular piece of code will be intellectually stimulating to me. If so, I write it myself without thinking about whether Claude might have done it faster. If not, I let Claude write it. Currently I'd estimate maybe 70% of the code falls in the first category, and the remaining 30% is something I would've used a lot of my own willpower to get started anyways.
Also, when I write code myself, I still ask Claude to review it. It's faster than asking a human colleague to review it, so you can have Claude review often. Just today after a five-minute review Claude said a piece of code I wrote had four bugs, three of which were hallucinations and one was a real bug. I honestly do think it would have taken me a bit more than five minutes to find that one real bug.
I'm a few years younger than the OP, but I remember the early Internet days. I started with Perl CGI scripts, ASP, even some early server side JS, in the form of Netscape Livewire.
Over the past couple months, I've created several applications with Claude Code. Personal projects that would've taken me weeks, months, or possibly forever, since I generally get distracted and move on to something else. I write pretty decent specs, break things into phases, and make sure each phase is solid before moving on to the next.
I have Claude build things in frameworks I would've never tried myself, just because it can. I do actually look at the code. Some of it is slop. In a few cases, it looks like it works, but it'll be a totally naive or insecure implementation. If I really don't like how it did something, I'll revert and give it another attempt. I also have other AIs review it and make suggestions.
It's fun, but I ultimately gain little intellectual satisfaction from it. It's not like the old days at all. I don't feel like I'm growing my skill set. Yes, I learned "something", but it's more about the capabilities of AI, not the end result.
Still, I'm convinced this is the future. Experienced developers are in the best position to work with AI. We also may not have a choice.
This past week I found and fixed a bug that happens once in 40,000 transactions working with Claude Code - Opus 4.6. Our legacy app was designed around 2008 and has had zillions of band aids added since then. Nobody (~700 person company) has been able to reliably reproduce this issue to confidently claim that they know what the cause is and how to definitively fix it. That all changed yesterday. I spent today writing up summaries that were shared far and wide. My wizard status is yet again renewed.
For fun and education purposes, learning and satisfaction are understandable.
For work, companies won't support it. Get it done. Fast. That's the new norm.
I disagree. I need to be able to support what I ship and answer to the details of what it does and why it does it. I can only truly do that if I write it myself.
There should also be a symbiotic relationship at a job. Yes, they get something from me, but I should also get something… learning and some amount of satisfaction… in addition to the paycheck. I can get a paycheck anywhere.
It’s not the “new norm” unless employees accept it as the new normal. I don’t know why anyone would accept a completely one-sided situation like that.
> I need to be able to support what I ship and answer to the details of what it does and why it does it. I can only truly do that if I write it myself.
How do you function on a team, where you have to maintain code others have written?
We talk to each other. If someone wrote something I don't understand, I defer to them. If someone wrote something who is no longer with the company, we trying to make sense of it, and in some cases end up re-writing some things.
There are only 3 or 4 of us working on most of the code I touch. 3 of us have worked together in some form or another for close to 20 years.
You're paid by a company to create software, so they can use it to drive business value and make a profit. You did so effortlessly. But it didn't make you feel personally fulfilled. So you're going to go back and re-do it, so you feel better?
How do you think your company's CEO is going to feel when you tell them you could be finishing the software much faster, but you'd rather not, because it feels better to do it by hand?
I think it depends what you're building. I find it fun, once in a while, an engineer to "not go shoeless" and get some of things I need done.
Your choices are not limited to one extreme or the other.
I'm over 50 now and feel like this as well. Haven't used Claude yet but used Codex a bunch, and it's been SO MUCH fun going over all the old perl & shell scripting stuff that I used to do years ago before I got into healthcare time and morphed to a hobby sysadmin.
Staying up and re-learning what I used to love long ago has given me a new found passion as well. Even if I do vibe code some scripts, at least I have the background now to go through them and make sure they make sense. They're things I'm using in my own homelab and not something that I'm trying to spin up a Github repo for. I'm not shipping anything. I'm refreshing my old skills and trying to bring some of them up to date. An unfortunate reality is that my healthcare career is going to be limited due to multiple injuries along the way, and I need to try to be as current as I can in case something happens. My safety net is limited.
A real-life scene that made me chuckle last weekend…
“Oh shit, Hey Babe did you close my laptop?”
My not-very-technical friend as we returned home from a Sunday afternoon trip to the park with the kids to find his Claude Code session had been thwarted.
- Grocery List with some tracking of frequent purchases
- Health Log for medical history, doc appointments and past visits
- Habits Tracker with trends I’m interested
- Daily Wisdom Reader instead of having multiple ebooks to keep track of where I'm at
- A task manager similar to the old LifeBalance app
- A Home Inventory app so that I can track what I have, warranty, and maintenance
- An ios watch app to see when I'm asleep so that it can turn off my music or audiobook
- An ios watch chess tactics trainer app
- some games
Many of these are similar to paid offerings, but those didn't check off all the features I really wanted, so I vibe-coded my own. They all do what I want, the way I want it to.
This is the reason. I have just been vibe-coding my way for a few months now, got almost all the tools (except Browser and Mail) that I use daily, designed by me (with the help of LLM).
And the biggest thing is that: software the way we want is much easier. No ads. No monthly cost.
I tried to execute a project in 1986 and was told it was impossible. Every few years as tech has improved I tried again, but it was still impossible. CD-ROM, CD-I, Web, Wiki, even AI a few years ago... But 2 weeks ago I taught myself to vibe code, and I built it. 40 years of planning and a few days of work. I'm freakin' thrilled.
Maybe the internet has made me too cynical, and I'm glad people seem to be having a good time, but at time of posting I can't help but notice that almost every comment here is suspiciously vague as to what, exactly, is being coded. Still better than the breathless announcements of the death of software engineering, but quite similar in tone.
The other week I used Copilot to write a program that scans all our Amazon accounts and regions, collects services and versions, and finds the ones going EOL within a year. The data on EOL dates is scraped from several sources and kept in JSON. There's about 16 different AWS API calls used. It generates reports in markdown, json, and csv, so humans can read the markdown (flags major things, explains stuff), and the csv can be used to triage, prioritize, track work over time. The result is deduplicated, sorted, consolidated (similar entries), and classified. I can automatically send reports to teams based on a regex of names or tags. This is more data than I get from AWS Health Dashboard, and can put it into any format I want, across any number of accounts/regions.
Afaik there are no open source projects that do this. AWS has a behemoth of a distributed system you can deploy in order to do something similar. But I made a Python script that does it in an afternoon with a couple of prompts.
I am currently using a Claude skill that I have been building out over the last few days that runs through my Amazon PPC campaigns and does a full audit. Suggestions of bid adjustments, new search terms and products to advertise against and adjustment to campaign structures. It goes through all of the analytics Amazon provides, which are surprisingly extensive, to find every search term where my product shows up, gets added to cart and purchased.
It's the kind of thing that would be hours of tedious work, then even more time to actually make all the changes to the account. Instead I just say "yeah do all of that" and it is done. Magic stuff. Thousands of lines of Python to hit the Amazon APIs that I've never even looked at.
And it doesn't freak you out that you're relying on thousands of lines of code that you've never looked at? How do you verify the end result?
I wouldn't trust thousands of lines of code from one of my co-workers without testing
It's thousands of lines of variation on my own hand-tooling, run through tests I designed, automated by the sort of onboarding docs I should have been writing years ago.
Do you trust the assembly your compiler puts out? The machine code your assembler puts out? The virtual machine it runs on? Thousands of lines of code you've never looked at...
None of that is generated by an LLM prone to hallucination and is perfectly deterministic unless there's a hardware problem.
And yes, I have occasionally run into compiler bugs in my career. That's one reason we test.
> None of that is generated by an LLM
How did you verify that?
> prone to hallucination
You know humans can hallucinate?
> is perfectly deterministic
We agree then that you can verify, test, and trust the deterministic code an LLM produces without ever looking at it.
> That's one reason we test
That's one way we can trust and verify code produced by an LLM. You can't stop doing all the other things that aren't coding.
I get there's a difference. Shitty code can be produced by LLMs or humans. LLMs really can pump out the shitty code. I just think the argument that you cant trust code you haven't viewed is not a good argument. I very much trust a lot of code I've never seen, and yes I've been bitten by it too.
Not trying to be an ass, more trying to figure out how im going to deal for the next decade before retirement age. Uts going to be a lot of testing and verification I guess
"Trust"? God no. That's why I have a debugger
In the past month, in my spare time, I've built:
- A "semantically enhanced" epub-to-markdown converter
- A web-based Markdown reader with integrated LLM reading guide generation (https://i.imgur.com/ledMTXw.png)
- A Zotero plugin for defining/clarifying selected words/sentences in context
- An epub-to-audiobook generator using Pocket TTS
- A Diddy Kong Racing model/texture extractor/viewer (https://i.imgur.com/jiTK8kI.png)
- A slimmed-down phpBB 2 "remake" in Bun.js/TypeScript
- An experimental SQLite extension for defining incremental materialized views
...And many more that are either too tiny, too idiosyncratic, or too day-job to name here. Some of these are one-off utilities, some are toys I'll never touch again, some are part of much bigger projects that I've been struggling to get any work done on, and so on.
I don't blame you for your cynicism, and I'm not blind to all of the criticism of LLMs and LLM code. I've had many times where I feel upset, skeptical, discouraged, and alienated because of these new developments. But also... it's a lot of fun and I can't stop coming up with ideas.
> I’m ready to retire. ... Fast forward decades and Claude Code is giving me that same energy and drive. I love it. It feels like it did back then. I’m chasing the midnight hour and not getting any sleep.
Of course you love it, you don't have to worry about retirement anymore.
Give me your 401k, then tell you feel about Claude Code.
Yes! Although 60 is still a decade away, I've spent a fair few evenings vibe-coding a FOSS dependency-free raw git repo browser.[1] Never would have even started such a project without LLMs because:
* Implementing a raw Git reader is daunting.
* Codifying syntax highlighting rules is laborious.
* Developing a nice UI/UX is not super enjoyable for me.
* Hardening with latest security measures would be tricky.
* Crafting a templating language is time-consuming.
Being able to orchestrate and design the high-level architecture while letting the LLM take care of the details is extremely rewarding. Moving all my repositories away from GitLab, GitHub, and BitBucket to a single repo under my own control is priceless.
[1]: https://repo.autonoma.ca/treetrek/
A great thing you can do with LLMs:
"in (language I'm familiar with) I use (some pattern or whatever) what's the equivalent in (other language)?"
It's really great for doing bits and then get it to explain or you look and see what's wrong and modify it and learn.
This resonates. The emotional side of returning to coding is real.
With Claude Code specifically, I've noticed that the longer it runs autonomously, the more cost anxiety creeps in. You stop thinking about the problem and start watching the token counter.
What finally let me stop worrying and just build again was building a hard budget limit outside the app — not just alerts, but an actual kill switch.
Glad you found the spark. It's worth protecting.
I introduced my dad to claude code. He doesn’t even code, but now it’s a more welcoming and rewarding experience from the get-go. He’s happy, became more comfortable with linux.
Occasionally I remote in to help fix something, but the coding agent really takes a load off my back, and he can start learning without knowing where the endpoints are.
I remember before style sheets existed. Webites were all nested tables and font tags. I built a video website before YouTube be even existed. Claude code and AI is definitely an exciting time.
And transparent 1 pixel gifs :-)
don't forget VRML there are dozens of us
Need to align something. Simple! :)
Opposite here. I was excited by writing code and worked on open source side projects consistently. Somehow, I've not done anything since around August 2025.
I have a sense that AI could have something to do with it.
AI is degrading the status of our profession; its perception in the public eye.
At the same time, it is stealing our work and letting cretins pretend to be software engineers.
It's a bad taste in the mouth.
51 year old electrical engineer here, same thing! (minus the retiring part cause finances)
It's given me the guts to be a solo-founder (for now). I
Just checked out MoveOMeter.com Great idea - and I get how pitching to "an old coot" like my parents would get a laugh out of them before an insulting hurtful pass. Very clever positioning - I'd lean in on that. Your audience is there and waiting - which is tricky since your customer is actually the sales person and you need to give them the training up front to close the deal with their elder. Nice work!
I've always dabbled in electronics, as a hobbyist. I've never had any formal courseware or training in it.
But I have been haranguing Claude/Gemini to help me on an analog computer project for some months now that has sent me on a deep dive into op-amps and other electronics esoterica that I had previously only dabbled a bit in.
Along the way I've learned about relaxation oscillators, using PWM to multiply two voltages, integrating, voltage-following…
I could lean on electronics.stackexchange (where my Google searches often lead) but 1) I first have to know what I am even searching for and 2) even the EEs disagree on how to solve a problem (as you might expect) so I am still with no clear answer. Might as well trust a sometimes hallucinating LLM?
I guess I like the first point above the best—when the LLM just out of the blue (seemingly) suggests a PWM multiplier when I was thinking log/anti-log was the only way to multiply voltages. So I get to learn a new topology.
Or I'm focused on user-adjustable pots for setting machine voltages and the LLM suggests a chip with its own internal 2.45V reference that you can use to get specific voltages without burdening the user to dial it in, own a multimeter. So I get to learn about a chip I was unfamiliar with.
It just goes on an on.
(And, Mr. Eater, I only let the magic smoke out once so far, ha ha.)
It's a lot of fun. I'm also an old timer.
I think it's also somewhat addictive. I wonder if that's part of what's at play here.
A coworker that never argues with you, is happy to do endless toil... sometimes messes up but sometimes blows your mind...
The promise/potential of ever-refining skills and agents drives this compulsion for me. "NEXT time it will be even better. And NOW it's set up to avoid the pitfalls I faced last time." You can feel the exponential engine-building.
I'm not a SWE. I'm a mechanical engineer who spends his life in excel. So when I first made my own node editor app and then asked Claude to read that for my workflow in my second project.... I felt like God herself.
From what I've seen, and of course the models get better everyday, if you have very simple grunt work that needs to be done. Coding agents are basically magic. The moment something gets either difficult or subjective, coding agents love to add completely incorrect solutions.
Try to tell Claude Code to refactor some code and see if it doesn't just delete the entire file and rewrite it. Sure that's cute, but it's absolutely not okay in a real software environment.
I do find this stuff great for hobbyist projects. I don't know if I'd be willing to put money on the line yet
As a parent to two young kids and in more of a leadership position at work, Claude allows me to grind through my backlog of ideas in minutes between other tasks, and see which ones take flight.
As a solo dev, using LLMs for coding has made me a better programmer for sure!
I can ask an LLM for specific help with my codebase and it can explain things in context and provide actual concrete relevant examples that make sense to me.
Then I can ask again for explanations about idiomatic code patterns that aren't familiar for me.
Working on my own, I don't get that feedback and code review loop.
Working with new languages and techniques, or diving into someone else's legacy code base is no longer as daunting with an LLM to ask for help!
I have had the opposite experience.
When it was just asking ChatGPT questions it was fine, I was having fun, I was able to unblock myself when I got non-trivial errors much quicker, and I still felt like I was learning stuff.
With Codex or Claude Code, it feels like I'm stuck LARPing as a middle manager instead of actually solving problems. Sometimes I literally just copy stuff from my assigned ticket into Claude and tell it to do that, I awkwardly wait for a bit, test it out to see if it's good enough, and make my pull request. It's honestly kind of demoralizing.
I suppose this is just the cost of progress; I'm sure there were people that loved raising and breeding horses but that's not an excuse to stop building cars.
I loved being able to figure out interesting solutions to software problems and hacking on them until something worked, and my willingness to do the math beforehand would occasionally give me an edge. Instead, now all I do is sit and wait while I'm cuckolded out of my work, and questioning why I bothered finishing my masters degree if the expectation now is to ship slop code lazily written by AI in a few minutes.
It was a good ride while it lasted; I got almost fifteen years of being paid to do my favorite thing. I should count my blessings that it lasted that long, though I'm a little jealous of people born fifteen years earlier who would be retiring now with their Silicon Valley shares. Instead, I get to sit here contemplating whether or not I can even salvage my career for the next five years (or if I need to make a radical pivot).
Are you 60?
No, I'm in my mid 30's. Unless I win the lottery (which seems unlikely considering I don't buy lottery tickets), or I managed to get some obscenely lucky with shares at a startup, I realistically will need to work for at least twenty more years before retiring.
Same! After years in engineering management I'm building so many small side projects thanks to Claude Code. I'm creating at a breakneck pace. Claude Code has mostly raised the level of abstraction so I can focus much more on the creative aspect of building which has been so much fun.
There are definitely a lot of limitations with Claude Code, but it's fun to work through the issues, figure out Claude's behavior, and create guardrails and workarounds. I do think that a lot of the poor behavior that agents exhibit can be fixed with more guardrails and scaffolding... so I'm looking forward to the future.
Getting claude to build mathematical models for me and running simulations really got me back into doing sciency things too. It's the model that's important, not the boilerplate each time!
I'm 38 years old, and as a manager, it's gradually become difficult to find joy in coding. Claude Code has helped me rediscover that pleasure. Now, all I want to do is code every day and use up my quota.
I've also been loving the speed Claude has enabled me to move at, and now agree that the coding part of SWE has become LLM-wrangling instead. I now see interacting with an LLM, to build all parts of software, as the new "frontend".
Following this idea, what do people think "backend" work will involve? Building and tweaking models, and the infra around them? Obviously everyone will shift more into architecture and strategy, but in terms of hands-on technical work I'm interested in where people see this going.
I’ve been trying to learn a lot about domain driven design, I think knowledge crunching will be a huge part of the new software development role.
Was chatting with a friend about this:
"I used to write java code and the compiler turned it into JVM bytecode.
Now I write in English and the LLMs compile it into whatever language I want."
Although as one HN commenter pointed out: English is a pretty bad programming language as it's way more ambiguous than most programming languages.
The English language has the ability to be ambiguous, but I bet AI use will change the way we use the English language colloquially, to say more specifically what we mean. I worked as a home inspector for a while. Writing for an LLM is very similar to writing a home inspection report or legal brief (or talking to a group of teenagers). Navigate the minefield with very specific intention.
I’m a 13 year lurker, first time commenter (Not sure why this post compelled me). I don’t think this is a genuine take. I don’t see how a 60 year old has any kind of joy for actual software creation suddenly from llms. It might be a joy in seeing software automatically be created but it’s definitely not doing the work. (I may be biased, I left the field 5 years ago) I doubt he’s spending any time fixing the software to make it near usable for anyone besides himself and the semi-working state the llm gave him. Meaning he’s going to have 10 or more half-finished projects again.
I agree. This seems more like an excitement or joy after getting a new toy more that actual process of creating something. Particularly when person uses LLM in a pure vibe code approach where they have no idea what's happening in the code.
Bummer of a first post!
Same, early 50s and this is like the heyday of coding where you could rapidly iterate on things and actively make leaps and bounds of progress. Super fun.
Another +1 from me at 62 years. My problem is this has led to me feeling like I am tech lead for a team of a dozen excellent developers, but I have no task for them!
It's taken over my life, I am in a leadership position at faang but i'm daydreaming about getting back to my claude sessions at work.
Congratulations! Are you still coding VB using Claude? Or something else.
I've heard this from so many greybeards... including me!
“Hell-ya brother”
100% agree even with half your experience.
I'll be 38 next month. I always wonder what I'm do in 30 more years and I cannot see myself NOT coding. Happy to see that spark is alive and well within you.
Glad to see this. I was tired of seeing posts that are on the extremes - "death of software by AI" vs "AI can't do this and that".
I took a break from software, and over the last few years, it just felt repetitive, like I was solving or attempting to solve the same kinds of problems in different ways every 6 months. The feeling of "not a for loop again", "not a tree search again", "not a singleton again". There's an exciting new framework or a language that solves a problem - you learn it - and then there are new problems with the language - and there is a new language to solve that language's problem. And it is necessary, and the engineer in me does understand the why of it, but over time, it just starts to feel insane and like an endless loop. Then you come to an agreement: "Just build something with what I know," but you know so much that you sometimes get stuck in analysis paralysis, and then a shiny new thing catches your engineer or programmer brain. And before you get maintainable traction, I would have spent a lot of time, sometimes quitting even before starting, because it was logistically too much.
Claude Code does make it feel like I am in my early twenties. (I am middle-aged, not in 60s)
I see a lot of comments wondering what is being built -
Think about it like this, and you can try it in a day.
Take an idea of yours, and better if it is yours - not somebody else's - and definitely not AI's. And scope it and ground it first. It should not be like "If I sway my wand, an apple should appear". If you have been in software for long, you would have heard those things. Don't be that vague. You have to have some clarity - "wand sway detection with computer vision", "auto order with X if you want a real apple", etc.. AI is a catalyst and an amplifier, not a cheat code. You can't tell it, "build me code where I have tariffs replacing taxes, and it generates prosperity". You can brainstorm, maybe find solutions, but you can't break math with AI without a rigorous theory. And if you force AI without your own reasoning, it will start throwing BS at you.
There is this idea in your mind, discuss it with ChatGPT, Gemini, or Claude. See the flaws in the idea - discover better ideas. Discuss suggestions for frameworks, accept or argue with AI. In a few minutes, you ask it to provide a Markdown spec. Give it to Claude Code. Start building - not perfect, just start. Focus on the output. Does it look good enough for now? Does it look usable? Does it make sense? Is the output (not code) something you wanted? That is the MVP to yourself. There's a saying - customers don't care about your code, but that doesn't mean you shouldn't. In this case, make yourself the customer first - care about the code later (which in an AI era is like maybe a 30min to an hour later)
And at this point, bring in your engineer brain. Typically, at this point, the initial friction is gone, you have code and something that is working for you in real - not just on a paper or whiteboard. Take a pause. Review, ask it to refactor - make it better or make it align with your way, ask why it made the decisions it made. I always ask AI to write unit tests extensively - most of which I do not even review. The unit tests are there just to keep it predictable when I get involved, or if I ask AI to fix something. Even if you want to remove a file from the project, don't do it yourself - acclimatize to prompting and being vague sometimes. And use git so that you can revert when AI breaks things. From idea to a working thing, within an hour, and maybe 3-4 more hours once you start reviews, refactors, and engineering stuff.
I also use it for iterative trading research. It is just an experiment for now, but it's quite interesting what it can do. I give it a custom backtesting engine to use, and then give it constraints and libraries like technical indicators and custom data indicators it can use (or you could call it skills) - I ask it to program a strategy (not just parameter optimize) - run, test, log, define the next iteration itself, repeat. And I also give it an exact time for when it should stop researching, so it does not eat up all my tokens. It just frees up so much time, where you can just watch the traffic from the window or think about a direction where you want AI to go.
I wanted to incorporate astrological features into some machine learning models. An old idea that I had, but I always got crapped out because of the mythological parts and sometimes mystical parts that didn't make sense. With AI, I could ask it to strip out those unwanted parts, explain them in a physics-first or logic-first way, and get deeper into the "why did they do this calculation", "why they reached this constant", and then AI obviously helps with the code and helps explain how it matches and how it works - helps me pin point the code and the theories. Just a few weeks ago, I implemented/ported an astronomy library in Go (github.com/anupshinde/goeph) to speed up my research - and what do I really know about astronomy! But the outputs are well verified and tested.
But, in my own examples, will I ever let AI unilaterally change the custom backtesting engine code? Never. A single mistake, a single oversight, can cost a lot of real money and wasted time in weeks or months. So the engine code is protected like a fortress. You should be very careful with AI modifying critical parts of your production systems - the bug double-counting in the ledger is not the same as a "notification not shown". I think managers who are blanket-forcing AI on their employees are soon going to realize the importance of the engineering aspect in software
Just like you don't trust just any car manufacturer or just any investment fund, you should not blindly trust the AI-generated code - otherwise, you are setting yourself up to get scammed.
btw how good are any of these tools for embedded programming? we need a new era for hardware enthusiasts. my dad made plenty of fun things in the 80s but it was at the tail end of the newess that came from radiokits and other gadgets that flooded the market due to the uchip
I've never built anything outside of a python notebook before, but Claude Code felt like magic to me.
Same here, 60 and few months and I'm excited about AI
all the insane and/or speculative projects that i never did because they would require heavy lift but with vague outcomes are now in progress. it's glorious.
try asking claude to write in VB6. Make some Active Server Pages. Use COM components. Why not? We can do things "better" now, but what does that matter when you can do the same things as before, but better?
I'm so excited to be able to continue build things when I'm living on the streets. I'm glad to know that drive to create will always be with me and keep me warm during winters.
You can't speak this kind of truth on hacker news!
But, uh, yeah... I've been noticing a growing divide between people like OP who are either already retired or are wealthy enough that they could if they wanted to who absolutely love the new world of LLMs, and people who aren't currently financially secure and realize that LLMs are going to snatch their career away. Maybe not this year, but not too far out either.
I'm enjoying the new era of agentic-coding all your ideas, but it's been obvious to me for a while that jobs are going to tend towards ones where you're liked by the decisionmaker or capital owner and kept around to be the middleman decider-delegator to others/AI/robots.
Have warned my friends about this already.
What I think is lost on ones like OP, is that yes, they are financially secure in the current climate. But if the future that everyone seems to be ushering in does come true, even ones like OP will be in a different state of security.
How does the saying go again? "It takes a village to reach financially secure retirement"
Curious, what are you building?
exactly need some goal here ;)
Re-calibrate your bot
I expect to have at least 15 more years in the workforce and I hate that I have to live through this "revolution". I worry about what will be final balance of lives improved vs lives worsened.
I get hate on only using cli. Glad someone else see's a different perspective
I’m on a field trip chaperoning my kid. I get a couple slack messages asking for some tweaks to a UI. I type a couple words into a Github AI Agent Session while riding the bus. Fixes are deployed to our staging env in 10 minutes.
Fucking wild.
I don't play games anymore. I just work on whacky ideas with LLMs. I even nuked my gaming PC and installed ollama+rocm to play with local models, run openclaw there to experiment with that too. It's a lot of fun. I feel like agents are particularly useful for people who are ADD and want to work on 10 things at once.
I’m on my 40s and building a platform to support my late cognitive decline. Tools that shaped human existence.
I'd like to hear more
Would love to hear more, if you are happy sharing!
I have this idea that probably violates some law of computing but I am really stubborn to make it happen somehow.
I want a game that generates its own mechanics on the fly using AI. Generates itself live.
Infinite game with infinite content. Not like no mans sky where everything is painfully predictable and schematic to a fault. No. Something that generates a whole method of generating. Some kind of ultra flexible communication protocol between engine and AI generator that is trained to program that protocol.
Develop it into a framework.
Use that framework to create one game. A dwarf fortress adventure mode 2.0
I have no other desires, I have no other goals, I don’t care. I or better yet - someone else, must do it.
It sounds doable. An AI can be made to keep modifying a game's codebase. I imagine it'd be easiest to separate out a scripting layer for game mechanics & behavior that AI can iterate quickly on, although of course it could more riskily modify the engine itself.
Then you could open voting up to a community for a weekly mechanics-change vote (similar to that recent repo where public voting decided what the AI would do next), and AI will implement it with whatever changes it sees fit.
Honestly, without some dedicated human guidance and taste, it would probably be more of a novelty that eventually lost its shine.
Building things as I read this.
My main worry is: what is the license on the code produced by Claude (or any other coding agent)? It seems like, if it was trained using open-source software, then the resulting code needs to be open-source as well and it should be compatible with the original source. Artwork produced by an AI cannot be copyrighted, but apparently code can be?
If the software produced is for internal use, the point is probably moot. But if it isn't, this seems like a question that needs to be answered ASAP.
Everything in this post is proof that Anthropic will kill it when they go public. I believe in it, so does everyone else.
I am 37;
Claude Code and it's parallels have extinguished multiple ones.
I was able to steer clear of the Bitcoin/NFT/Passport bros but it turns out they infiltrated the profession and their starry puppy delusional eyes are trying to tell me that iteration X of product Y released yesterday evening is "going to change everything".
They have started redefining what "I have build this" actually means, and they have outjerked the executives by slinging outrageous value creation narratives.
> I’m chasing the midnight hour and not getting any sleep.
You are 60; go spend some time with your grand-kids, smell a flower, touch grass forget chasing anything at this age cause a Tuesday like the others things are gonna wrap up.
Absolutely sincerely.
The ageism in this comment is revolting.
I see many comments here about Claude and I get the same feeling I get when I see comments about MacOS: it's nice that you're content with it, but I don't trust Apple/Anthropic for a fraction of an angstrom.
Wake me when we have ethically trained, open source models that run locally. Preferably high-quality ones.
Perhaps I shouldn't say this but I feel that with the current LLMs I've found "my people" :)
This is the way. It's the most fun computers have been in decades.
I have bipolar disorder. The more frustrating aspects of coding have historically affected me tenfold (sometimes to the point of severe mania). Using Claude Code has been more like an accessibility tool in that regard. I no longer have to do the frustrating bits. Or at the very least, that aspect of the job is thoroughly diminished. And yes - coding is "fun again".
I think coding can be an endurance sport sometimes. There are a lot of points at which you have to bang your head against a wall for hours or days to figure out the smallest issue. Having an agent do that frustrating part definitely lowers the endurance needed to stay productive on a project.
"Just when I thought I was out, they pull me back in"
Older here, equally excited. It's like programming with a team of your best buddies who are smarter than you but humble and eager to collaborate.