The future of software development is software developers

2025-12-2919:14421566codemanship.wordpress.com

I’ve been a computer programmer all-told for 43 years. That’s more than half the entire history of electronic programmable computers. In that time, I’ve seen a lot of things chang…

I’ve been a computer programmer all-told for 43 years. That’s more than half the entire history of electronic programmable computers.

In that time, I’ve seen a lot of things change. But I’ve also seen some things stay pretty much exactly the same.

I’ve lived through several cycles of technology that, at the time, was hailed as the “end of computer programmers”.

WYSIWYG, drag-and-drop editors like Visual Basic and Delphi were going to end the need for programmers.

Wizards and macros in Microsoft Office were going to end the need for programmers.

Executable UML was going to end the need for programmers.

No-Code and Low-Code platforms were going to end the need for programmers.

And now, Large Language Models are, I read on a daily basis, going to end the need for programmers.

These cycles are nothing new. In the 1970s and 1980s, 4GLs and 5GLs were touted as the end of programmers.

And before them, 3GLs like Fortran and COBOL.

And before them, compilers like A-0 were going to end the need for programmers who instructed computers in binary by literally punching holes in cards.

But it goes back even further, if we consider the earliest (classified) beginning of electronic programmable computers. The first of them, COLOSSUS, was programmed by physically rewiring it.

Perhaps the engineers who worked on that machine sneered at the people working on the first stored-program computers for not being “real programmers”.

In every cycle, the predictions have turned out to be very, very wrong. The end result hasn’t been fewer programmers, but more programs and more programmers. It’s a $1.5 trillion-a-year example of Jevons Paradox.

And here we are again, in another cycle.

“But this time it’s different, Jason!”

Yes, it certainly is. Different in scale to previous cycles. I don’t recall seeing the claims about Visual Basic or Executable UML on the covers of national newspapers. I don’t recall seeing entire economies betting on 4GLs.

And there’s another important distinction: in previous cycles, the technology worked reliably. We really could produce working software faster with VB or with Microsoft Access. This is proving not to be the case with LLMs, which – for the majority of teams – actually slow them down while making the software less reliable and less maintainable. It’s a kind of LOSE-LOSE in most cases. (Unless those teams have addressed the real bottlenecks in their development process.)

But all of this is academic. Even if the technology genuinely made a positive difference for more teams, it still wouldn’t mean that we don’t need programmers anymore.

The hard part of computer programming isn’t expressing what we want the machine to do in code. The hard part is turning human thinking – with all its wooliness and ambiguity and contradictions – into computational thinking that is logically precise and unambiguous, and that can then be expressed formally in the syntax of a programming language.

That was the hard part when programmers were punching holes in cards. It was the hard part when they were typing COBOL code. It was the hard part when they were bringing Visual Basic GUIs to life (presumably to track the killer’s IP address). And it’s the hard part when they’re prompting language models to predict plausible-looking Python.

The hard part has always been – and likely will continue to be for many years to come – knowing exactly what to ask for.

Edgar Dijkstra called it nearly 50 years ago: we will never be programming in English, or French, or Spanish. Natural languages have not evolved to be precise enough and unambiguous enough. Semantic ambiguity and language entropy will always defeat this ambition.

And while pretty much anybody can learn to think that way, not everybody’s going to enjoy it, and not everybody’s going to be good at it. The demand for people who do and people who are will always outstrip supply.

Especially if businesses stop hiring and training them for a few years, like they recently have. But these boom-and-bust cycles have also been a regular feature during my career. This one just happens to coincide with a technology hype cycle that presents a convenient excuse.

There’s no credible evidence that “AI” is replacing software developers in significant numbers. A combination of over-hiring during the pandemic, rises in borrowing costs, and a data centre gold rush that’s diverting massive funds away from headcount, are doing the heavy lifting here.

And there’s no reason to believe that “AI” is going to evolve to the point where it can do what human programmers have to do – understand, reason and learn – anytime soon. AGI seem as far away as it’s always been, and the hard part of computer programming really does require general intelligence.

On top of all that, “AI” coding assistants are really nothing like the compilers and code generators of previous cycles. The exact same prompt is very unlikely to produce the exact same computer program. And the code that gets generated is pretty much guaranteed to have issues that a real programmer will need to be able to recognise and address.

When I write code, I’m executing it in my head. My internal model of a program isn’t just syntactic, like an LLM’s is. I’m not just matching patterns and predicting tokens to produce statistically plausible code. I actually understand the code.

Even the C-suite has noticed the correlation of major outages and incidents proceeding grand claims about how much of that company’s code is “AI”-generated.

The folly of many people now claiming that “prompts are the new source code”, and even that entire working systems can be regenerated from the original model inputs, will be revealed to be the nonsense that it is. The problem with getting into a debate with reality is that reality always wins. (And doesn’t even realise it’s in a debate.)

So, no, “AI” isn’t the end of programmers. I’m not even sure, 1-3 years from now, that this current mania won’t have just burned itself out, as the bean counters tot up the final scores. And they always win.

To folks who say this technology isn’t going anywhere, I would remind them of just how expensive these models are to build and what massive losses they’re incurring. Yes, you could carry on using your local instance of some small model distilled from a hyper-scale model trained today. But as the years roll by, you may find not being able to move on from the programming language and library versions it was trained on a tad constraining.

For this reason, I’m skeptical that hyper-scale LLMs have a viable long-term future. They are the Apollo Moon missions of “AI”. In the end, quite probably just not worth it. Maybe we’ll get to visit them in the museums their data centres might become?

The foreseeable future of software development is one where perhaps “AI” – in a much more modest form (e.g., a Java coding assistant built atop a basic language model) – is used to generate prototypes, and maybe for inline completion on production code and those sorts of minor things.

But, when it matters, there will be a software developer at the wheel. And, if Jevons is to be believed, probably even more of us.

Employers, if I were you, I might start hiring now to beat the stampede when everyone wakes up from this fever dream.

And then maybe drop me a line if you’re interested in skilling them up in the technical practices that can dramatically shrink delivery lead times while improving reliability and reducing the cost of change, with or without “AI”. That’s a WIN-WIN-WIN.


Read the original article

Comments

  • By solaire_oa 2025-12-3016:1914 reply

    Most people in this thread are quibbling about the exact degree of utility LLMs provide, which a tedious argument.

    What's more interesting to me is, per the article, the concern regarding everyone who is leaning into LLMs without realizing (or downplaying) the exorbitant, externalized cost. Our current LLM usage is being subsidized to the point of being free by outside investment. One day when the well runs dry, you must be able to either pay the actual cost (barring grand technology breakthroughs), or switch back to non-LLM workflows. I run local LLMs infrequently, and every single prompt makes my beefy PC sounds like a jet engine taking off. It's a great reminder to not become codependent.

    • By PTOB 2025-12-3018:011 reply

      As someone who works on the design and construction of datacenters, I cannot stress enough how apropos this comment is. Even before the first conversation in your IDE starts, the load on national and local government resources, local utility capacity, and roadway infrastructure is enormous. We're all paying whether we're using the tools or not.

      • By compiler-devel 2025-12-3020:253 reply

        Nearly nobody cares about the load on “national and local government resources, local utility capacity, and roadway infrastructure” for any other day-to-day activity. Why should they care about the same for AI which for most people is “out there online” somewhere? Related my, crypto bros worried about electricity usage only so far as its expense went and whether they could move closer to hydro dams.

        • By mcbishop 2025-12-3020:462 reply

          The parent comment's point is that we _should_ care because cheap frontier-model access (that many of us have quickly become hopelessly dependent on) might be temporary.

          • By zcw100 2025-12-3120:263 reply

            It's amazing that anyone that has seen anything in technology in the last 30 years can say, "better be careful. They might stop subsidizing this and then it's gunna get expensive!" is ridiculous. I can buy a 1Tb flash drive for $100. Please, even with every reason to amortize the hardware over the longest horizon possible are only going out 6 years. 64K should be enough for anyone right?

            • By mcbishop 2026-01-023:36

              I think the heavy investor subsidization / speculation makes this different. The high cost of early 1Tb flash drives was largely borne by buyers.

            • By sensanaty 2026-01-1412:58

              Yeah, I can't wait to buy some RAM for my PC! Oh, wait, the AI companies are buying up all the RAM sticks on the planet and driving up their prices to comical highs, surely these beacons of ethics and morality won't do the same with their services that are actively hemorrhaging Billions of dollars, they're providing these services to us out of the goodness of their black hearts and not any kind of monetary incentive after all!

            • By Ma8ee 2025-12-3120:40

              Yes, hardware has become cheaper, but services all enshittify the moment the investors start to ask for some return.

          • By Terretta 2025-12-3115:561 reply

            If expert devs have junior devs to assign code to, that you review and integrate, do they become “hopelessly dependent” on junior devs?

            My experience of expert devs is those who are happy to have extra leverage are not slowed much by having to do it themselves.

            In no cases have I seen experts become “dependent” on the junior devs.

            • By Ma8ee 2025-12-3120:421 reply

              They do quite soon after they have become managers or product owners or “architects”.

              • By LtWorf 2026-01-020:01

                Those were probably senior only in age.

        • By milkytron 2025-12-3022:43

          They should care because they are expensive. If we become dependent on something that is expensive, we have to maintain a certain level of economic productivity to sustain our dependence.

          For AI, once these companies or shareholders start demanding profit, then users will be footing the bill. At this rate, it seems like it'll be expensive without some technological breakthrough as another user mentioned.

          For other things, like roads and public utilities, we have to maintain a certain level of economic productivity to sustain those as well. Roads for example are expensive to maintain. Municipalities, states, and the federal government within the US are in lots of debt associated with roads specifically. This debt may not be a problem now, but it leaves us vulnerable to problems in the future.

        • By PTOB 2025-12-3022:37

          > Nearly nobody cares about ...

          That's an accurate and sad truth about humanity in general, isn't it? We all feel safer and saner if we avoid thinking about how things really are. It's doubly true if our hands are dirty to some extent.

          At the same time, I submit that ignoring the effectiveness of very small contingents of highly motivated people is a common failure mode of humanity in general. Recall that "nearly nobody" also describes "people who are the President of the United States." Observe how that tiny rounding error of humanity is responsible for quite a bit of the way the world goes - for good or ill. Arguably, that level of effectiveness doesn't even require much intelligence.

          > Why should they care about the same for AI which for most people is “out there online” somewhere?

          Well, some will be smart enough to see the problem. Some portion thereof will be wise enough to see a solution. And a portion of those folks will be motivated enough to implement it. That's all that's required. Very simple even if it's not very easy or likely.

    • By Mongoose 2025-12-3018:133 reply

      I always liken it to using Uber in ~2012. It was fun to get around major metro areas for dirt cheap. But then prices rose dramatically over the next decade+ as the company was forced to wean itself off of VC subsidies.

      • By zaphirplane 2026-01-019:03

        It’s common since year dot fir new businesses to compete on price to attract customers and gain market share. It wasn’t invented by uber

      • By dawnerd 2025-12-3018:452 reply

        Same with Airbnb. Oh and Moviepass. Those were the days.

        • By doxeddaily 2025-12-3019:12

          I watched a friend of mine minmax Moviepass so hard. They were so doomed.

        • By XenophileJKO 2025-12-3019:343 reply

          Except none of those cost structures are based primarily on a resource that gets cheaper over time.. a.k.a. compute.

          • By dawnerd 2025-12-3022:29

            Computer isn’t getting cheaper, growth right now is supply constrained to memory which if you haven’t seen the news recently…

          • By coffeebeqn 2025-12-3021:03

            Training is getting exponentially more expensive. And inference isn’t that cheap unless you can do it locally

          • By Ma8ee 2025-12-3120:44

            The energy demand doesn’t decrease.

      • By mkozlows 2025-12-3019:521 reply

        ... and people kept using Uber.

        • By insane_dreamer 2025-12-312:302 reply

          Uber and Lyft put all the taxis out of business and now cost as much as the taxis they displaced

          • By signatoremo 2025-12-3112:021 reply

            Ever notice that even where Uber doesn’t operate most of ride sharing alternatives work pretty much the same way? Go to South Asia, China, Middle East, or South East Asia.

            Consumers pick those services because of what Uber pioneered — trust and convenience. You know exactly how much you pay, you pay everything upfront, you know you are dropped off where you need to be. There are of course exceptions, but exceptions they are.

            Cost maybe the initial selling point but people stick with Uber and similar services despite higher cost, not because they don’t have other options.

            • By zaphirplane 2026-01-019:06

              > because of what Uber pioneered — trust

              I really dislike the retro fitting of history. I’ve read more occurrences of serious SA by uber drivers and zero for normal taxi in the last few years

          • By wojciii 2025-12-3112:45

            Not everywhere. Here the government fucked Uber etc. big time because it required the companies to pay for taxi licenses if I remember correctly. That is if they want to deliver a taxi service.

    • By magic_hamster 2025-12-3018:471 reply

      There's a lot in this comment that doesn't exactly fit.

      First of all, there could be other solutions, such as B2B subsidizing individual user plans, or more fine grained model tiering per cost.

      Also, yes you can get some access for free, but even today the higher tiers of proprietary models is around $200/mo for individual users, which might still be subsidized but is definitely not free, and is quite a chunk of money at $2400 a year!

      I don't know what your setup is at the moment, but it's possible more efficient hardware and stack are available that you're not utilizing. Of course this depends on what models you're trying to run.

      I think that smaller models will become a lot better, and hardware will become more optimized as well. We're starting to see this with NPUs and TPUs.

      All this means running models will cost less, and maybe upgrading the power grid will also reduce cost of energy, making it more affordable.

      I don't see any way that AI will go away because it "hits a wall". We have long passed the point of no return.

      • By nosianu 2025-12-3019:46

        You are looking at it from the individual's PoV, but the OP is using the bird view from high above. It is the total amount of effort deployed today already to provide all the existing AI services, which is enormous. Data centers, electricity, planning/attention (entities focused on AI have less time to work on something else), components (Nvidia shortage, RAM shortage), etc.

        This is not about finance, but about the real economy and how much of it, and/or its growth, is diverted to AI. The real economy is being reshaped, influencing a lot of other sectors independent of AI use itself. AI heavily competes with other uses for many kinds of actual real resources - without having equally much to show for it yet.

        Just an example: https://www.technologyreview.com/2025/05/20/1116327/ai-energ... (it is worth skip-reading it all, the headline on its own is useless)

    • By JeremyNT 2025-12-3016:442 reply

      This is a good point but you can see the price "ceiling" by examining the prices for PCs that can effectively run local models. A DGX Spark is ~$4k (plus power) for example.

      That's not nothing, but it's still not very much to pay compared to e.g. the cost of a FTE.

      • By __float 2025-12-3017:001 reply

        You're counting the cost of running the model, but what about training it? You can't count the compute and data costs at $0.

        • By philipkglass 2025-12-3017:191 reply

          You can assume that already-published open weights models are available at $0, regardless of how much money was sunk into their original development. These models will look increasingly stale over time but most software development doesn't change quickly. If a model can generate capable and up-to-date Python, C++, Java, or Javascript code in 2025 then you can expect it to still be a useful model in 2035 (based on the observation that then-modern code in these languages from 2015 works fine today, even if styles have shifted).

          • By sarchertech 2025-12-3017:332 reply

            >2025-2035

            Depending on other people to maintain backward compatibility so that you can keep coding like it’s 2025 is its own problematic dependency.

            You could certainly do it but it would be limiting. Imagine that you had a model trained on examples from before 2013 and your boss wants you to take over maintenance for a React app.

            • By nikkwong 2025-12-3018:061 reply

              You're all referencing the strange idea in a world where there would be no open-weight coding models trained in the future. Even in a world where VC spending vanished completely, coding models are such a valuable utility that I'm sure at the very least companies/individuals would crowdsource them on a reoccurring basis, keeping them up to date.

              The value of this technology has been established, it's not leaving anytime soon.

              • By sarchertech 2025-12-312:471 reply

                SOTA models cost hundreds of millions to train. I doubt anyone is crowdsourcing that.

                And that’s assuming you already have a lot of the infrastructure in place.

                • By nikkwong 2025-12-316:011 reply

                  I think faang and the like would probably crowdsource it given that they would—according to the hypothesis presented—would only have to do it every few years, and ostensibly are realizing improved developer productivity from them.

                  • By sarchertech 2025-12-3113:58

                    I don’t think the incentive to open source is there for $200 million LLM models the same way it is for frameworks like React.

                    And for closed source LLMs, I’ve yet to see any verifiable metrics that indicate that “productivity” increases are having any external impact—looking at new products released, new games on Steam, new startups founded etc…

                    Certainly not enough to justify bearing the full cost of training and infrastructure.

            • By nuancebydefault 2025-12-3020:47

              2013 was pre-LLM. If devs continue relying on LLMs and their training would stop (which i would find unlikely), still the tools around the LLMs will continue to evolve and new language features will get less attention and would only be used by people who don't like to use LLMs. Then it would be a race of popularity between new language (features) and using LLMs steering 'old' programming languages and APIs. Its not always the best technology that wins, often its the most popular one. You know what happened during the browser wars.

      • By losvedir 2025-12-3018:021 reply

        That can't come anywhere close to running the current SotA models, though.

        • By Imustaskforhelp 2025-12-3019:52

          Most things don't require SotA models.

          But still, right now, you don't have to worry as even these SotA models are subsidized really so much and you can just use them for free on websites and then if you don't even want to type, go use a cheaper model or even a free model with something like opencode even to then act as a mini agent of things

          Usually I just end up it being more focused in a single file which isn't really the best practise but its usually for prototyping purposes anyway so it ends up being really good

          uv scripts are good for python, and I usually create golang single main.go files as well as I feel like it can be a binary, compile fast and cross compilation and still easy and simple so yeah :)

    • By kenforthewin 2025-12-3017:05

      I find the cost discussion to be exceedingly more tedious. This would be a more compelling line of thinking if we didn't have highly effective open-weight models like qwen3-coder, glm 4.7 etc. which allow us to directly measure the cost of running inference with large models without confounding factors like VC money. Regardless of the cost of training, the models that exist right now are cheap and effective enough to push the conversation right back to "quibbling about the exact degree of utility LLMs provide".

    • By robotnikman 2025-12-3020:42

      >I run local LLMs infrequently, and every single prompt makes my beefy PC sounds like a jet engine taking off. It's a great reminder to not become codependent.

      I would try setting the GPU to run at a lower power level. I set my GPU power level to 80% and it becomes much quieter, and only runs maybe 5% slower at most.

      Also I 100% agree with the rest of your comment. We can only power the current growth we are seeing for so long.

    • By 12345hn6789 2025-12-311:191 reply

      LLMs cannot generate coherent sentences

      LLMs writing prose is too robotic

      LLMs output is too dependent on prompts to be interesting

      LLMs take too much RAM to run effectively

      LLMs take too much electricity to run locally

      LLMs work locally but are a bit too slow for my taste

      LLMs output mostly correct code but it isn't applicable to my codebase

      LLMs make tool calls to pull in additional context

      LLMs outputted code works for most developers but not my codebase <---- you are currently here

      • By fuy 2025-12-3110:061 reply

        isn't this template supposed to mean that all the previous considerations are now obsolete?

        • By 12345hn6789 2026-01-0421:09

          I guess you could argue that the standard LLM sentence structure is too robotic but prompting mostly fixes that.

          The rest is no longer true, indeed

    • By akkad33 2025-12-3020:43

      So what'll happen to all these companies building on top of openai license. I don't hear these warnings in professional circles, only online

    • By whazor 2025-12-3018:05

      A competitive coding like devstral 2 runs fast enough to be very helpful: https://www.hardware-corner.net/devstral-2-hardware-requirem...

      The required hardware is fits the budget for a professional developer.

      Putting LLMs in the cloud allows the hardware to be utilised better and to have sharding of big models.

    • By lelele 2025-12-3017:393 reply

      > One day when the well runs dry, you must be able to either pay the actual cost

      What multiple of the current cost do you expect? Currently, GitHub Copilot and ChatGPT for Business cost $19/month and €29/month respectively. Even a 10×–20× increase will still be economically viable in a professional setting if the tools continue to save hours of work.

      • By automatic6131 2025-12-3020:241 reply

        These tools (e.g. Chatgpt pro) lose money at $200/month

        So expect, maybe, $1000 a month? Until your business is dependant on these LLMs. Then they can extract basically all your margin lol

        • By lelele 2025-12-3110:51

          At a $1,000/month price point, wouldn't the economics start favoring buying GPUs and running local LLMs? Even if they're weaker, local models can still cover enough use cases to justify the switch.

      • By mock-possum 2025-12-3018:43

        Claude max is like $100/mo, and if you’re a daily user you’re likely going to need max

      • By cl0ckt0wer 2025-12-3018:11

        Remember when Uber was cheap and showed you the surge pricing multiple?

    • By fnord77 2026-01-0416:38

      The cost is coming down fast. You can get a $2000 desktop machine (AMD 395) that can run effectively chatGPT 3.5 levels of LLMs at over 100 tokens per second.

    • By anthonypasq 2025-12-3019:29

      if you wrote this comment 70 years ago when computers were the size of rooms, it would make a lot of sense, and yet we know how history played out where everyone has a super computer in their pocket.

      for some reason it feels like people are under the assumption that hardware isnt going to improve or something?

    • By Imustaskforhelp 2025-12-3019:48

      Writing my comment on this post, I kind of feel like LLM's are like similar to wordpress/drag and drop tool although its more inconsistent too perhaps not sure

      I 100% share the codependent path too and had written a similar comment some 2-3 days ago but these AI companies which provide are either seriously negative/loss making subsidizing or they are barely net zero. I doubt that it will continue and so the bubble will burst I guess and prices of these will rise perhaps

      We will see perhaps something like google which can feed on advertising can perhaps still provide such subsidies for a longer time but the fact of the matter is that I have no alleigance to any model as some might have and I will simply shift to the cheapest thing which can still provide me / be enough for my queries or prototypes mostly I suppose.

    • By bogzz 2025-12-3016:40

      I am sorry, but this kind of level-headed and realistic take is completely unacceptable on hackernews, and you should be ashamed of yourself. This is not a place for rational discussion when it comes to LLMs.

      LLMs are amazing and they will change everything, and then everything will be changed.

  • By snickerer 2025-12-3010:2115 reply

    After working with agent-LLMs for some years now, I can confirm that they are completely useless for real programming.

    They never helped me solve complex problems with low-level libraries. They can not find nontrivial bugs. They don't get the logic of interwoven layers of abstractions.

    LLMs pretend to do this with big confidence and fail miserably.

    For every problem I need to turn my brain to ON MODE and wake up, the LLM doesn't wake up.

    It surprised me how well it solved another task: I told it to set up a website with some SQL database and scripts behind it. When you click here, show some filtered list there. Worked like a charm. A very solved problem and very simple logic, done a zillion times before. But this saved me a day of writing boilerplate.

    I agree that there is no indication that LLMs will ever cross the border from simple-boilerplate-land to understanding-complex-problems-land.

    • By spicyusername 2025-12-3014:013 reply

          I can confirm that they are completely useless for real programming
      
      And I can confirm, with similar years of experience, that they are not useless.

      Absolutely incredible tools that have saved hours and hours helping me understand large codebases, brainstorm features, and point out gaps in my implementation or understanding.

      I think the main disconnect in the discourse is that there are those pretending they can reliably just write all the software, when anyone using them regularly can clearly see they cannot.

      But that doesn't mean they aren't extremely valuable tools in an engineer's arsenal.

      • By Gud 2025-12-3014:201 reply

        Same. I started coding before hitting puberty, and Im well into my 30s.

        If you know the problem space well, you can let LLMs(I use Claude and ChatGPT) flesh it out.

        • By lelele 2025-12-3017:471 reply

          > I use Claude and ChatGPT

          Both for code? For me, it's Claude only for code. ChatGPT is for general questions.

          • By Gud 2025-12-3018:14

            Yes, I use them in tandem. Generally Claude for coding and ChatGPT when I run out of tokens in Claude.

            I also use ChatGPT to summarise my project. I ask it to generate mark down and PDFs, explaining the core functionality.

      • By englishspot 2025-12-3020:58

        I feel like I have to be strategic with my use of claude code. things like frequently clearing out sessions to minimize context, writing the plan out to a file so that I can review it more effectively myself and even edit it, breaking problems down into consumable chunks, attacking those chunks in separate sessions, etc. it's a lot of prep work I have to do to make the tool thrive. that doesn't mean it's useless, though.

    • By bdcravens 2025-12-3013:04

      "real programming"

      Perhaps you're doing some amazing low-level work, but it feels like you're way overestimating how much of our industry does that. A massive amount of developers show up to work every day and just stitch together frameworks and libraries.

      In many ways, it feels similar to EVs. Just because EVs aren't yet, and may never be, effective to moving massive amounts of cargo in a day with minimal refueling, doesn't mean that they aren't an effective solution for the bulk of drivers who have an average commute of 40 miles a day.

    • By JeremyNT 2025-12-3016:481 reply

      > After working with agent-LLMs for some years now, I can confirm that they are completely useless for real programming

      This is a bit of no-true-scottsman, no? For you "real programming" is "stuff LLMs are bad at," but a lot of us out in the real world are able to effectively extract code that meets the requirements of our day jobs from tossing natural language descriptions into LLMs.

      I actually find the rise of LLM coding depressing and morally problematic (re copyright / ownership / license laundering), and on a personal level I feel a lot of nostalgia for the old ways, but I simply can't levy an "it's useless" argument against this stuff with any seriousness.

      • By nosianu 2025-12-3020:06

        I only use it sparingly thus far, and for small things, but I don't find it depressing at all - but timely.

        All those many, many languages, frameworks, libraries, APIs and there many many iterations, soooo much time lost on minute details. The natural language description, even highly detailed down to being directly algorithmic, is a much better level for me. I have gotten more and more tired of coding, but maybe part of it is too much Javascript and its quickly changing environment and tools, for too many years (not any more though). I have felt that I'm wasting way too much time chasing all those many, many details for quite some time.

        I'm not pro-high-level-programming per se - I started a long time ago with 8 bit assembler and knowing every one of the special registers and RAM cells. I cherish the memories of complex software fitting on a 1.44 MB floppy. But it had gotten just a bit too extreme with all the little things I had to pay attention to that did not contribute to solving the actual (business) problem.

        I feel it's a bit early even if it's already usable, but I hope they can get at least one more giant leap out of AI in the next decade or so. I am quite happy to be able to concentrate on the actual task, instead of the programming environment minutiae, which has exploded in size and complexity across platforms.

    • By perhapsAnLLM 2025-12-3014:172 reply

      "they are completely useless for real programming"

      You and I must have completely different definitions of "real programming". In this very comment, you described a problem that the model solved. The solution may not have involved low-level programming, or discovering a tricky bug entrenched in years-worth of legacy code, but still a legitimate task that you, as a programmer, would've needed to solve otherwise. How is that not "real programming"?

      • By re_chief 2025-12-3014:361 reply

        I wouldn't describe the LLM's actions in the example as "solving a problem" so much as "following a well-established routine". If I were to, for instance, make a PB&J sandwich, I wouldn't say that what I'm doing is "real cooking" even if it might technically fit the definition.

        If an LLM controlling a pair of robot hands was able to make a passable PB&J sandwich on my behalf, I _guess_ that could be useful to me (how much time am I really saving? is it worth the cost? etc.), but that's very different from those same robo-hands filling the role of a chef de cuisine at a fine dining restaurant, or even a cook at a diner.

        • By theshrike79 2025-12-3015:581 reply

          In this analogy you're clearly a private chef with clients who have very specific wishes and allergies.

          The rest of us are just pumping out CRUD-burgers off the API assembly line. Not exactly groundbreaking stuff.

          LLMs are really good with burgers, but not so much being a private chef.

          • By sarchertech 2025-12-3017:501 reply

            Every useful CRUD app becomes its own special snowflake with time and users.

            Now if your CRUD app never gets any users sure it stays generic. But we’ve had low code solutions that solve this problem for decades.

            LLMs are good at stuff that probably should have been low code in the first place, but couldn’t be for reasons. That’s useful, but it comes with a ton of trade offs. And these kind of solutions covet a lot less ground than you’d think.

            • By theshrike79 2025-12-3020:07

              I'm old enough to remember the "OMG low-code is going to take our jeeeerbbs!" panic :D

              Like LLMs they took away a _very_ specific segment of software, Zapier, n8n, NodeRED etc. do some things in a way that bespoke apps can't - but they also hit a massive hard wall where you either need to do some really janky shit or just break out Actual Code to get forward.

      • By nuancebydefault 2025-12-3020:55

        "real programming" hits a "true scottsman" snare with me.

    • By underdeserver 2025-12-3012:362 reply

      People are saying Codex 5.2 fullsolved crypto challenges in 39C3 CTF last weekend.

      Three months ago I would have agreed with you, but anecdotal evidence says Codex 5.2 and Opus 4.5 are finally there.

      • By embedding-shape 2025-12-3013:041 reply

        You'll get a vastly different experience the more you use these tools and learn their limitations and how you can structure things effectively to let them do their job better. But lots of people, understandably, don't take the time to actually sit down and learn it. They spend 30 seconds on some prompt not even a human would understand, and expect the tooling to automatically spend 5 hours trying its hardest at implementing it, then they look at the results and conclude "How could anyone ever be productive with this?!".

        People say a lot of things, and there is a lot of context behind what they're saying that is missing, so then we end up with conversations that basically boil down to one person arguing "I don't understand how anyone cannot see the value in this" with another person thinking "I don't understand how anyone can get any sort of value out of this", both missing the other's perspective.

        • By broast 2025-12-3014:10

          Prompt engineering is just good transfer notes and ticket writing, which is something a majority of the devs I've worked with don't enjoy or excel at

      • By dent9 2025-12-3014:171 reply

        I've been using Codex and Claude Sonnet for many months now for personal (Codex) and work (Sonnet) and I agree. Three months ago these tools were highly usable, now with Codex 5.2 and Sonnet 4.5 I think we're at the point where you can confidently rely on them to analyze your repo codebase and solve, at the very least, small scoped problems and apply any required refactor back throughout the codebase.

        6-12+ months ago the results I was getting with these tools were highly questionable but in the last six months the changes have been pretty astounding

        • By baq 2025-12-3016:052 reply

          Sonnet is dumb as a bag of bricks compared to Opus, perhaps you meant Opus? I never use sonnet for anything anymore, it’s either too verbose or just can’t handle tasks which Opus one shots.

          • By boredtofears 2025-12-3016:53

            These anecdotes feel so worthless. I notice almost no difference between the two and get generally high quality results from either. This is also a worthless anecdote. I'm guessing what kind of codebase you are working in matters a lot as well as the tasks you're giving it.

          • By dent9 2025-12-3016:311 reply

            I use the Copilot extension in VS Code, which links back to my enterprise GitHub account, where I have Claude Sonnet 4.5 available amongst other things. I'm not familiar with Opus. I just open the Copilot Chat window in my VS Code, configure it to use Sonnet 4.5, tell it what I need and it writes the responses and code for me. I'm not using it for large tasks. Most of my usage is "examine this codebase and tell me how to fix xyz problem" or "look at this source code file and show me the code to implement some feature, make sure to examine the entire codebase for insight into how it should be integrated with the rest of the project"

            There's other more advanced coding AI tools but this has accomplished most all of my needs so far

            • By lelele 2025-12-3017:521 reply

              The Copilot extension in VS Code includes Opus as well. It costs three times as much as Claude, so I'd expect it to perform better or be able to handle more complex tasks, but if you're happy with Claude - I am too - more power to you.

              • By lelele 2025-12-319:22

                s/Claude/Sonnet/

    • By solumunus 2025-12-3014:152 reply

      It’s crazy how different my experience is. I think perhaps it’s incredibly important what programming language you are using, what your project and architecture is like. Agents are making an extraordinary contribution to my productivity. If they jacked my Claude Code subscription up to $500/month I would be upset but almost certainly would keep paying it, that’s how much value it brings.

      I’m in enterprise ERP.

      • By merlincorey 2025-12-3019:352 reply

        It sounds like you use your personal Claude Code subscription for work of your employer, but that is not something I would ever consider doing personally so I imagine I must be mistaken.

        Can you elaborate slightly on what you pay for personally and what your employer pays for with regards to using LLMs for Enterprise ERP?

        • By rcbdev 2025-12-3110:281 reply

          Freelancers regularly use tools such as Copilot and Claude, it's always handled professionally and in agreement with their customers. I've seen other freelancers do it plenty of times in the last 1-2 years at my customer sites.

          Why so narrow minded?

          • By merlincorey 2025-12-3120:461 reply

            I'm inquisitive not narrow minded.

            The GP didn't mention anything about freelancing so unless you know them or stalked them you are perhaps being narrow minded here.

            • By rcbdev 2026-01-0121:251 reply

              They also never said anything about being employed.

              You are being narrow minded here.

              • By merlincorey 2026-01-0122:251 reply

                Again, I disagree and reaffirm my being full of inquisitiveness.

                You are being downright unpleasant and I don't think we should continue this conversation further until you open your mind.

        • By solumunus 2026-01-016:581 reply

          I own my own business.

          • By merlincorey 2026-01-0122:29

            Interesting and thanks for clarifying that aspect. I have a few more questions if you would be able to answer any of them at any level of detail I would appreciate it.

            How much would you be willing to pay to continue using Claude on a monthly basis before you stopped?

            Do you currently maintain the new (as of two weeks ago) cash reserve to ensure it continues working when limits are reached and how much do you reserve for said reserve?

            Finally, do you send your customer's code or data directly to Claude or do you use it indirectly on generic stuff and then manually specialize the outputs?

      • By embedding-shape 2025-12-3015:252 reply

        Even more important than those things, is how well you can write and communicate your ideas. If you cannot communicate your ideas so a human could implement it as you wanted it to without asking extra questions, a LLM isn't gonna be able to.

        • By doxeddaily 2025-12-3019:191 reply

          As someone who has managed engineers for many years I find those skills immediately applicable to the LLM domain. If you aren't used to communicating what you are trying to build to other engineers I think using the AI is harder as you need to develop those skills.

          • By embedding-shape 2025-12-3113:48

            I'd take it a step further and say that for any engineer who is used to collaborating with others, engineers or not, should have these skills already, but as most of us know, communication is a generally lacking skill among the population at large, even among engineers too.

        • By onemoresoop 2025-12-3015:47

          Natural language programming has arrived in my opinion. If you're not a developer or have any experience programming it won't help much

    • By fragmede 2025-12-3012:463 reply

      > After working with agent-LLMs for some years now, I can confirm that they are completely useless for real programming.

      "completely useless" and "real programming" are load bearing here. Without a definition to agree on for those terms, it's really hard not to read that as you're trying to troll us by making a controversial unprovable claim that you know will get people that disagree with you riled up. What's especially fun is that you then get to sneer at the abilities of anybody making concrete claims by saying "that's not real programming".

      How tiresome.

      • By sod22 2025-12-3013:222 reply

        Who cares about semantics.

        Ultimately it all boils down to the money - show me the money. OAI have to show money and so do its customers from using this tool.

        But nope, the only thing out there where it matters is hype. Nobody is on an earnings call clearly showing how they had a numerical jump in operating efficiency.

        Until I see that, this technology has a dated shelf life and only those who already generate immense cash flows will fund its continued existence given the unfavourable economics of continued reinvestment where competition is never-ending.

        • By theshrike79 2025-12-3015:501 reply

          The "real programming" people are moving the goalposts of their no true scotsman fallacy so fast they're leaving Roadrunner style dust behind them.

          Yes, there are things LLMs can't do at all, some where they are actively dangerous.

          But also there are decently sized parts of "software development" where any above average LLM can speed up the process as long as whoever is using it knows hot to do so and doesn't fight the tool.

          • By sod22 2025-12-3016:461 reply

            Who cares. Focus on what matters. OAI knows this considering they are dedicating a lot of their resources toward figuring out how to become profitable.

            • By 8note 2025-12-3020:28

              isnt OAI only unprofitable because they are putting all their money into more training?

              the product market fit for LLMs is already clearly found, there's just no moat to it. tokens are a commodity

        • By beeboop0 2025-12-3018:00

          maybe they'll ask cut off the free tiers in 2026 and the only thing left will be China and open router

      • By mrwrong 2025-12-3013:39

        agreed. we should instead be sneering at the AI critics because "you're holding it wrong"

    • By furyofantares 2025-12-3020:18

      > After working with agent-LLMs for some years now

      Some years? I don't remember any agents being any good at all before just over 1 year ago with Cursor and stuff really didn't take off until Claude Code.

      Which isn't to say you weren't working with agent-LLMs before that, but I just don't know how relevant anything but recent experience is.

    • By bwfan123 2025-12-3017:25

      > I can confirm that they are completely useless for real programming

      Can you elaborate on "real programming" ?

      I am assuming you mean the simplest hard problem that is solved. The value of the work is measured in those terms. Easy problems have boilerplate solutions and have been solved numerous times in the past. LLMs excel here.

      Hard problems require intricate woven layers of logic and abstraction, and LLMs still struggle since they do not have causal models. The value however is in the solution of these kinds of problems since the easy problems are assumed to be solved already.

    • By lr4444lr 2025-12-3016:28

      > After working with agent-LLMs for some years now, I can confirm that they are completely useless for real programming. > They never helped me solve complex problems with low-level libraries. They can not find nontrivial bugs. They don't get the logic of interwoven layers of abstractions.

      This was how I felt until about 18 months ago.

      Can you give a single, precise example where modern day LLMs fail as woefully as you describe?

    • By beeboop0 2025-12-3017:55

      i had to disable baby Ceph (Deepseek 3.1) from writing changes in Continue because he's like a toddler. But, he did confirm some solutions and wrote a routine and turn me on to some libraries, etc

      so I see what you're saying. he comes up with the wrong answers a lot to a problem involving a group of classes in related files

      however it's Continue, so it can read files in vs code which is really nice and that helps a lot with its comprehension so sometimes it does find the issue or at least the nature of the issue

      I tend to give it bug n-1 to pre digest while I work on bug n

    • By constantcrying 2025-12-3011:501 reply

      >After working with agent-LLMs for some years now, I can confirm that they are completely useless for real programming.

      >They never helped me solve complex problems with low-level libraries. They can not find nontrivial bugs. They don't get the logic of interwoven layers of abstractions.

      >LLMs pretend to do this with big confidence and fail miserably.

      This is true for most developers as well. The mean software developer, especially if you outsource, has failure modes worse than any LLM and round-trip time is not seconds but days.

      The promise of LLMs is not that they solve the single most difficult tasks for you instantly, but that they do the easy stuff well enough that they replace offshore teams.

      • By shafyy 2025-12-3011:552 reply

        > The promise of LLMs is not that they solve the single most difficult tasks for you instantly, but that they do the easy stuff well enough that they replace offshore teams.

        But that's exactly the *promise* of LLMs by the hypepeople behind it.

        • By constantcrying 2025-12-3011:582 reply

          >But that's exactly the promise of LLMs by the hypepeople behind it.

          I do not know and do not care what the "hypepeople" say. I can tell you that, by pure logic alone, LLMs will be superior at simple and routine tasks sooner, which means they will compete with outsourced labor first.

          LLMs need to be measured against their competition and their competition right now is outsourced labor. If an LLM can outperform an offshore team at a fraction of the cost, why would any company choose the offshore team? Especially when the LLM eliminates some of the biggest problems with offshore teams (communication barriers, round trip times).

          If LLMs take any programmer jobs they will at the very beginning make those outsourced jobs obsolete, so the only relevant question is whether they have done that or are in the process of doing so. If they don't, then their impact will be minimal, if they do, then their impact will be massive. I think that this line of thinking is a far better benchmark then asking whether an LLM gets X or Y question wrong Z% of the time.

          • By kvemkon 2025-12-3013:061 reply

            > If an LLM can outperform an offshore team at a fraction of the cost,..

            And "a few moments later" happens the same as with those "cost effective" clouds.

            [1] https://www.heise.de/en/news/IDC-Many-companies-want-partly-...

            [2] https://www.idc.com/resource-center/blog/storm-clouds-ahead-... (original)

            • By andrekandre 2025-12-3018:19

              in the end, it all comes down to roi; if spending x dollars a month brings in an additional 5x revenue then its gonna be worth?

              then again, i have some suspicion that alot of consumer-focused end products using llms in the backend (hello chatbots) expecting big returns for all those tokens spent may have some bad news coming... if the bubble starts popping i'm guessing it starts there...

          • By baq 2025-12-3016:082 reply

            Outsourced devs wielding smart models are even cheaper than onshore and the models lift all boats wrt capability.

            The bottleneck will soon be ideas for the things to build.

            • By shafyy 2025-12-3117:28

              > The bottleneck will soon be ideas for the things to build

              No, it won't. The utility of LLMs is already growing asymptotically now...

            • By constantcrying 2025-12-3016:27

              >Outsourced devs wielding smart models are even cheaper than onshore

              But they do not compete. They have totally different jobs.

        • By theshrike79 2025-12-3015:471 reply

          I bet you trusted the Blockchain bros and were first in line to buy NFTs too. No?

          Why would you trust the hype when you can verify this stuff yourself pretty easily.

          • By shafyy 2025-12-3117:26

            Obviously I am calling that promise bullshit...

    • By wiz21c 2025-12-3016:19

      Claude is currently porting my rust emulator to WASM. It's not easy at all, it struggles, I need to guide it quite a lot but it's way easier to let him do it than me learning yet another tech. For the same result I have 50% the mental load...

    • By dawnerd 2025-12-3018:49

      The idea they're good for development is propped up a lot by people able to have a react + tailwind site spun up fast. You know what also used to be able to scaffold projects quickly? The old init scripts and generators!

  • By mohsen1 2025-12-2920:5632 reply

    I really really want this to be true. I want to be relevant. I don’t know what to do if all those predictions are true and there is no need (or very little need) for programmers anymore.

    But something tells me “this time is different” is different this time for real.

    Coding AIs design software better than me, review code better than me, find hard-to-find bugs better than me, plan long-running projects better than me, make decisions based on research, literature, and also the state of our projects better than me. I’m basically just the conductor of all those processes.

    Oh, and don't ask about coding. If you use AI for tasks above, as a result you'll get very well defined coding task definitions which an AI would ace.

    I’m still hired, but I feel like I’m doing the work of an entire org that used to need twenty engineers.

    From where I’m standing, it’s scary.

    • By dataviz1000 2025-12-2922:033 reply

      I was a chef in Michelin-starred restaurants for 11 years. One of my favorite positions was washing dishes. The goal was always to keep the machine running on its 5-minute cycle. It was about getting the dishes into racks, rinsing them, and having them ready and waiting for the previous cycle to end—so you could push them into the machine immediately—then getting them dried and put away after the cycle, making sure the quality was there and no spot was missed. If the machine stopped, the goal was to get another batch into it, putting everything else on hold. Keeping the machine running was the only way to prevent dishes from piling up, which would end with the towers falling over and breaking plates. This work requires moving lightning fast with dexterity.

      AI coding agents are analogous to the machine. My job is to get the prompts written, and to do quality control and housekeeping after it runs a cycle. Nonetheless, like all automation, humans are still needed... for now.

      • By conductr 2025-12-305:00

        If it requires an expert engineer/dishwasher to keep the flow running perfectly, the human is the bottleneck in the process. This sounds a lot more like the past before AI to me. What AI does is just give you enough dishes that they don’t need to be washed at all during dinner service. Just let them pile up dirty or throw them away and get new dishes tomorrow it’s so immaterial to replace that washing them doesn’t always make sense. But if for some reason you do want to reuse them, then, it washes and dries them for you too. You just look over things at the end and make sure they pass your quality standards. If they left some muck on a plate or lipstick on a cup, just tell it not to let that happen again and it won’t. So even your QC work gets easier over time. The labor needed to deal with dirty dishes is drastically reduced in any case.

      • By leptons 2025-12-309:28

        > humans are still needed... for now

        "AI" doesn't have a clue what to do on its own. Humans will always be in the loop, because they have goals, while the AI is designed to placate and not create.

        The amount of "AI" garbage I have to sift through to find one single gem is about the same or more work than if I had just coded it myself. Add to that the frustration of dealing with a compulsive liar, and it's just a fucking awful experience for anyone that actually can code.

      • By potamic 2025-12-309:511 reply

        Humans are still needed, but they just got down-skilled.

        • By chii 2025-12-3010:20

          > got down-skilled.

          who's to say that it's a down?

          Orchestrating and doing higher level strategic planning, such that the sub-tasks can be AI produced, is a skill that might be higher than programming.

    • By bigstrat2003 2025-12-308:233 reply

      > Coding AIs design software better than me, review code better than me, find hard-to-find bugs better than me, plan long-running projects better than me, make decisions based on research, literature, and also the state of our projects better than me.

      That is just not true, assuming you have a modicum of competence (which I assume you do). AIs suck at all these tasks; they are not even as good as an inexperienced human.

      • By embedding-shape 2025-12-3010:281 reply

        For all we know, you both could comparing using a Nokia 3310 and a workstation PC based on the hardware, but you both just say "this computer is better than that computer".

        There are a ton of models out there, ran in a ton of different ways, that can be used in different ways with different harnesses, and people use different workflows. There is just so many variables involved, that I don't think it's neither fair nor accurate for anyone to claim "This is obviously better" or "This is obviously impossible".

        I've been in situations where I hit my head against some hard to find bug for days, then I put "AI" (but what? No one knows) to it and it solves it in 20 minutes. I've also asked "AI" to do trivial work that it still somehow fucked up, even if I could probably have asked a non-programmer friend to do it and they'd be able to.

        The variance is great, and the fact that system/developer/user prompts matter a lot for what the responses you get, makes it even harder to fairly compare things like this without having the actual chat logs in front of you.

        • By mrwrong 2025-12-3013:422 reply

          > The variance is great

          this strikes me as a very important thing to reflect on. when the automobile was invented, was the apparent benefit so incredibly variable?

          • By embedding-shape 2025-12-3015:571 reply

            > was the apparent benefit so incredibly variable?

            Yes, lots of people were very vocally against horseless-carriages, as they were called at the time. Safety and public nuisance concerns were widespread, the cars were very noisy, fast, smoky and unreliable. Old newspapers are filled with opinions about this, from people being afraid of horseless-carriages spooking other's horses and so on. The UK restricted the adoption of cars at one point, and some Canton in Switzerland even banned cars for a couple of decades.

            Horseless-carriages was commonly ridiculed for being just for "reckless rich hobbyists" and similar.

            I think the major difference is that cars produced immediate, visible externalities, so it was easy for opposition to focus on public safety in public spaces. In contrast, AI has less physically visible externalities, although they are as important, or maybe even more important, than the ones cars introduced.

            • By mrwrong 2025-12-3111:261 reply

              yeah I agree about the negative externalities but I'm curious about the perceived benefits. did anybody argue that cars were actually slower than horse and carriage? (were they at first?)

              • By embedding-shape 2025-12-3113:541 reply

                The cars were obviously faster than the typical horse transportation and I don't think anyone tried to argue against that, but laws typically restricted cars so they couldn't go faster than horses, at least in highly populated areas like cities. As others mentioned too, the benefit of not needing roads to go places were highlighted as a drawback of cars too. People argued that while cars might go faster, the result would be that the world would be worse off in total.

                • By mrwrong 2026-01-0216:34

                  sure but my point is people could agree they were faster at least. that is decidedly not true for LLMs. maybe due to alignable vs non-alignable differences

          • By throw4847285 2025-12-3014:221 reply

            Is this a trick question? Yes it was. A horse could go over any terrain while a car could only really go over very specific terrain designed for it. We had to terraform the world in order to make the automobile so beneficial. And it turned out that this terraforming had many unintended consequences. It's actually a pretty apt comparison to LLMs.

            • By mrwrong 2025-12-3111:28

              who would I be trying to trick if it was? you didn't answer the question anyways. I'm not wondering whether cars were seen as strictly better than horses in all situations. I'm wondering if people disagreed so vehemently about whether cars were faster road transportation than horses

      • By gentooflux 2025-12-309:241 reply

        LLMs generate the most likely code given the problem they're presented and everything they've been trained on, they don't actually understand how (or even if) it works. I only ever get away with that when I'm writing a parser.

        • By chii 2025-12-3010:183 reply

          > they don't actually understand how

          but if it empirically works, does it matter if the "intelligence" doesn't "understand" it?

          Does a chess engine "understand" the moves it makes?

          • By goatlover 2025-12-3011:102 reply

            It matters if AGI is the goal. If it remains a tool to make workers more productive, then it doesn't need to truly understand, since the humans using the tools understand. I'm of the opinion AI should have stood for Augmented (Human) Intelligence outside of science fiction. I believe that's what early pioneers like Douglas Engalbert thought. Clearly that's what Steve Jobs and Alan Kay thought computing was for.

            • By victorbjorklund 2025-12-3011:59

              AGI is such a meaningless concept. We can’t even fully design what human intelligence is (and when a human fails it meaning they lack human intelligence). It’s just philosophy.

            • By theshrike79 2025-12-3016:09

              AGI is about as well defined as "full self-driving" :D

              It's an useless philosophical discussion.

          • By gentooflux 2025-12-3012:332 reply

            If it empirically works, then sure. If instead every single solution it provides beyond a few trivial lines falls somewhere between "just a little bit off" and "relies entirely on core library functionality that doesn't actually exist" then I'd say it does matter and it's only slightly better than an opaque box that spouts random nonsense (which will soon include ads).

            • By simonw 2025-12-3012:591 reply

              Those are 2024-era criticisms of LLMs for code.

              Late 2025 models very rarely hallucinate nonexistent core library functionality - and they run inside coding agent harnesses so if they DO they notice that the code doesn't work and fix it.

              • By mrwrong 2025-12-3013:43

                get ready to tick those numbers over to 2026!

            • By theshrike79 2025-12-3016:10

              This sounds like you're copy-pasting code from ChatGPT's web interface, which is very 2024.

              Agentic LLMs will notice if something is crap and won't compile and will retry, use the tools they have available to figure out what's the correct way, edit and retry again.

          • By jvanderbot 2025-12-3010:41

            This is a semantic dead end when discussing results and career choices

      • By lelanthran 2025-12-3010:41

        Depends on how he defined "better". If he uses the word "better" to mean "good enough to not fail immediately, and done in 1/10th of the time", then he's correct.

    • By foxygen 2025-12-2922:347 reply

      I think I've been using AI wrong. I can't understand testimonies like this. Most times I try to use AI for a task, it is a shitshow, and I have to rewrite everything anyway.

      • By qweiopqweiop 2025-12-309:451 reply

        Have you tried Opus 4.5 (or similar recent models)? With Claude code 2, it's actually harder to mess things up IMO

        • By techblueberry 2025-12-3012:242 reply

          I remember when about a year ago people were asking the same thing about gpt-4.5, the answer is always “yes, I’ve tried them all”

          • By meindnoch 2025-12-3013:27

            Ok, but have you tried claude-sonnet-GPT-codex-4.5-thinking-fast? That's the game changer. Anyone saying bad things about vibe coding without trying claude-sonnet-GPT-codex-4.5-thinking-fast is like a dinosaur to me, doomed to extinction. Seriously, give claude-sonnet-GPT-codex-4.5-thinking-fast a try, you'll thank me ;)

          • By qweiopqweiop 2025-12-3016:01

            Fair. Well personally they didn't work well for me (on a huge, complex codebase) until the latest batch. Now they do.

      • By weakfish 2025-12-2922:42

        Same. Seems to be the never ending theme of AI.

      • By threethirtytwo 2025-12-307:091 reply

        Try Claude. And partner with it on building something complex.

        • By dent9 2025-12-3015:43

          Yes you want Kiro which uses Claude models under the hood

      • By doug_durham 2025-12-2922:402 reply

        I don’t know about right/wrong. You need to use the tools that make you productive. I personally find that in my work there are dozens of little scripts or helper functions that accelerate my work. However I usually don’t write them because I don’t have the time. AI can generate these little scripts very consistently. That accelerates my work. Perhaps just start simple.

        • By JSDave 2025-12-301:03

          Instead of generating, exporting or copy pasting just seems more reliable to me and also takes very little time.

          I think what matters most is just what you're working on. It's great for crud or working with public APIs with lots of examples.

          For everything else, AI has been a net loss for me.

        • By 63stack 2025-12-3023:35

          > there are dozens of little scripts or helper functions that accelerate my work. However I usually don’t write them because I don’t have the time

          People who write things like this can't expect to be taken seriously.

          Before AI you didn't have time to write things that saved you time? So you just ended up spending (wasting) more time by going the long way? That was a better choice than just doing the thing that would have saved you time?

      • By CuriouslyC 2025-12-305:533 reply

        Do you tell AI the patterns/tools/architecture you want? Telling agents to "build me XYZ, make it gud!" is likely to precede a mess, telling it to build a modular monolith using your library/tool list, your preferred folder structure, other patterns/algorithms you use, etc will end you up with something that might have some minor style issues or not be perfectly canonical, but will be approximately correct within a reasonable margin, or is within 1-2 turns of being so.

        You have to let go of the code looking exactly a certain way, but having code _work_ a certain way at a coarse level is doable and fairly easy.

        • By dent9 2025-12-3015:421 reply

          We are way beyond this. Now you use your plain text prompt to generate a requirements spec that the AI will follow when implementing your project

          https://kiro.dev/

          • By CuriouslyC 2025-12-3015:551 reply

            Kiro is just trying to build a product around exactly what I'm talking about. I'm not a fan, because it's simultaneously too heavyweight and agents don't respect all the details of the specs it creates enough to make the time investment in super-detailed specs worthwhile.

            I have a spec driven development tool I've been working on that generates structured specs that can be used to do automatic code generation. This is both faster and more robust.

            • By dent9 2025-12-3016:20

              That sounds cool, please do share your tools when they're ready :)

        • By mkozlows 2025-12-307:161 reply

          Honestly, even this isn't really true anymore. With Opus 4.5 and 5.2 Codex in tools like Cursor, Claude Code, or Codex CLI, "just do the thing" is a viable strategy for a shockingly large category of tasks.

          • By CuriouslyC 2025-12-3014:451 reply

            Just do the thing can produce functional code, but even with Opus4.5/Codex5.2, there are still plenty of moments where the way it decides to do something is cringe.

            • By mkozlows 2025-12-3015:38

              Agree. But it's increasingly the case, IME, that for a a lot of tasks, you can start with that. If it does it well, great. If it does something stupid, it's easy enough to ask it to completely rework the stupid thing in a better way, and it can do it quickly. That's still a huge shift compared to the olden days (three months ago) where you needed to really break things down into small chunks for it to get to a success state.

        • By leptons 2025-12-309:311 reply

          >You have to let go of the code looking exactly a certain way, but having code _work_ a certain way at a coarse level is doable and fairly easy.

          So all that bullshit about "code smells" was nonsense.

          • By CuriouslyC 2025-12-3015:381 reply

            A lot of code smells matter more for humans than LLMs (and LLMs have their own unique code smells). For example, nested ternary operators are a great source of bugs in human code, but agents could care less, but humans handle multiple files with the same variable names and lots of duplicated code well, whereas this stuff confuses agents.

            • By leptons 2025-12-3017:44

              >but agents could care less,

              The phrase is "couldn't care less". If you "could care less" then you actually care about it. If you "couldn't care less" then there's no caring at all.

      • By mrwrong 2025-12-3013:441 reply

        have you tried using $NEWEST_MODEL ?

        • By threethirtytwo 2025-12-3014:181 reply

          It’s because depending on the person the newest model crossed the line into being useful for them personally. It’s not like a new version crosses the line for everyone. It happens gradually. Each version more and more people come into the fold.

          For me Claude code changed the game.

          • By mrwrong 2025-12-3111:301 reply

            yes, it is trivially true that each new person who recommends LLMs is a new person coming into the fold

            • By threethirtytwo 2025-12-3112:14

              You get new people recommending the latest version all the time to people who are unconvinced because that version is usually what brought them into the fold.

              What you’re mocking is somewhat of a signal of actual improvement of the models and that improvement as a result becoming useful to more and more people.

      • By bdangubic 2025-12-2922:44

        how much time/effort have you put in to educate yourself about how they work, what they excel at, what they suck at, what is your responsibility when you use them…? this effort is directly proportional to how well they will serve you

    • By belter 2025-12-2921:174 reply

      >> From where I’m standing, it’s scary.

      You are being fooled by randomness [1]

      Not because the models are random, but because you are mistaking a massive combinatorial search over seen patterns for genuine reasoning. Taleb point was about confusing luck for skill. Dont confuse interpolation for understanding.

      You can read a Rust book after years of Java, then go build software for an industry that did not exist when you started. Ask any LLM to write a driver for hardware that shipped last month, or model a regulatory framework that just passed... It will confidently hallucinate. You will figure it out. That is the difference between pattern matching and understanding.

      [1] https://en.wikipedia.org/wiki/Fooled_by_Randomness

      • By Verdex 2025-12-2921:29

        I've worked with a lot of interns, fresh outs from college, overseas lowest bidders, and mediocre engineers who gave up years ago. All over the course of a ~20 year career.

        Not once in all that time has anyone PRed and merged my completely unrelated and unfinished branch into main. Except a few weeks ago. By someone who was using the LLM to make PRs.

        He didn't understand when I asked him about it and was baffled as to how it happened.

        Really annoying, but I got significantly less concerned about the future of human software engineering after that.

      • By joefourier 2025-12-2921:272 reply

        Have you used an LLM specifically trained for tool calling, in Claude Code, Cursor or Aider?

        They’re capable of looking up documentation, correcting their errors by compiling and running tests, and when coupled with a linter, hallucinations are a non issue.

        I don’t really think it’s possible to dismiss a model that’s been trained with reinforcement learning for both reasoning and tool usage as only doing pattern matching. They’re not at all the same beasts as the old style of LLMs based purely on next token prediction of massive scrapes of web data (with some fine tuning on Q&A pairs and RLHF to pick the best answers).

        • By treespace8 2025-12-2921:432 reply

          I'm using Claude code to help me learn Godot game programming.

          One interesting thing is that Claude will not tell me if I'm following the wrong path. It will just make the requested change to the best of its ability.

          For example a Tower Defence game I'm making I wanted to keep turret position state in an AStarGrid2D. It produced code to do this, but became harder and harder to follow as I went on. It's only after watching more tutorials I figured out I was asking for the wrong thing. (TileMapLayer is a much better choice)

          LLMs still suffer from Garbage in Garbage out.

          • By jennyholzer3 2025-12-2922:41

            don't use LLMs for Godot game programming.

            edit: Major engine changes have occurred after the models were trained, so you will often be given code that refers to nonexistent constants and functions and which is not aware of useful new features.

          • By memoriuaysj 2025-12-2921:513 reply

            before coding I just ask the model "what are the best practices in this industry to solve this problem? what tools/libraries/approaches people use?

            after coding I ask it "review the code, do you see any for which there are common libraries implementing it? are there ways to make it more idiomatic?"

            you can also ask it "this is an idea on how to solve it that somebody told me, what do you think about it, are there better ways?"

            • By hansmayer 2025-12-2922:38

              > before coding I just ask the model "what are the best practices in this industry to solve this problem? what tools/libraries/approaches people use?

              Just for the fun of it, and so you lose your "virginity" so to speak, next time when the magic machine gives you the answer about "what it thinks", tell it its wrong in a strict language and scold it for misleading you. Tell it to give you the "real" best practices instead of what it spat out. Then sit back and marvel at the machine saying you were right and that it had mislead you. Producing a completely, somewhat, or slightly different answer (you never know what you get on the slot machine).

            • By manmal 2025-12-2922:27

              Both the before and after are better done manually. What you are describing is fine for the heck of it (I‘ve vibe coded a whisper related rust port today without having any actual rust skills), but I’d never use fully vibed software in production. That’s irresponsible in multiple ways.

            • By skydhash 2025-12-2922:00

              Do you also light candles and chant?

        • By belter 2025-12-2922:065 reply

          Ask a model to

          "Write a chess engine where pawns move backward and kings can jump like nights"

          It will keep slipping back into real chess rules. It learned chess, it did not understand the concept of "rules"

          Or

          Ask it to reverse a made up word like

          "Reverse the string 'glorbix'"

          It will get it wrong on the first try. You would not fail.

          Or even better ask it to...

          "Use the dxastgraphx library to build a DAG scheduler."

          dxastgraphx is a non existing library...

          Marvel at the results...tried in both Claude and ChatGPT....

          • By manmal 2025-12-2922:30

            I‘ve just tried the dxastgraphx one in pi with Opus 4.5. This was its response:

                I couldn't find a library called dxastgraphx in either pip (Python) or npm (JavaScript) package registries. This library doesn't appear to exist.
            
                Did you perhaps mean one of these popular DAG/graph libraries?
            
                Python:
                - networkx - comprehensive graph library with DAG support
                - graphlib - Python standard library (3.9+) with TopologicalSorter
                - dask - parallel computing with DAG task scheduling
            
                JavaScript/TypeScript:
                - graphlib - graph data structures
                - dagre - DAG layout
            
                Would you like me to build a DAG scheduler using one of these existing libraries, or would you like me to implement one from scratch? Let me know which language and approach you prefer.

          • By bossyTeacher 2025-12-3113:56

            Tried ChatGPT (free version). I answered correctly to the first question.

            Answer to second question:

            "I can do that, but there’s a catch: dxastgraphx is not a known or standard Python (or JS) library as of now. I don’t have any public API or documentation for it to target directly.

            To avoid guessing wrong, here are the two sensible options:"

          • By somebodythere 2025-12-3012:50

            Why would I ask the model to reverse the string 'glorbix,' especially in the context of software engineering?

          • By criticalfault 2025-12-2922:301 reply

            just tried to reverse the string you provided using Gemini. it worked fine on the first try

            • By eithed 2025-12-2922:531 reply

              In gemini:

              > The reverse of the string 'gssssyeekxxuoaqaaaiooo9aaa7zjx' is:

              xjz7aaa9oooiiaaaqauxxkyeessssg

              > Youve made a mistake

              ``` I see where I tripped up! Let's re-examine that long string carefully.

              The original string you gave me was: `gssssyeekxxuoaqaaaiooo9aaa7zjx`

              Let's look at the end of that string: `...aaa7zjx`. When we reverse it, it should start with: `xjz7aaa...`

              Here is the precise character-by-character reversal:

              *xjz7aaa9oooiiaaaqauxxkyeessssg*

              Wait—I see it now. In my previous response, I included an extra *'i'* and missed one *'o'* in the middle section. Let's do a strict count to get it perfect:

              * *Original:* `g` `ssss` `y` `ee` `k` `xx` `u` `o` `a` `q` `aaa` *`iooo`* `9` `aaa` `7` `z` `j` `x` * *Correct Reverse:* `x` `j` `z` `7` `aaa` `9` *`oooi`* `aaa` `q` `a` `o` `u` `xx` `k` `ee` `y` `ssss` `g`

              The correct reversed string is: *xjz7aaa9oooiiaaaqauoxxkyeessssg*

              Thank you for catching that. It’s easy for the eyes to glaze over when there are so many repeating vowels!

              Since I've stumbled a bit on these, would you like to give me one more string to see if I can get it right on the first try? ```

              After more back and fors it consistently fails in this task, even though when strictly dividing the tokens it will get this right. Yet the final answer is always wrong.

              • By knollimar 2025-12-3015:49

                Mine said it used python and got: xjz7aaa9oooiaaaqaouxxkeeyssssg

          • By baq 2025-12-308:311 reply

            You’re trying to interrogate a machine as you would a human and presenting this as evidence that machines aren’t humans. Yes, you’re absolutely right! And also completely missing the point.

            • By belter 2025-12-312:00

              The discussion is not about being human. Is about being fit for purpose...

      • By doug_durham 2025-12-2922:54

        Why would you expect an LLM or even a human to succeed in these cases? “Write a piece of code for a specification that you can’t possibly know about?” That’s why you have to do context engineering, just like you’d provide a reference to a new document to an engineer writing code.

      • By germandiago 2025-12-309:14

        This is exactly what happened to me: novel or uncommon = hallucinate or invent wrong.

        It is ok for getting snippets for example and saying (I did it). Please make this MVVM style. It is not perfect, but saves time.

        For very broad or novel reasoning, as of today... forget it.

    • By btbuildem 2025-12-2921:443 reply

      They do all those things you've mentioned more efficiently than most of us, but they fall woefully short as soon as novelty is required. Creativity is not in their repertoire. So if you're banging out the same type of thing over and over again, yes, they will make that work light and then scarce. But if you need to create something niche, something one-off, something new, they'll slip off the bleeding edge into the comfortable valley of the familiar at every step.

      I choose to look at it as an opportunity to spend more time on the interesting problems, and work at a higher level. We used to worry about pointers and memory allocation. Now we will worry less and less about how the code is written and more about the result it built.

      • By keyle 2025-12-2921:572 reply

        Take food for example. We don't eat food made by computers even though they're capable of making it from start to finish.

        Sure we eat carrots probably assisted by machines, but we are not eating dishes like protein bars all day every day.

        Our food is still better enjoyed when made by a chef.

        Software engineering will be the same. No one will want to use software made by a machine all day every day. There are differences in the execution and implementation.

        No one will want to read books entirely dreamed up by AI. Subtle parts of the books make us feel something only a human could have put right there right then.

        No one will want to see movies entirely made by AI.

        The list goes on.

        But you might say "software is different". Yes but no, in the abundance of choice, when there will be a ton of choice for a type of software due to the productivity increase, choice will become more prominent and the human driven software will win.

        Even today we pick the best terminal emulation software because we notice the difference between exquisitely crafted and bloated cruft.

        • By doug_durham 2025-12-2922:513 reply

          You should look at other engineering disciplines. How many highway over passes have unique “chef quality” designs? Very few. Most engineering is commodity replications of existing designs. The exact same thing applies to software engineering. Most of us engineers are replicating designs that came earlier. LLMs are good at generating the rote designs that make up the bulk of software by volume. Who benefit from an artisanal REST interface? The best practices were codified over a decade ago.

          • By bccdee 2025-12-305:23

            > How many highway over passes have unique “chef quality” designs?

            Have you ever built a highway overpass? That kind of engineering is complex and interdisciplinary. You need to carry out extensive traffic pattern analysis and soil composition testing to even know where it should go.

            We're at a point where we've already automated all the simple stuff. If you want a website, you don't type out html tags. You use Squarespace or Wordpress or whatever. If you need a backend, you use Airtable. We already spend most of our time on the tricky stuff. Sure, it's nice that LLMs can smooth the rough edges of workflows that nobody's bothered to refine yet, but the software commodities of the world have already been commodified.

          • By keyle 2025-12-2923:041 reply

            Just like cooking in the middle ages. As the kitchen, hygiene, etc. got better, so did the chefs and so did the food.

            This is just a transition.

            re-Rest API, you're right. But again, we use roombas to vacuum when the floor layout is friendly to them. Not all rooms can be vacuumed by roombas. Simple Rest api can be emitted one shot from an LLM and there is no room for interpretation. But ask a future LLM to make a new kind of social network and you'll end up with a mash up of the existing ones.

            Same thing, you and I won't use a manual screwdriver when we have 100 screws to get in, and we own an electric drill.

            That didn't reinvent screws nor the assembly of complex items.

            I'm keeping positive in the sense that LLMs will enable us to do more, and to learn faster.

            The sad part about vibe coding is you learn very little. And to live is to learn.

            You'll notice people vibecoding all day become less and less attached to the product they work on. That's because they've given away the dopamine hits of the many "ha-ha" moments that come from programming. They'll lose interest. They won't learn anymore and die off (career wise).

            So, businesses that put LLM first will slowly lose talent over time, and business that put developers first will thrive.

            It's just a transition. A fast one that hits us like a wall, and it's confusing, but software for humans will be better made by humans.

            I've been programming since the 80s. The level of complexity today is bat shit insane. I welcome the LLM help in managing 3 code bases of 3 languages spread across different architectures (my job) to keep sane!

            • By brulard 2025-12-311:01

              I disagree with the vibecoding take. Its a new skill that absolutely has a place in developers skillset and it may be of great importance for some kinds of projects. You can learn so much by vibecoding little projects that otherwise would never see the light of day.

          • By germandiago 2025-12-308:03

            There is a part of this that is true. But when you get the nuanced parts of every "replicated design" or need the tweaks or what the AI gave you is just wrong, that deteriorates quality.

            For many tasks it is ok, for others it is just a NO.

            For software maintenance and evolution I think it won't cut it.

            The same way a Wordpress website can do a set of useful things. But when you need something specific, you just drop to programming.

            You can have your e-commerce web. But you cannot ask it to give you a "pipeline excution as fast as possible for calculating and solving math for engineering task X". That needs SIMD, parallelization, understanding the niche use you need, etc. which probably most people do not do all the time and requires specific knowledge.

        • By apt-apt-apt-apt 2025-12-2923:11

          Is your argument that we only want things that are hand-crafted by humans?

          There are lots of things like perfectly machined nails, tools, etc. that are much better done by machines. Why couldn't software be one of those?

      • By skydhash 2025-12-2921:56

        > So if you're banging out the same type of thing over and over again, yes, they will make that work light and then scarce.

        The same thing over and over again should be a SaaS, some internal tool, or a plugin. Computers are good at doing the same thing over and over again and that's what we've been using them for

        > But if you need to create something niche, something one-off, something new, they'll slip off the bleeding edge into the comfortable valley of the familiar at every step.

        Even if the high level description of a task may be similar to another, there's always something different in the implementation. A sports car and a sedan have roughly the same components, but they're not engineered the same.

        > We used to worry about pointers and memory allocation.

        Some still do. It's not in every case you will have a system that handle allocations and a garbage collector. And even in those, you will see memory leaks.

        > Now we will worry less and less about how the code is written and more about the result it built.

        Wasn't that Dreamweaver?

      • By 9dev 2025-12-2921:521 reply

        I think your image of LLMs is a bit outdated. Claude Code with well-configured agents will get entirely novel stuff done pretty well, and that’s only going to get better over time.

        I wouldn’t want to bet my career on that anyway.

    • By Deep-Blue 2025-12-304:152 reply

      As of today NONE of the known AI codebots can solve correctly ANY of the 50+ programming exercises we use to interview fresh grads or summer interns. NONE! Not even level 1 problems that can be solved in fewer than 20 lines of code with a bit of middle school math.

      • By NitpickLawyer 2025-12-307:552 reply

        After 25+ years in this field, having interviewed ~100 people for both my startup and other companies, I'm having a hard time believing this. You're either in an extremely niche field (such as to make your statement irrelevant to 99.9% of the industry), or it's hyperbole, or straight up bs.

        Interviewing is an art, and IME "gotcha" types of questions never work. You want to search for real-world capabilities, and like it or not the questions need to match those expectations. If you're hiring summer interns and the SotA models can't solve those questions, then you're doing something wrong. Sorry, but having used these tools for the past three years this is extremely ahrd to believe.

        I of course understand if you can't, but sharing even one of those questions would be nice.

        • By heldrida 2025-12-3021:41

          I agree, it’s hard to believe. Hopefully the original comment author can share one of those questions.

        • By bdangubic 2025-12-3021:43

          I would live to see just one

      • By cheevly 2025-12-306:201 reply

        I promise you that I can show you how to reliably solve any of them using any of the latest OpenAI models. Email me if you want proof; josh.d.griffith at gmail

        • By utopiah 2025-12-306:382 reply

          I'd watch that show ideally with few base rules though, e.g.

          - the problems to solve must NOT be part of the training set

          - the person using the tool (e.g. OpenAI, Claude, DevStral, DeepSeek, etc) must NOT be able to solve problems alone

          as I believe otherwise the 1st is "just" search and the 2nd is basically offloading the actual problem solving to the user.

          • By ehnto 2025-12-307:54

            > the person using the tool (e.g. OpenAI, Claude, DevStral, DeepSeek, etc) must NOT be able to solve problems alone

            I think this is a good point, as I find the operators input is often forgotten when considering the AIs output. If it took me an hour and decades of expertise to get the AI to output the right program, did the AI really do it? Could someone without my expertise get the same result?

            If not, then maybe we are wasting our time trying to mash our skills through vector space via a chat interface.

          • By cheevly 2025-12-3014:24

            Im talking generalized solutions that solve all of them.

    • By to11mtm 2025-12-2922:331 reply

      It's definitely scary in a way.

      However I'm still finding a trend even in my org; better non-AI developers tend to be better at using AI to develop.

      AI still forgets requirements.

      I'm currently running an experiment where I try to get a design and then execute on an enterprise 'SAAS-replacement' application [0].

      AI can spit forth a completely convincing looking overall project plan [1] that has gaps if anyone, even the AI itself, tries to execute on the plan; this is where a proper, experienced developer can step in at the right steps to help out.

      IDK if that's the right way to venture into the brave new world, but I am at least doing my best to be at a forefront of how my org is using the tech.

      [0] - I figured it was a good exercise for testing limits of both my skills prompting and the AI's capability. I do not expect success.

      • By dent9 2025-12-3015:441 reply

        AI does not forget requirements when you use a spec driven AI tool like Kiro

        • By songodongo 2025-12-3022:06

          Are you on the Kiro marketing team?

    • By chii 2025-12-3010:163 reply

      > I’m basically just the conductor of all those processes.

      a car moves faster than you, can last longer than you, and can carry much more than you. But somehow, people don't seem to be scared of cars displacing them(yet)? Perhaps autodriving would in the near future, but there still needs to be someone making decisions on how best to utilize that car - surely, it isn't deciding to go to destination A without someone telling them.

      > I feel like I’m doing the work of an entire org that used to need twenty engineers.

      and this is great. A combine harvester does the work of what used to be an entire village for a week in a day. More output for less people/resources expended means more wealth produced.

      • By embedding-shape 2025-12-3010:241 reply

        > a car moves faster than you, can last longer than you, and can carry much more than you. But somehow, people don't seem to be scared of cars displacing them(yet)?

        People whose life were based around using horses for transportation were very scared of cars replacing them though, and correctly so, because horses for transportation is something people do for leisure today, not necessity. I feel like that's a more apt analogy than comparing cars to any human.

        > More output for less people/resources expended means more wealth produced.

        This is true, but it probably also means that this "more wealth produced" will be more concentrated, because it's easier to convince one person using AI that you should have half of the wealth they produce, rather than convincing 100 people you should have half of what they produce. From where I'm standing, it seems to have the same effects (but not as widespread or impactful, yet) as industrialization, that induced that side-effect as well.

        • By jvanderbot 2025-12-3010:38

          Analogies are not going to work. Bug it's just as likely that, in the worst case, we are stage coach drivers who have to use cars when we just really love the quiet slowness of horses.

      • By wiether 2025-12-3011:31

        And parent is scared of being made redundant by AI because they need their job to pay for their car, insurance, gas and repairs.

      • By lelanthran 2025-12-3010:391 reply

        > a car moves faster than you, can last longer than you, and can carry much more than you. But somehow, people don't seem to be scared of cars displacing them(yet)?

        ???

        Cars replaced horses, not people.

        In this scenario you are the horse.

        • By aprilthird2021 2025-12-3010:501 reply

          Well no, you'd be the horse driver who becomes a car driver

          • By lelanthran 2025-12-3011:011 reply

            > Well no, you'd be the horse driver who becomes a car driver

            Well, that's the crux of the argument. The pro-AI devs are making the claim that devs are the horse-drivers, the anti-AI is making the claim that devs are the horses themselves.

            There is no objective way to verify who is right in this case, we just have to see it play out.

            • By aprilthird2021 2025-12-3018:50

              I don't really understand what you are saying... Anyways glad you got what I am saying at least

    • By Desafinado 2025-12-301:07

      That's kind of the point of the article, though.

      Sure LLMs can churn out code, and they sort of work for developers who already understand code and design, but what happens when that junior dev with no hard experience builds their years of experience with LLMs?

      Over time those who actually understand what the LLMs are doing and how to correct the output are replaced by developers who've never learned the hard lessons of writing code line by line. The ability to reason about code gets lost.

      This points to the hard problem that the article highlights. The hard problem of software is actually knowing how to write it, which usually takes years, sometimes up to a decade of real experience.

      Any idiot can churn out code that doesn't work. But working, effective software takes a lot of skill that LLMs will be stripping people of. Leaving a market there for people who have actually put the time in and understand software.

    • By jayd16 2025-12-2922:23

      My experience with these tools is far and away no where close to this.

      If you're really able to do the work of a 20 man org on your own, start a business.

    • By gingersnap 2025-12-3010:00

      This is not how I think about it. Me and the coding assistant is better then me or the coding assistant separately.

      For me its not about me or the coding assistant, its me and the coding assistant. But I'm also not a professional coder, i dont identify as a coder. I've been fiddling with programming my whole life, but never had it as title, I've more worked from product side or from stakeholder side, but always got more involved, as I could speak with the dev team.

      This also makes it natural for me to work side-by-side with the coding assistant, compared maybe to pure coders, who are used to keeping the coding side to themselves.

    • By zsoltkacsandi 2025-12-309:371 reply

      I have been using the most recent Claude, ChatGPT and Gemini models for coding for a bit more than a year, on a daily basis.

      They are pretty good at writing code *after* I thoroughly described what to do, step by step. If you miss a small detail they get loose and the end result is a complete mess that takes hours to clean up. This still requires years of coding experience, planning ahead in head, you won't be able to spare that, or replace developers with LLMs. They are like autocomplete on steroids, that's pretty much it.

      • By dent9 2025-12-3015:471 reply

        Yes what you are describing is exactly what Kiro solves

        • By zsoltkacsandi 2025-12-318:18

          > Through Kiro, we reinvented how developers work with AI agents.

          Even according to it’s documentation it is still built for developers, so my point still stands. You need dev experience to use this tool, same as other LLM-based coding tools.

    • By germandiago 2025-12-307:24

      I am sorry to say you are not a good programmer.

      I mean, AIs can drop something fast the same way you cannot beat a computer at adding or multiplying.

      After that, you find mistakes, false positives, code that does not work fully, and the worse part is the last one: code that does not work fully but also, as a consequence, that you do NOT understand yet.

      That is where your time shrinks: now you need to review it.

      Also, they do not design systems better. Maybe partial pieces. Give them something complex and they will hallucinate worse solutions than what you already know if you have, let us say, over 10 years of experience programming in a language (or mabye 5).

      Now multiply this unreliability problem as the code you "AI-generate" grows.

      Now you have a system you do not know if it is reliable and that you do not understand to modify. Congrats...

      I use AI moderately for the tasks is good at: generate some scripts, give me this small typical function amd I review it.

      Review my code: I will discard part of your mistakes and hallucinations as a person that knows well the language and will find maybe a few valuable things.

      Also, when reviewing and found problems in my code I saw that the LLMs really need to hallucinate errors that do not exist to justify their help. This is just something LLMs seem to not be accurate at.

      Also, when problems go a bit more atypical or past a level of difficulty, it gets much more unreliable.

      All in all: you are going to need humans. I do not know how many, I do not know how much they will improve. I just know that they are not reliable and this "generate-fast-unreliable vs now I do not know the codebase" is a fundamental obstacle that I think it is if not very difficult, impossible to workaround.

    • By khalic 2025-12-2921:09

      I feel you, it's scary. But the possibilities we're presented with are incredible. I'm revisiting all these projects that I put aside because they were "too big" or "too much for a machine". It's quite exciting

    • By CraigJPerry 2025-12-309:361 reply

      >> Coding AIs design software better than me

      Absolutely flat out not true.

      I'm extremely pro-faster-keyboard, i use the faster keyboards in almost every opportunity i can, i've been amazed by debugging skills (in fairness, i've also been very disappointed many times), i've been bowled over by my faster keyboard's ability to whip out HTML UI's in record time, i've been genuinely impressed by my faster keyboard's ability to flag flaws in PRs i'm reviewing.

      All this to say, i see lots of value in faster keyboard's but add all the prompts, skills and hooks you like, explain in as much detail as you like about modularisation, and still "agents" cannot design software as well as a human.

      Whatever the underlying mechanism of an LLM (to call it a next token predictor is dismissively underselling its capabilities) it does not have a mechanism to decompose a problem into independently solvable pieces. While that remains true, and i've seen zero precursor of a coming change here - the state of the art today is equiv to having the agent employ a todo list - while this remains true, LLMs cannot design better than humans.

      There are many simple CRUD line of business apps where they design well enough (well more accurately stated, the problem is small/simple enough) that it doesn't matter about this lack of design skill in LLMs or agents. But don't confuse that for being able to design software in the more general use case.

      • By tatjam 2025-12-3010:20

        Exactly, for the thing that has been done in Github 10000x times over, LLMs are pretty awesome and they speed up your job significantly (it's arguable if you would be better off using some abstraction already built if that's the case).

        But try to do something novel and... they become nearly useless. Not like anything particularly difficult, just something that's so niche it's never been done before. It will most likely hallucinate some methods and call it a day.

        As a personal anecdote, I was doing some LTSpice simulations and tried to get Claude Sonnet to write a plot expression to convert reactance to apparent capacitance in an AC sweep. It hallucinated pretty much the entire thing, and got the equation wrong (assumed the source was unit intensity, while LTSpice models AC circuits with unit voltage. This surely is on the internet, but apparently has never been written alongside the need to convert an impedance to capacitance!).

    • By Herring 2025-12-2922:19

      Try have your engineers pick up some product work. Clients do NOT want to talk to bots.

    • By lelanthran 2025-12-3010:381 reply

      > Coding AIs design software better than me, review code better than me, find hard-to-find bugs better than me, plan long-running projects better than me, make decisions based on research, literature, and also the state of our projects better than me.

      They don't do any of that better than me; they do it poorer and faster, but well enough for most of the time.

      • By dent9 2025-12-3015:46

        Then you are using the wrong AI tools or using them poorly

    • By tom_m 2025-12-302:21

      There will be a need. Don't worry. Most people still haven't figured out how to properly read and interpret instructions. So they build things incorrectly - with or without AI

      Seriously. The bar is that low. When people say "AI slop" I just chuckle because it's not "AI" it's everyone. That's the general state of the industry.

      So all you have to do is stay engaged, ask questions, and understand the requirements. Know what it is you're building and you'll be fine.

    • By conartist6 2025-12-3012:383 reply

      More than any other effect they have LLMs breed something called "learned helplessness". You just listed a few things it may stay better than you at, and a few things that it is not better than you at and never will be.

      Planning long running projects and deciding are things only you can do well!! Humans manage costs. We look out for our future. We worry. We have excitement, and pride. It wants you to think none of these things matter of course, because it doesn't have them. It says plausible things at random, basically. It can't love, it can't care, it won't persist.

      WHATEVER you do don't let it make you forget that it's a bag of words and you are someing almost infinitely more capable, not in spite of human "flaws" like caring, but because of them :)

      • By conartist6 2025-12-3012:591 reply

        Plus I think I've almost never see so little competition for what I think are the real prizes! Everyone's off making copies of copies of copies of the same crappy infrastructure we already have. They're busy building small inconsequential side projects so they can say they built something using an LLM.

        • By embedding-shape 2025-12-3013:081 reply

          > They're busy building small inconsequential side projects

          Unironically, sending a program to build those for me have send me almost endless amount of time. I'm a pretty distracted individual, and pretty anal about my workflow/environment, so lots of times I've spent hours going into rabbit-holes to make something better, when I could have just sucked it up and do it the manual way instead, even if it takes mental energy.

          Now, I can still do those things, but not spend hours, just a couple of minutes, and come back after 20-30 minutes to something that lets me avoid that stuff wholesale. Once you start stacking these things, it tends to save a lot of time and more importantly, mental energy.

          So the programs by themselves are basically "small inconsequential side projects" because they're not "production worthy and web scale SaaS ready to earn money", but they help me and others who are building those things in a big way.

          • By throw4847285 2025-12-3014:191 reply

            But isn't that exactly the kind of learned helplessness being discussed? As a fellow distracted individual, I have seen instant gratification erode all of my most prized hobbies and skills. Why read a book when I can scroll on my phone? My distress tolerance is lower than ever. LLMs feel like a bridge too far, for me anyway.

            • By embedding-shape 2025-12-3014:411 reply

              Nothing has been eroded for me, in fact it had the opposite effect. It's easier to get into new hobbies, easier to develop skills, I value reading on my own more than I did before. At least for me, LLMs act as multipliers of what I can and want to do, it hasn't removed my passion for music production, 3D, animation or programming one bit, if anything it's fueled those passions and let me do stuff within them faster and better.

              • By throw4847285 2025-12-3015:001 reply

                Nothing I could make would be very good. So the only reason I would, say, write, is in order to write, not to have produced an essay. Hobbies are ways to pass time productively. If it took less time, it wouldn't be a better use of time, but a worse one.

                • By embedding-shape 2025-12-3015:511 reply

                  It's not about being able to do more faster, but be able to faster get help doing what you wanted to do. For example, before LLMs, if I wanted to figure out how to do something with a specific analog synth I basically spent time reading manuals and browsing internet forums, piecing together whatever I could find into something actionable, sometimes slightly wrong, but at least in the right direction.

                  Nowadays, I fire off the LLM to figure it out for me, then try out what I get back, and I can move on to actually having fun playing on the synth, rather than trying to figure out how to do what I wanted to do.

                  The end goal for me with my hobbies is more or less the same, have fun. But for me the fun is not digging through manuals, it is to "do" or "use" or "perform" or whatever. I like music production because I like to make music, not because I like digging through manuals for some arcane knowledge.

                  • By throw4847285 2025-12-3017:141 reply

                    But looking up information via an LLM is an entirely different category of usage. I have no problem with that (well, much less of a problem).

                    • By embedding-shape 2025-12-3018:29

                      The point is "things that used to take me hours, can now be done by a magic computer program in the background, while I do other things". It's applicable for small unix utilities I create to make my development UX better, it's applicable for when I'm doing music production and it's applicable in a wide-range of tasks both professionally and for my hobbies.

                      It saves me from stuff I find boring yet necessary, so I can focus more on the fun stuff. I guess this was the overall point I was trying to make in this comment-chain.

      • By cootsnuck 2025-12-3018:06

        Yea I've been seeing very similar behavior from people. They think of themselves as static, unchanging, uncreative but view LLMs as some kind of unrelenting and inevitable innovative force...

        I think it's people's anxieties and fears about the uncertainty about the value of their own cognitive labor demoralizing them and making them doubt their own self-efficacy. Which I think is an understandable reaction in the face of trillion dollar companies frothing at the mouth to replace you with pale imitations.

        Best name I could think of calling this narrative / myth is people believing in "effortless AI": https://www.insidevoice.ai/p/effortless-ai

      • By XenophileJKO 2025-12-3019:552 reply

        You are still in denial of what an LLM actually is capable of in the near-mid term.

        In the current architecture there are mathmatical limitations on what it can do with information. However, tool use and external orchestration allow it to work around many (maybe all) those limitations.

        The current models have brittle parts and some bad tendencies.. but they will continue to eat up the executive thought ladder.

        I think it is better to understand this and position yourself higher and higher on that chain while learning what are the weak areas in the current generation of models.

        Your line of thinking is like hiding in a corner while the water is rising. You are right, it is a safe corner, but probably not for long.

        • By conartist6 2025-12-3116:37

          I don't think the limitations on what it can do are mathematical at all. It has no, faith, no conviction, no sense of self. No philosophy, no ability to learn. How could it undertake a major effort?

        • By conartist6 2025-12-3022:431 reply

          I'm as high on the chain as it is possible to get! I don't use AI at all. Models help people follow, but I'm leading. Bite me.

          • By XenophileJKO 2025-12-3023:351 reply

            No reason to be uncivil.

            Just so we are clear, you are saying you don't use it at all, but you are providing advice about it? Specifically detailing with certainty that the current state of the art has or doesn't have certain traits or abilities.

            • By conartist6 2025-12-3113:38

              Yes. I'm not providing advice on how to use it, I'm providing advice on whether or not to use it. A million people cried out that I would be obsolete. I would be replaced: left behind. Career suicide one said LOL.

              I think I'm the perfect person to be qualified to stand up and say "if they tell you you can't live without it, they are lying to your face." Only someone who has lived without it as I have would be in a position to know

    • By deadbabe 2025-12-2922:05

      Where the hell was all this fear when the push for open source everything got fully underway? When entire websites were being spawned and scaffolded with just a couple lines of code? Do we not remember all those impressive tech demos of developers doing massive complex thing with "just one line of code"? How did we not just write software for every kind of software problem that could exist by now?

      How has free code, developed by humans, become more available than ever and yet somehow we have had to employ more and more developers? Why didn't we trend toward less developers?

      It just doesn't make sense. AI is nothing but a snippet generator, a static analyzer, a linter, a compiler, an LSP, a google search, a copy paste from stackoverflow, all technologies we've had for a long time, all things developers used to have to go without at some point in history.

      I don't have the answers.

    • By goodpoint 2025-12-309:07

      > Coding AIs design software better than me, review code better than me, find hard-to-find bugs better than me, plan long-running projects better than me, make decisions based on research, literature, and also the state of our projects better than me

      ChatGPT, is that you?

    • By scellus 2025-12-2922:07

      Perfect economic substitution in coding doesn't happen for a long time. Meanwhile, AI appears as an amplifier to the human and vice versa. That the work will change is scary, but the change also opens up possibilities, many of them now hard to imagine.

    • By bborud 2025-12-3012:01

      Notice who makes these predictions that programmers will become irrelevant.

    • By heliumtera 2025-12-2921:592 reply

      Stop freaking out. Seriously. You're afraid of something completely ridiculous.

      It is certainly more eloquent than you regarding software architecture (which was a scam all along, but conversation for another time). It will find SOME bugs better than you, that's a given.

      Review code better than you? Seriously? What you're using and what you consider code review? Assume I could identify one change broke production and you reviewed the latest commit. I am pinging you and you better answer. Ok, Claude broke production, now what? Can you begin to understand the difference between you and the generative technology? When you hop on the call, you will explain to me with a great deal of details what you know about the system you built, and explain decision making and changes over time. You'll tell about what worked and what didn't. You will tell about the risks, behavior and expectations. About where the code runs, it's dependencies, users, usage patterns, load, CPU usage and memory footprint, you could probably tell what's happening without looking at logs but at metrics. With Claude I get: you're absolutely right! You asked about what it WAS, but I told you about what it WASN'T! MY BAD.

      Knowledge requires a soul to experience and this is why you're paid.

      • By mywittyname 2025-12-2922:461 reply

        We use code rabbit and it's better than practically any human I've worked with at a number of code review tasks, such as finding vulnerabilities, highlighting configuration issues, bad practices, etc. It's not the greatest at "does this make sense here" type questions, but I'd be the one answering those questions anyway.

        Yeah, maybe the people I've worked with suck at code reviews, but that's pretty normal.

        Not to say your answer is wrong. I think the gist is accurate. But I think tooling will get better at answering exactly the kind of questions you bring up.

        Also, someone has to be responsible. I don't think the industry can continue with this BS "AI broke it." Our jobs might devolve into something more akin to a SDET role and writing the "last mile" of novel code the AI can't produce accurately.

        • By andrekandre 2025-12-3019:43

            > We use code rabbit and it's better than practically any human
          
          code rabbit does find things occasionally, but it also calls things 'critical' that arent and flags issues that dont actually exist and even lies in replies sometimes...

          it also is extremely verbose to the point of being slog to go through... and the haikus: they are so cringe and infantilizing...

          maybe its our config, but code rabbit has been underwhelming...

      • By anonymars 2025-12-2922:511 reply

        > Review code better than you? Seriously?

        Yes, seriously (not OP). Sometimes it's dumb as rocks, sometimes it's frighteningly astute.

        I'm not sure at which point of the technology sigmoid curve we find ourselves (2007 iPhone or 2017 iPhone?) but you're doing yourself a disservice to be so dismissive

        • By heliumtera 2025-12-2923:05

          Copilot reviews are enabled company wide and comments must be resolved manually. I wish I could be so dismissive lol I cannot, literally do not have the ability to be dismissive

    • By robofanatic 2025-12-3018:26

      >I really really want this to be true. I want to be relevant

      Think of yourself as a chef and LLMs as ready to eat meals or a recipe app. Can ready to eat meals OR recipe apps put a chef out of business?

    • By bob1029 2025-12-307:41

      The AI is pretty scary if you think most of software engineering is about authoring individual methods and rubber ducking about colors of paint and brands of tools.

      Once you learn that it's mostly about interacting with a customer (sometimes this is yourself), you will realize the AI is pretty awful at handling even the most basic tasks.

      Following a product vision, selecting an appropriate architecture and eschewing 3rd party slop are examples of critical areas where these models are either fundamentally incapable or adversely aligned. I find I have to probe ChatGPT very hard to get it to offer a direct implementation of something like a SAML service provider. This isn't a particularly difficult thing to do in a language like C# with all of the built in XML libraries, but the LLM will constantly try to push you to use 3rd party and cloud shit throughout. If you don't have strong internal convictions (vision) about what you really want, it's going to take you for a ride.

      One other thing to remember is that our economies are incredibly efficient. The statistical mean of all information in sight of the LLMs likely does not represent much of an arbitrage opportunity at scale. Everyone else has access to the same information. This also means that composing these systems in recursive or agentic styles means you aren't gaining anything. You cannot increase the information content of a system by simply creating another instance of the same system and having it argue with itself. There usually exists some simple prompt that makes a multi agent Rube Goldberg contraption look silly.

      > I’m basically just the conductor of all those processes.

      "Basically" and "just" are doing some heroic weight lifting here. Effectively conducting all of the things an LLM is good at still requires a lot of experience. Making the constraints live together in one happy place is the hard part. This is why some of us call it "engineering".

    • By 63stack 2025-12-2921:568 reply

      This reads like shilling/advertisement.. Coding AIs are struggling for anything remotely complex, make up crap and present it as research, write tests that are just "return true", and won't ever question a decision you make.

      Those twenty engineers must not have produced much.

      • By pfannkuchen 2025-12-304:447 reply

        I think part of what is happening here is that different developers on HN have very different jobs and skill levels. If you are just writing a large volume of code over and over again to do the same sort of things, then LLMs probably could take your job. A lot of people have joined the industry over time, and it seems like the intelligence bar moved lower and lower over time, particularly for people churning out large volumes of boilerplate code. If you are doing relatively novel stuff, at least in the sense that your abstractions are novel and the shape of the abstraction set is different from the standard things that exist in tutorials etc online, then the LLM will probably not work well with your style.

        So some people are panicking and they are probably right, and some other people are rolling their eyes and they are probably right too. I think the real risk is that dumping out loads of boilerplate becomes so cheap and reliable that people who can actually fluently design coherent abstractions are no longer as needed. I am skeptical this will happen though, as there doesn’t seem to be a way around the problem of the giant indigestible hairball (I.e as you have more and more boilerplate it becomes harder to remain coherent).

        • By mekoka 2025-12-3010:52

          Indeed, discussions on LLMs for coding sound like what you would expect if you asked a room full of people to snatch up a 20 kg dumbbell once and then tell you if it's heavy.

          > I think the real risk is that dumping out loads of boilerplate becomes so cheap and reliable that people who can actually fluently design coherent abstractions are no longer as needed.

          Cough front-end cough web cough development. Admittedly, original patterns can still be invented, but many (most?) of us don't need that level of creativity in our projects.

        • By stackghost 2025-12-308:042 reply

          Absolutely this, and TFA touches on the point about natural language being insufficiently precise:

          AI can write you an entire CRUD app in minutes, and with some back-and-forth you can have an actually-good CRUD app in a few hours.

          But AI is not very good (anecdotally, based on my experience) at writing fintech-type code. It's also not very good at writing intricate security stuff like heap overflows. I've never tried, but would certainly never trust it to write cryptography correctly, based on my experience with the latter two topics.

          All of the above is "coding", but AI is only good at a subset of it.

          • By bonesss 2025-12-3012:151 reply

            Generating CRUD is like solving cancer in mice, we already have a dizzying array of effective solutions… Ruby on Rails, Access 97, model first ORMs with GUI mappers. SharePoint lets anyone do all the things easily.

            The issue is and always has been maintenance and evolution. Early missteps cause limitations, customer volume creates momentum, and suddenly real engineering is needed.

            I’d be a lot more worried about our jobs if these systems were explaining to people how to solve all their problems with a little Emacs scripting. As is they’re like hyper aggressive tech sales people, happy just to see entanglements, not thinking about the whole business cycle.

            • By skydhash 2025-12-3012:581 reply

              Go with Laravel and some admin packages and you generate CRUD pages in minutes. And I think with Django, that is builtin.

              But I don’t think I’ve seen pure CRUD on anything other than prototype. Add an Identity and Access Management subsystem and the complexity of requirements will explode. Then you add integration to external services and legacy systems, and that’s where the bulk of the work is. And there’s the scalability issue that is always looming.

              Creating CRUD app is barely a level over starting a new project with the IDE wizard.

              • By stackghost 2025-12-3017:231 reply

                >Creating CRUD app is barely a level over starting a new project with the IDE wizard.

                For you, maybe. But for a non-progrmamer who's starting a business or just needs a website it's the difference between hiring some web dev firm and doing it themselves.

                • By andrekandre 2025-12-3018:421 reply

                    > it's the difference between hiring some web dev firm and doing it themselves.
                  
                  anecdote but i've had a lot of acquaintances who started at both "hiring some web dev firm" and "doing it themselves" with results largely being the same: "help me fix this unmaintainable mess and i will pay you x"...

                  jmo but i suspect llms will allow for the later to go further before the "help me" phase but i feel like that aint going away completely...

                  • By stackghost 2025-12-3022:28

                    Just like my previous comments, much depends on the specifics.

                    My wife's sister and her husband run a small retail shop in $large_city. My sister-in-law taught herself how to set up and modify a website with a shopify storefront largely with LLM help. Now they take online orders. I've looked at the code she wrote and it's not pretty but it generally works. There will probably never be a "help me fix this unmaintainable mess and I will pay you" moment in the life of that business.

                    The crux of my point is this: In 2015 she would have had to hire somebody to do that work.

                    This segment of the software industry is where the "LLMs will take our jerbs" argument is coming from.

                    The people who say "AI is junk and it can't do anything right" are simply operating in a different part of the industry.

          • By llmslave2 2025-12-3010:051 reply

            > and with some back-and-forth you can have an actually-good CRUD app in a few hours

            Perhaps the debate is on what constitutes "actually-good". Depends where the bar is I suppose.

            • By stackghost 2025-12-3022:29

              Beauty is in the eye of the beholder. Litigating our personal opinions about "actually-good" is irrelevant and pointless.

        • By IshKebab 2025-12-307:292 reply

          > different developers on HN have very different jobs and skill levels.

          Definitely this. When I use AIs for web development they do an ok job most of the time. Definitely on par with a junior dev.

          For anything outside of that they're still pretty bad. Not useless by any stretch, but it's still a fantasy to think you could replace even a good junior dev with AI in most domains.

          I am slightly worried for my job... but only because AI will keep improving and there is a chance it will be as good as me one day. Today it's not a threat at all.

          • By ryandrake 2025-12-307:481 reply

            Yea, LLMs produce results on par with what I would expect out of a solid junior developer. They take direction, their models act as the “do the research” part, and they output lots of code: code that has to be carefully scrutinized and refined. They are like very ambitious interns who never get tired and want to please, but often just produce crap that has to be totally redone or refactored heavily in order to go into production.

            If you think LLMs are “better programmers than you,” well, I have some disappointing news for you that might take you a while to accept.

            • By monsieurbanana 2025-12-3010:002 reply

              > LLMs produce results on par with what I would expect out of a solid junior developer

              This is a common take but it hasn't been my experience. LLMs produce results that vary from expert all the way to slightly better than markov chains. The average result might be equal to a junior developer, and the worst case doesn't happen that often, but the fact that it happens from time to time makes it completely unreliable for a lot of tasks.

              Junior developers are much more consistent. Sure, you will find the occasional developer that would delete the test file rather than fixing the tests, but either they will learn their lesson after seeing your wth face or you can fire them. Can't do that with llms.

              • By jvanderbot 2025-12-3010:451 reply

                I think any further discussion about quality just needs to have the following metadata:

                - Language

                - Total LOC

                - Subject matter expertise required

                - Total dependency chain

                - Subjective score (audited randomly)

                And we can start doing some analysis. Otherwise we're pissing into ten kinds of winds.

                My own subjective experience is earth shattering at webapps in html and css (because I'm terrible and slow at it), and annoyingly good but a bit wrong usually in planning and optimization in rust and horribly lost at systems design or debugging a reasonably large rust system.

                • By monsieurbanana 2025-12-3011:032 reply

                  I agree in that these discussions (this whole hn thread tbh) are seriously lacking in concrete examples to be more than holy wars 3.0.

                  Besides one point: junior developers can learn from their egregious mistakes, llms can't no matter how strongly worded you are in their system prompt.

                  In a functional work environment, you will build trust with your coworkers little by little. The pale equivalent in LLMs is improving system prompts and writing more and more ai directives that might or might not be followed.

                  • By ryandrake 2025-12-3016:16

                    This seems to be one of the huge weaknesses of current LLMs: Despite the words "intelligence" and "machine learning" we throw around, they aren't really able to learn and improve their skills without someone changing the model. So, they repeat the same mistakes and invent new mistakes by random chance.

                    If I was tutoring a junior developer, and he accidentally deleted the whole source tree or something egregious, that would be a milestone learning point in his career, and he would never ever do it again. But if the LLM does it accidentally, it will be apologetic, but after the next context window clear, it has the same chances of doing it again.

                  • By embedding-shape 2025-12-3013:101 reply

                    > Besides one point: junior developers can learn from their egregious mistakes, llms can't no matter how strongly worded you are in their system prompt.

                    I think if you set off an LLM to do something, and it does a "egregious mistake" in the implementation, and then you adjust the system prompt to explicitly guard against that or go towards a different implementation and you restart from scratch again yet it does the exact same "egregious mistake", then you need to try a different model/tool than the one you've tried that with.

                    It's common with smaller models, or bigger models that are heavily quanitized that they aren't great at following system/developer prompts, but that really shouldn't happen with the available SOTA models, I haven't had something ignored like that in years by now.

                    • By jvanderbot 2025-12-3013:29

                      And honestly this is precisely why I don't fear unemployment, but I do fear less employment overall. I can learn and get better and use LLMs as a tool. So there's still a "me" there steering. Eventually this might not be the case. But if automating things has taught me anything, it's that removing the person is usually such a long tail cost that it's cheaper to keep someone in the loop.

                      But is this like steel production or piloting (few highly trained experts are in the loop) or more like warehouse work (lots of automation removed any skills like driving or inventory work etc).

              • By AnimalMuppet 2025-12-3014:36

                I can in fact fire an LLM. It's even easier than firing a junior developer.

                Or rather, it's more like a contractor. If I don't like the job they did, I don't give them the next job.

          • By anthonypasq 2025-12-3022:25

            you say this as if web development isnt 90% of software.

        • By 1718627440 2025-12-3010:011 reply

          > If you are just writing a large volume of code over and over again

          But why would you do that? Wouldn't you just have your own library of code eventually that you just sell and sell again with little tweaks? Same money for far less work.

          • By embedding-shape 2025-12-3010:302 reply

            People, at least novice developers, tend to prefer fast and quick boilerplate that makes them look effective, over spending one hour sitting just thinking and designing, then implementing some simple abstraction. This is true today, and been true for as long as I've been in programming.

            Besides, not all programming work can be abstracted into a library and reused across projects, not because it's technically infeasible, but because the client doesn't want to, cannot for legal reasons or the developer process at the client's organization simply doesn't support that workflow. Those are just the reasons from the top of my head, that I've encountered before, and I'm sure there is more reasons.

            • By 1718627440 2025-12-3023:501 reply

              But people don't stay novices after years/decades. Of course when you write the boilerplate for the 20x time maybe you still accept that, but when you write it for the 2000x time, I bet you do the lazy thing and just copy it.

              > cannot for legal reasons or ...

              Sure, you can't copy trade secrets, but that's also not the boilerplate part. Copying e.g. a class hierarchy and renaming all the names and replacing the class contents that represent the domain, won't be a legal problem, because this is not original in the first place.

              • By embedding-shape 2025-12-3113:521 reply

                > But people don't stay novices after years/decades

                Some absolutely do. I know programmers who entered web development at the same time as me, and now after decades they're still creating typical CRUD applications for whatever their client today is, using the same frameworks and languages. If it works, makes enough money and you're happy, why change?

                > Copying e.g. a class hierarchy and renaming all the names and replacing the class contents that represent the domain, won't be a legal problem, because this is not original in the first place.

                Some code you produce for others definitively fall under their control, but obviously depends on the contracts and the laws of the country you're in. But I've written code for others that I couldn't just "abstract into a FOSS library and use in this project", even if it wasn't trade secrets or what not, just some utility for reducing boilerplate.

                • By 1718627440 2025-12-3122:06

                  > "abstract into a FOSS library and use in this project"

                  That is not what I meant. My idea was more like "copy ten lines from this project, then lines from that project, the class from here, but replace every line before the commit ...".

                  I shouldn't have used the word library, as I did not mean output from the linker, but rather a colloquial meaning of a loose connection of snippets.

        • By therobots927 2025-12-3014:08

          That’s a very good point I hadn’t heard explained that way before. Makes a lot of sense and explains a lot of the circular debates about AI that happen here daily.

        • By charcircuit 2025-12-308:034 reply

          >at least in the sense that your abstractions are novel and the shape of the abstraction set is different from the standard things that exist

          People shouldn't be doing this in the first place. Existing abstractions are sufficient for building any software you want.

          • By yetihehe 2025-12-308:311 reply

            > Existing abstractions are sufficient for building any software you want.

            Software that doesn't need new abstractions is also already existing. Everything you would need already exists and can be bought much more cheaply than you could do it yourself. Accounting software exists, unreal engine exists and many games use it, why would you ever write something new?

            • By charcircuit 2025-12-3010:15

              >Software that doesn't need new abstractions is also already existing

              This isn't true due to the exponential growth of how many ways you can compose existing abstractions. The chance that a specific permutation will have existing software is small.

          • By bryanrasmussen 2025-12-308:18

            I'm supposing that nobody who has a job is producing abstractions that are always novel, but there may be people who find abstractions that are novel for their particular field because it is something most people in that field are not familiar with, or that come up with novel abstractions (infrequently) that improve on existing ones.

          • By bckr 2025-12-308:39

            The new abstraction is “this corporation owns this IP and has engineers who can fix and extend it at will”. You can’t git clone that.

            But if there is something off the shelf that you can use for the task at hand? Great! The stakeholders want it to do these other 3000 things before next summer.

          • By mdemare 2025-12-3021:051 reply

            Software development is a bit like chess. 1. e4 is an abstraction available to all projects, 3. Nc3 is available to 20% of projects, while 15. Nxg5 is unique to your own project.

            Or, abstractions in your project form a dependency tree, and the nodes near the root are universal, e.g. C, Postgres, json, while the leaf nodes are abstractions peculiar to just your own project.

            • By charcircuit 2025-12-311:30

              The possible chess moves is already known ahead of time. Just because an AI can't make up a move like Np5 as a human could do, that doesn't mean anything AI can't play chess. It will be fine with just using the existing moves that have been found so far. The idea that we still need humans to come up with new chess moves is not a requirement for playing chess.

      • By aspenmartin 2025-12-2922:223 reply

        No it doesn’t read like shilling and advertisement, it’s tiring hearing people continually dismiss coding agents as if they have not massively improved and are driving real value despite limitations and they are only just getting started. I’ve done things with Claude I never thought possible for myself to do, and I’ve done things where Claude made the whole effort take twice as long and 3x more of my time. It’s not like people are ignoring the limitations, it’s that people can see how powerful the already are and how much more headroom there is even with existing paradigms not to mention the compute scaling happening in 26-27 and the idea pipeline from the massive hoarding of talent.

        • By jayd16 2025-12-2922:243 reply

          When prices go down or product velocity goes up we'll start believing in the new 20x developer. Until then, it doesn't align with most experiences and just reads like fiction.

          You'll notice no one ever seems to talk about the products they're making 20x faster or cheaper.

          • By hansmayer 2025-12-2922:314 reply

            +1 - I wish at least one of these AI boosters had shown us a real commercialised product they've built.

            • By aspenmartin 2025-12-2922:431 reply

              AI boosters? Like people are planted by Sam Altman like the way they hire crowds for political events or something? Hey! Maybe I’m AI! You’re absolutely right!

              In seriousness: I’m sure there are projects that are heavily powered by Claude, myself and a lot of other people I know use Claude almost exclusively to write and then leverage it as a tool when reviewing. Almost everyone I hear that has this super negative hostile attitude references some “promise” that has gone unfulfilled but it’s so silly: judge the product they are producing and maybe just maybe consider the rate of progress to _guess_ where things are heading

              • By hansmayer 2025-12-2923:077 reply

                I never said "planted", that is your own assumption, albeit a wrong one. I do respect it though, as it is at least a product of a human mind. But you don't have to be "planted" to champion an idea, you are clearly championing it out of some kind of conviction, many seem to do. I was just giving you a bit of reality check.

                If you want to show me how to "guess where things are heading" / I am actually one of the early adopters of LLMs and have been engineering software professionally for almost half my life now. Why do you think I was an early adopter? Because I was skeptical or afraid of that tech? No, I was genuinely excited. Yes you can produce mountains of code, even more so if you were already an experienced engineer, like myself for example.

                Yes you can even get it to produce somewhat acceptable outputs, with a lot of effort at prompting it and fatigue that comes with it. But at the end of the day, as an experienced engineer, I am not being more productive with it, I will end up being less productive because of all the sharp edges I have to take care of, all the sloppily produced code, unnecessary bloat, hallucinated or injected libraries etc.

                Maybe for folks who were not good at maths or had trouble understanding how computers work this looks like a brave new world of opportunities. Surely that app looks good to you, how bad can it be? Just so you and other such vibe-coders understand, here is a parallel.

                It is actually fairly simple for a group of aviation enthusiasts to build a flying airplane. We just need to work out some basic mechanics, controls and attach engines. It can be done, I've seen a couple of documentaries too. However, those planes are shit. Why? Because me and my team of enthusiast dont have the depth of knowledge of a team of aviation engineers to inform my decisions.

                What is the tolerance for certain types of movements, what kind of materials do I need to pick, what should be my maintenance windows for various parts etc. There are things experts can decide on almost intuitively, yet with great precision, based on their many years of craft and that wonderful thing called human intelligence. So my team of enthusiasts puts together an airplane. Yeah it flies. It can even be steered. It rolls, pitches and yawns. It takes off and lands. But to me it's a black-box, because I don't understand many, many factors, forces, pressures, tensors, effects etc that are affecting an airplane during it's flight and takeoff. I am probably not even aware WHAT I should be aware of. Because I dont have that deep educaiton about mechanical engineering, materials, aerodynamics etc. Neither does my team. So my plane, while impressive to me and my team, will never take off commercially, not unless a team of professionals take it over and remakes it to professional standards. It will probably never even fly in a show. And if me or someone on my team dies flying it, you guessed it - our insurance sure as hell won't cover the costs.

                So what you are doing with Claude and other tools, while it may look amazing to you, is not that impressive to the rest of us, because we can see those wheels beginning to fall off even before your first take off. Of course, before I can even tell that, I'd have to actually see your airplane, it's design plans etc. So perhaps first show us some of those "projects heavily powered by Claude" and their great success, especially commercial one (otherwise its a toy project), before you talk about them.

                The fact that you are clearly not an expert on the topic of software engineering should guide you here - unless you know what you are talking about, it's better to not say anything at all.

                • By user34283 2025-12-309:291 reply

                  How would you know whether he is an expert on the topic of software engineering or not?

                  For all I know, he is more competent than you; he figured out how to utilize Claude Code in a productive way, which is a point for him.

                  I'd have to guess whether you are an expert working on software not well suited for AI, or just average with a stubborn attitude towards AI and potentially not having tried the latest generation of models and agentic harnesses.

                  • By llmslave2 2025-12-3010:022 reply

                    > How would you know whether he is an expert on the topic of software engineering or not?

                    Because of their views on the effectiveness of AI agents for generating code.

                    • By user34283 2025-12-3010:273 reply

                      Considering those views are shared by a number of high profile, skilled engineers, this is obviously no basis for doubting someone's expertise.

                      • By mekoka 2025-12-3011:41

                        I think it's worth framing things back to what we're reacting to. The top poster said:

                        > I really really want this to be true. I want to be relevant. I don’t know what to do if all those predictions are true and there is no need (or very little need) for programmers anymore.

                        The rest of the post is basically their human declaration of obsolescence to the programming field. To which someone reacted by saying that this sounds like shilling. And indeed it does for many professional developers, including those that supplement their craft with LLMs. Declaring that you feel inadequate because of LLMs only reveals something about you. Defending this position is a tell that puts anyone sharing that perspective in the same boat: you didn't know what you were doing in the first place. It's like when someone who couldn't solve the "invert a binary tree" problem gets offended because they believed they were tricked into an impossible task. No, you may be a smart person that understands enough of the rudiment of programming to hack some interesting scripts, but that's actually a pretty easy problem and failing to solve it indeed signals that you lack some fundamentals.

                        > Considering those views are shared by a number of high profile, skilled engineers, this is obviously no basis for doubting someone's expertise.

                        I've read Antirez, Simon Willison, Bryan Cantrill, and Armin Ronacher on how they work or want to work with AI. From none I've got this attitude that they're no longer needed as part of the process.

                      • By hansmayer 2025-12-3011:08

                        > Considering those views are shared by a number of high profile, skilled engineers, this is obviously no basis for doubting someone's expertise

                        Again, a lot of fluff, a lot of of "a number ofs", "highly this, highly that". But very little concrete information. What happened to the pocket PhDs promised for this past summer? Where are the single-dude billion dollar companies built with AI tools ? Or even a multiple-dudes billion dollar companies ? What are you talking about?

                      • By llmslave2 2025-12-3010:442 reply

                        I've yet to see it from someone who isn't directly or indirectly affiliated with an organisation that would benefit from increased AI tool adoption. Not saying it's impossible, but...

                        Whereas there are what feels like endless examples of high profile, skilled engineers who are calling BS on the whole thing.

                        • By victorbjorklund 2025-12-3012:061 reply

                          You can say the same about people saying the opposite. I haven’t heard from a single person who says AI can’t write code that does not a financially interest directly or indirectly in humans writing code.

                          • By llmslave2 2025-12-3012:10

                            Nobody says AI "can't write code". It very clearly can.

                        • By user34283 2025-12-3021:56

                          That seems rather disingenuous to me. I see many posts which clearly come from developers like you and me who are happy with the results they are getting.

                          Every time people on here comment something about "shilling" or "boosters". It would seem to me that in the rarest of cases someone shares their opinion to profit from it, while you act like that is super common.

                    • By aspenmartin 2025-12-3020:35

                      Right: they disagree with me and so must not know what they’re talking about. Hey guess how I know neither of you are all as good as you think you are: your egos! You know what the brightest people at the top of their respective fields have in common? They tend not to think that new technologies they don’t understand how to use are dumb and they don’t think everyone who disagrees with them is dumb!

                • By aspenmartin 2025-12-3020:25

                  > you are clearly not an expert on the topic of software engineering should guide you here - unless you know what you are talking about, it's better to not say anything at all.

                  Yikes, pretty condescending. Also wrong!

                  IMO you are strawmanning pretty heavily here.

                  Believe it or not, using Claude to improve your productivity is pretty dissimilar to vibe coding a commercial airplane(?) which I would agree is probably not FAA approved.

                  I prefer not to toot my own horn, but to address an idea you seem to have that I don’t know math or CS(?) I have a PhD in astrophysics and a decade of industry experience in tech and other domains so I’m fairly certain I know how math and computers work but maybe not!

                • By threethirtytwo 2025-12-307:052 reply

                  I’m an expert in what I do. A professional, and few people can do what I do. I have to say you are wrong. AI is changing the game. What you’ve written here might’ve been more relevant about 9 months ago, but everything has changed.

                  • By leptons 2025-12-309:043 reply

                    This is a typical no-proof "AI"-boosting response, and from an account created only 35 days ago.

                    • By threethirtytwo 2025-12-3013:523 reply

                      Right I’m a bot made to promote AI like half the people on this thread.

                      I don’t know if you noticed a difference from other hype cycles but other ones were speculative. This one is also speculative but the greater divide is that the literal on the ground usefulness of AI is ALREADY going to change the world.

                      The speculation is that the AI will get better and will no longer need hand holding.

                      • By afishhh 2025-12-3014:271 reply

                        I'm having a lot of trouble understanding what you're trying to convey. You say there's a difference from previous "speculation" but also that it's still speculation. Then you go on to write "ALREADY going to" which is future tense (speculation), even clarifying what the speculation is.

                        Is this sarcasm, ragebait, or a serious argument?

                        • By threethirtytwo 2025-12-3018:30

                          Serious.

                          So let me explain it more clearly. AI as it is now is already changing the game. It will reduce the demand of swes across every company as an eventuality if we hold technological progress fixed. There is no speculation here. This comes from on the ground evidence from what I see day to day and what I do and my experience pair programming things from scratch with AI.

                          The speculation is this: if we follow the trendlines of AI improvement for the past decade and a half, the projection of past improvement indicates AI will only get better and better. It’s a reasonable speculation, but it is nonetheless speculative. I wouldn’t bet my life on continuous improvement of AI to the point of AGI but it’s now more than ever before a speculation that is not unrealistic.

                      • By leptons 2025-12-3022:351 reply

                        >AI is ALREADY going to change the world.

                        Nice slop response. This is the same thing said about blockchain and NFTs, same schtick, different tech. The only thing "AI" has done is convince some people that it's a magical being that knows everything. Your comments seem to be somewhere on that spectrum. And, sure what if it isn't changing the world for the better, and actually makes things much worse? You're probably okay with that too, I guess, as long as your precious "AI" is doing the changing.

                        We've seen what social media and every-waking-hour access to tablets and the internet has done to kids - so much harm that some countries have banned social media for people under a certain age. I can see a future where "AI" will also be banned for minors to use, probably pretty soon too. The harms from "AI" being able to placate instead of create should be obvious, and children shouldn't be able to use it without adult supervision.

                        >The speculation is that the AI will get better and will no longer need hand holding.

                        This is nonsense. No AI is going to produce what someone wants without telling it exactly what to do and how to do it, so yes, it will always need hand holding, unless you like slurping up slop. I don't know you, if you aren't a bot, you might just be satisfied with slop? It's a race to the bottom, and it's not going to end up the way you think it will.

                        • By threethirtytwo 2025-12-312:561 reply

                          >This is nonsense. No AI is going to produce what someone wants without telling it exactly what to do and how to do it, so yes, it will always need hand holding, unless you like slurping up slop. I don't know you, if you aren't a bot, you might just be satisfied with slop? It's a race to the bottom, and it's not going to end up the way you think it will.

                          You're not thinking clearly. A couple years ago we didn't even have AI who could do this, then chatGPT came out we had AI who could barely do it, then we had AI who could do simple tasks with A lot of hand holding, now we have AI who can do complex human tasks with minimal hand holding. Where do you think the trendline is pointing.

                          Your hypothesis is going against all evidence. It's more wishful thinking and irrational. It's a race to the bottom because you wish it will be a race to the bottom, and we both know the trendline is pointing in the opposite direction.

                          >We've seen what social media and every-waking-hour access to tablets and the internet has done to kids - so much harm that some countries have banned social media for people under a certain age. I can see a future where "AI" will also be banned for minors to use, probably pretty soon too. The harms from "AI" being able to placate instead of create should be obvious, and children shouldn't be able to use it without adult supervision.

                          I agree AI is bad for us. My claim is it's going to change the world and it is already replacing human tasks. That's all. Whether that's good or bad for us is an ORTHOGANOL argument.

                          • By leptons 2025-12-3120:39

                            I use AI every day, and it's honestly crap. No it isn't significantly improving, it's hitting a wall. Every new model release is getting less and less good, so no, the "trendline" is not going up as much as you seem to think it is. It's plateaued. The only way "AI" is going to change the world is if stupid people put it in places that it really shouldn't be, thinking it will solve problems and not create even more problems.

                    • By aspenmartin 2025-12-3020:321 reply

                      Proof of what? Should you also have to prove you are not a bot sponsored by short-sellers? It’s all so so silly, anti-AI crowds on HN rehash so many of the same tired arguments it’s ridiculous:

                      - bad for environment: how? Why? - takes all creative output and doesn’t credit: common crawl has been around for decades and models have been training for decades, the difference is that now they’re good. Regurgitating training data is a known issue for which there are mitigations but welcome to the world of things not being as idealistic as some Stallman-esque hellscape everyone seems to want to live in - it’s bad and so no one should use it and any professionals who do don’t know what they’re doing: I have been so fortunate to personally know some of the brightest minds on this planet (Astro departmentments, AI research labs) and majority of them use AI for their jobs.

                      • By leptons 2025-12-3022:272 reply

                        >Should you also have to prove you are not a bot sponsored by short-sellers?

                        On a 35 day-old account, yes. Anything "post-AI" is suspect now.

                        The rest of your comment reads like manufactured AI slop, replying to things I never even wrote in my one sentence comment. And no surprise coming from an account created 1 day ago.

                        • By threethirtytwo 2025-12-3112:351 reply

                          I think it’s quite obvious I’m not writing AI slop.

                          The latest chatgpt for example will produce comments that are now distinguishable from the real thing only because they’re much better written. It’s insane that the main visible marker rn is that the arguments and writings it crafts are superior then what your average joe can write.

                          My shit writing can’t hold a candle and that’s pretty obvious. AI slop is not accepted here but I can post an example of what AI slop will now look like, if AI responded to you it would look like this:

                          Fair to be skeptical of new accounts. But account age and “sounds like AI” are not workable filters for truth. Humans can write like bots, bots can write like humans, and both can be new. That standard selects for tenure, not correctness.

                          More importantly, you did not engage any claim. If the position is simply “post-AI content from new accounts is suspect,” say that as a moderation concern. But as an argument, suspicion alone does not refute anything.

                          Pick one concrete claim and say why it is wrong or what evidence would change your mind. Otherwise “this reads like slop” is just pattern matching. That is exactly the failure mode being complained about.

                          • By leptons 2026-01-0118:05

                            I accused another user of writing AI slop in this specific thread, and here you are inserting yourself as if you are replying to comment I made to the other user. You certainly seem desperate to boost "AI" as much as you can. Your 37 day old account is also just as suspect as their 3 day old account. I'm not engaging with you any more so replying is kind of pointless.

                  • By skydhash 2025-12-3013:011 reply

                    > I’m an expert in what I do. A professional, and few people can do what I do

                    Are you an astronaut?

                    • By threethirtytwo 2025-12-3013:551 reply

                      Obviously not troll, I know I’m bragging. But I have to emphasize that it is not some stupid oh “only domain experts know AI is shit. Everyone else is too stupid to understand how bad it is” That is patently wrong.

                      Few people can do what I do and as a result I likely make more money than you. But now with AI… everyone can do what I do. It has leveled the playing field… what I was before now matters fuck all. Understand?

                      I still make money right now. But that’s unlikely to last very long. I fully expect it to disappear within the next decade.

                      • By dent9 2025-12-3015:241 reply

                        You are wrong. People like yourself will likely be smart enough to stay well employed into the future. It's the folks who are arguing with you trying to say that AI is useless who will quickly lose their jobs. And they'll be all shocked Pikachu face when they get a pink slip while their role gets reassigned to an AI agent

                        • By elktown 2025-12-3016:002 reply

                          > It's the folks who are arguing with you trying to say that AI is useless who will quickly lose their jobs.

                          Why is it that in every hype there are always the guys like you that want to punish the non-believers? It's not enough to be potentially proven correct, your anger requires the demise of the heretics. It was the same story for cryptocurrencies.

                          • By hansmayer 2025-12-3017:521 reply

                            He/she is probably one of those poor souls working for an AI-wrapper-startup who received a ton of compensation in "equity", which will be worth nothing when their founders get acquihired, Windsurf style ;) But until then, they get to threaten us all with the impending doom, because hey, they are looking into the eye of the storm, writing Very Complex Queries against the AI API or whatever...

                            • By threethirtytwo 2025-12-3018:221 reply

                              Isn’t this the same type of emotional response he’s getting accused for? You’re speculating that he will be “punished” just as he speculated for you.

                              There’s emotions on both sides and the goal is to call it out, throw it to the side and cut through into the substance. The attitude should be: Which one of us is actually right? Rather than: I’m right and you’re a fucking idiot attitude I see everywhere.

                              • By hansmayer 2025-12-3019:101 reply

                                Mate, I could not care less if he/her got "punished" or not. I was just assuming what might be driving someone to go and try and answer each and every one of my posts with very low quality comments, reeking of desperation and "elon-style" humour (cheap, cringe puns). You are assuming too much here.

                                • By threethirtytwo 2025-12-3019:56

                                  Maybe he was just assuming something negative as well.

                                  Both certainly look very negative and over the top.

                          • By threethirtytwo 2025-12-3018:19

                            Not too dissimilar to you. I wrote long rebuttals to you points and you just descended into put downs, stalking and false accusations. You essentially told me to fuck off from all of HN in one of your posts.

                            So it’s not like your anger is any better.

                • By dent9 2025-12-3015:201 reply

                  Bro idk why you waste your time writing all this. No one cares that you were an early adopter, all that means is that you used the rudimentary LLM implementations that were available from 2022-2024 which are now completely obselete. Whatever experience you think you have with AI tools is useless because you clearly haven't kept up with the times. AI platforms and tools have been changing quickly. Every six months the capabilities have massively improved.

                  Next time before you waste ten minutes typing out these self aggrandizing tirades maybe try asking the AI to just write it for you instead

                  • By 63stack 2025-12-3023:41

                    Maybe he's already ahead of you by not using current models, 2026 models are going to make 2025 models completely obsolete, wasting time on them is dumb.

                • By satisfice 2025-12-305:53

                  Hear hear!

                • By llmslave2 2025-12-306:401 reply

                  This is such a fantastic response. And outsiders should very well be made aware what kind of plane they are stepping into. No offence to the aviation enthusiasts in your example but I will do everything in my power to avoid getting on their plane, in the same way I will do everything in my power to avoid using AI coded software that does anything important or critical...

                  • By andrekandre 2025-12-3018:53

                      > but I will do everything in my power to avoid getting on their plane
                    
                    speaking of airplanes... considering how much llm usage is being pushed top-down in many places, i wonder how long until some news drops of some catastrophic one-liner got through via llm generated code...

            • By threethirtytwo 2025-12-307:012 reply

              Are you joking? You realize entire companies and startups are littered with ppl who only use AI.

              • By hansmayer 2025-12-308:081 reply

                > littered with ppl who only use AI

                "Littered" is a great verb to use here. Also I did not ask for a deviated proxy non-measure, like how many people who are choking themselves to death in a meaningless bullshit job are now surviving by having LLMs generate their spreadsheets and presentations. I asked for solid proof of succesful, commercial products built up by dreaming them up through LLMs.

                • By threethirtytwo 2025-12-3014:061 reply

                  The proof is all around you. I am talking about software professionals not some bullshit spread sheet thing.

                  What I’m saying is this: From my pov Everyone is using LLMs to write code now. The overwhelming majority of software products in existence today are now being changed with LLM code.

                  The majority of software products being created from scratch are also mostly LLM code.

                  This is obvious to me. It’s not speculation, where I live and where I’m from and where I work it’s the obvious status quo. When I see someone like you I’m thinking because the change happened so fast you’re one of the people living in a bubble. Your company and the people around you haven’t started using it because the culture hasn’t caught up.

                  Wait until you have that one coworker who’s going at 10x speed as everyone else and you find out it’s because of AI. That is what will slowly happen to these bubbles. To keep pace you will have to switch to AI to see the difference.

                  I also don’t know how to offer you proof. Do you use google? If so you’ve used products that have been changed by LLM code. Is that proof? Do you use any products built by a start up in the last year? The majority of that code will be written by an LLM.

                  • By hansmayer 2025-12-3017:191 reply

                    > Your company and the people around you haven’t started using it because the culture hasn’t caught up.

                    We have been using LLMs since 2021, if I havent repeated that enough in these threads. What culture do I have to catch up with? I have been paying top tier LLM models for my entire team since it became an option. Do you think you are proselytizing to the un-initiated here? That is a naive view at best. My issue is that the tools are at best a worse replacement for the pre-2019 google search and at worst a huge danger in the hands of people who dont know what they are doing.

                    • By threethirtytwo 2025-12-3018:37

                      Doesn’t make sense to me. If it’s bad why pay for the tool?

                      Obviously your team disagrees that it’s a worse replacement for google or else why demand it against your will?

                      > at worst a huge danger in the hands of people who dont know what they are doing.

                      I agree with this. But the upside negates this and I agree with your own team on that.

                      Btw if you’re paying top dollar for AI.. your developers are unlikely using it as a google search replacement. At top dollar AI is used as an agent. What it ends up doing is extremely different from a google search in this mode. That may be good or bad but it is a distinctly different outcome then a google search and that makes your google analogy ill fitted to what your team is actually using it for.

            • By dent9 2025-12-3015:081 reply

              Have you had your head in the sand for the past two years?

              At the recent AWS conference, they were showcasing Kiro extensively with real life products that have been built with it. And the Amazon developers all allege that they've all been using Kiro and other AI tools and agents heavily for the past year+ now to build AWS's own services. Google and Microsoft have also reported similar internal efforts.

              The platforms you interact with on a daily basis are now all being built with the help of AI tools and agents

              If you think no one is building real commercial products with AI then you are either blind or an idiot or both. Why don't you just spend two seconds emailing your company AWS ProServe folks and ask them, I'm sure they'll give you a laundry list of things they're using AI for internally and sign you up for a Kiro demo as well

              • By 63stack 2025-12-3023:451 reply

                Amazon, Google and Microsoft are balls deep invested in AI, a rational person should draw 0 conclusions in them showcasing how productive they are with it.

                I'd say it's more about the fear of their $50billion+ investments not paying off is creeping up on them.

                • By aspenmartin 2025-12-3114:10

                  It’s ok to have this prior but these are not speculative tools and capabilities, they exist today. If you remain unimpressed by them that’s fine, but to deny real people (not bots!) and real companies (we measure lots of stuff, I’ve seen the data at a large MAANG and have used their internal and external tools) get serious benefits _today_ and we still have about 4 more orders of magnitude to scale _existing_ paradigms, the writing on the wall is so obvious. It’s fine and reasonable to be skeptical and there are so many serious serious societal risks and issues to worry about and champion but to me if your position is akin to “this is all hype” it makes absolutely no sense to me

            • By CuriouslyC 2025-12-305:472 reply

              I'm sure you're interacting with a ton of tools built via agents, ironically even in software engineering people are trying to human-wash AI code due to anti-AI bias by people who should know better (if you think 100% of LLM outputs are "slop" with no quality consideration factored in, you're hopelessly biased). The commercialized seems like an arbitrary and pointless bar, I've seen some hot garbage that's "commercialized" and some great code that's not.

              • By andsoitis 2025-12-306:58

                > The commercialized seems like an arbitrary and pointless bar

                The point is that without mentioning specific software that readers know about, there isn’t really a way to evaluate a claim of 20x.

              • By hansmayer 2025-12-3011:031 reply

                > I'm sure you're interacting with a ton of tools built via agents, ironically even in software engineering people are trying to human-wash AI code due to anti-AI bias

                Please just for fun - reach out to for example Klarna support via their website and tell me how much of your experience can be attributed to an anti-AI bias and how much to the fact that the LLMs are a complete shit for any important production use cases.

                • By dent9 2025-12-3015:261 reply

                  My man here is reaching out to Klarna Support, this tells a LOT about his life decision making skills which clearly shine through as well in his comments on the topic of AI

                  • By 63stack 2025-12-3023:49

                    Klarna functions as a payment provider as well, not just a payday loan service (which you are implying I assume). This comment says more about you.

          • By aspenmartin 2025-12-2922:441 reply

            Who is saying anything about 20x? Sorry did I miss something here?

            • By jayd16 2025-12-2922:46

              > work of an entire org that used to need twenty engineers.

              From the OP. If you think that's too much then we agree.

          • By doug_durham 2025-12-2922:43

            You’ve never read Simon Willison’s blog? His repo is full of work that he’s created with LLM’s. He makes money off of them. There are plenty of examples you just need to look.

        • By threethirtytwo 2025-12-306:591 reply

          The paradigm shift hit the world like a wall. I know entire teams where the manager thinks AI is bullshit and the entire team is not allowed to use AI.

          I love coding. But reality is reality and these fools just aren’t keeping pace with how fast the world is changing.

          • By goatlover 2025-12-3011:231 reply

            Or we're in another hype cycle and billions of dollars are being pumped in to sustain the current bubble with a lot of promises about how fast the world is changing. Doesn't mean AI can't be a useful tool.

            • By aspenmartin 2025-12-3021:18

              When people say “hype cycle” that can mean so many different things. That valuations are too high and many industry “promises” are wrong is maybe true but to me it’s irrelevant, this isn’t speculative, I think most posters who are positive on agents in these threads are talking about two things: current, existing tools, and the existing rate of progress. Check out e.g. Epoch.ai for great industry analyses. To compare AI to crypto is disingenuous, they are completely different and crypto is a technology that fundamentally makes no sense in a world where governments want to (and arguably should) control money supply. You may or may not agree on that take but AI is something that governments will push aggressively and see as crucial to national security/control. It means this is not going away

        • By hansmayer 2025-12-2922:305 reply

          > I’ve done things with Claude I never thought possible for myself to do,

          That's the point champ. They seem great to people when they apply them to some domain they are not competent it, that's because they cannot evaluate the issues. So you've never programmed but can now scaffold a React application and basic backend in a couple of hours? Good for you, but for the love of god have someone more experienced check it before you push into production. Once you apply them to any area where you have at least moderate competence, you will see all sorts of issues that you just cannot unsee. Security and performance is often an issue, not to mention the quality of code....

          • By cmrdporcupine 2025-12-2922:43

            This is remarkably dismissive and comes across as arrogant. In reality they assist many people with expert skills in a domain in getting things done in areas they are competent in, without getting bogged down in tedium.

            They need a heavy hand to police to make sure they do the right thing. Garbage in, garbage out.

            The smarter the hand of the person driving them, the better the output. You see a problem, you correct it. Or make them correct it. The stronger the foundation they're starting from, the better the production.

            It's basically the opposite of what you're asserting here.

          • By wiseowise 2025-12-307:361 reply

            > So you've never programmed but can now scaffold a React application and basic backend in a couple of hours?

            Ahaha, weren’t you the guy who wrote an opus about planes? Is this your baseline for “stuff where LLMs break and real engineering comes into the room”? There’s a harsh wake up call for you around the corner.

            • By hansmayer 2025-12-307:551 reply

              What wake up call mate? I've been on board as early adopter with GH Copilot closed beta since 2021, it was around time when you did not even hear about the LLMs. I am just being realistic about the limits of the technology. In the 90s, we did not need to convince people about the Internet. It just worked. Also - what opus? Have the LLMs affected your attention span so much, that you consider what typically an primary school first-grader would read during their first school class, an "opus" no less? No wonder you are so easily impressed.

              • By matthewmacleod 2025-12-308:261 reply

                I expect it’s your “I’m an expert and everyone else is merely an idiot child” attitude that’s probably making it hard to take you seriously.

                And don’t get me wrong - I totally understand this personality. There are a similar few I’ve worked with recently who are broadly quite skeptical of what seems to be an obvious fact to me - their roles will need to change and their skillsets will have to develop to take advantage of this new technology.

                • By hansmayer 2025-12-3010:472 reply

                  I am a bit tired of explaining, but I run my own company, so its not like I have to fear my "roles and responsibilities" changing - I am designing them myself. I also am not a general skeptic of the "YAGNI" type - my company and myself have been early adopters on many trends. Those that made sense of course. We also tried to be early adopters of LLMs, all the way since 2021. And I am sorry if that sounds arrogant to you, but anyone still working on them and with them to me looks like the folks who were trying to build computers and TVs with the vaccuum tubes. With the difference that vaccuum tubes computers were actually useful at the time.

                  • By dent9 2025-12-3015:31

                    95% of companies fail. Yours will too, don't worry. Amazon themselves have already been using in-house versions of this to build AWS for over a year https://kiro.dev/ you can either continue adopting AI in your company or you can start filing your company bankruptcy papers

                  • By cheevly 2025-12-3014:11

                    What would you need to see to change your mind? I can generate at mind-boggling scale. What’s your threshold for realizing you might not have explored every possible vector for AI capabilities?

          • By oooyay 2025-12-305:11

            > That's the point champ.

            Friendly reminder that this style of discourse is not very welcome on HN: https://news.ycombinator.com/newsguidelines.html

          • By threethirtytwo 2025-12-307:084 reply

            What you wrote here was relevant about 9 months ago. It’s now outdated. The pace and velocity of improvement of Ai can only be described as violent. It is so fast that there are many people like you who don’t get it.

            • By leptons 2025-12-309:181 reply

              The last big release from OpenAI was a big giant billion-dollar flop. Its lackluster update was written about far and wide, even here on HN. But maybe you're living in an alternate reality?

              • By threethirtytwo 2025-12-3014:341 reply

                I use Claude code.

                My experience comes from the fact that after over a decade of working as a swe I no longer write code. It’s not some alternate reality thing or reading headlines. It’s my daily life that has changed.

                • By andrekandre 2025-12-3019:061 reply

                    > I no longer write code
                  
                  do you review it before checking it in?

                  • By threethirtytwo 2025-12-314:51

                    Have you used AI before? Agentic systems are set up so it gives you a diff before even making committing to a change. Sounds like you haven’t really used AI agentically yet.

            • By hansmayer 2025-12-308:001 reply

              Yeah, sure buddy :)

              • By baq 2025-12-308:224 reply

                Disrespect the trend line and get rolled over by the steamroller. Labs are cooking and what is available commercially is lobotomized for safety and alignment. If your baseline of current max capability is sonnet 4.5 released just this summer you’re going to be very surprised in the next few months.

                • By hansmayer 2025-12-3010:531 reply

                  Right, like I was steamrolled by the "Team of Pocket Ph.D Experts" announced earlier this year with ChatGPT 5 ? Remember that underwhelming experience? The Grok to which you could "paste your entire source code file"? The constantly debilitating Claude models? Satya Nadella desperately dropping down to a PO role and bypassing his executives to try and micro-manage Copilot product development because the O365 Copilot experience is experiencing a MASSIVE pushback globally from teams and companies forced to use it ? Or is there another steamrolling coming around? What is this time? Zuckerberg implements 3D avatars in a metaverse with legs that can walk around and talk to us via LLMs? And then they sit down at virtual desks and type on virtual keyboards to produce software? Enlighten me please!

                  • By threethirtytwo 2025-12-3014:271 reply

                    First examine your post. Can you create a 3D avatar with legs that can walk and talk?

                    If not, then for this area you’ve been steam rolled.

                    Anyway main point is, you’re looking at the hype headlines which are ludicrous. Where most optimists come from is that they are using it in the daily to code. To them it’s right in front of their eyes.

                    I’m not sure what your experience is but my opinion on AI doesn’t come from speculation. It comes from on the ground experience on how AI currently has changed my job role completely. If I hold the technology to be fixed and to not improve into the future then my point still stands. I’m not speculating. Most AI optimists aren’t speculating.

                    The current on the ground performance is what’s causing the divide. Some people have seen it fully others only have a rudimentary trial.

                    • By elktown 2025-12-3014:582 reply

                      I have a hard time trusting the judgement of someone writing this:

                      > I no longer write code. I’ve been a swe for over a decade. AI writes all my code following my instructions. My code output is now expected to be 5x what it was before because we are now augmented by AI. All my coworkers use AI. We don’t use ChatGPT we use anthropic. If I didn’t use AI I would be fired for being too slow.

                      https://news.ycombinator.com/item?id=46175628

                      • By baq 2025-12-3015:181 reply

                        You should drop the prejudice and focus to be aware of the situation. This is happening all over the world, most people who have crossed this bridge just don’t share, just like they don’t share that they’ve brushed their teeth this morning.

                        • By elktown 2025-12-3015:471 reply

                          I think I'll keep defaulting to critical thinking rather than some kinda pseudo-religious "crossing the bridge" talk.

                          • By baq 2025-12-3017:001 reply

                            Just a metaphore - used to code by hand, now he doesn't, but still produces software. Keep religion out of this.

                            • By elktown 2025-12-3017:082 reply

                              No one shrugs off 5x like brushing one's teeth in the morning. That makes no sense.

                              • By baq 2025-12-3022:35

                                You're confusing critical thinking with having an axe to grind it seems. Bye.

                              • By threethirtytwo 2025-12-3017:221 reply

                                People are sharing it. Look at this entire thread. It’s so conflicted.

                                We have half the thread saying it’s 5x and the other half saying they’re delusional and lack critical thinking.

                                I think it’s obvious who lacks critical thinking. If half the thread is saying on the ground AI has changed things and the other half just labels everyone as crazy without investigation… guess which one didn’t do any critical thinking?

                                Last week I built an app that cross compiled into Tauri and electron that’s essentially a google earth clone for farms. It uses mapbox and deckgl and you can play back gps tracks of tractor movements and the gps traces change color as the tractor moves in actual real time. There’s pausing, seeking, bookmarking, skipping. All happening in real time because it’s optimized to use shader code and uniforms to do all these updates rather than redrawing the layers. There’s also color grading for gps fix values and satellite counts which the user can switch instantaneously to with zero slow down on tracks with thousands and thousands of points. It all interfaces with an API that scans gcp storage for gps tracks and organizes it into a queryable api that interfaces with our firebase based authentication. The backend is deployed by terraform and written in strictly typed typescript and it’s automatically deployed and checked by GHA. Of course the electron and tauri app have GUI login interfaces that work fully correctly with the backend api and it all looks professionally designed like a movie player merged with Google earth for farm orchards.

                                I have rudimentary understanding for many of the technologies involved in the above. But I was able to write that whole internal tool in less than a week thanks to AI. I couldn’t have pulled it off without rudimentary understanding of the tech so some novice swe couldn’t really do it without the optimizations I used but that’s literally all I needed. I never wrote shader code for prod in my life and left to its own devices the AI would have come up with an implementation that’s too laggy to work properly.

                                That’s all that’s needed. Some basic high level understanding and AI did everything else and now our company has an internal tool that is polished beyond anything that would’ve been given effort to before AI.

                                I’m willing to bet you didn’t use AI agents in a meaningful way. Maybe copying and pasting some snippets of code into a chatbot and not liking the output. And then you do it every couple of weeks to have your finger on the pulse of AI.

                                Go deeper. Build an app with AI. Hand hold it into building something you never built before. It’s essentially a pair programming endeavor. Im willing to bet you haven’t done this. Go in with the goal of building something polished and don’t automatically dismiss it when the AI does something stupid (it inevitably will) Doing this is what actual “critical thinking” is.

                                • By elktown 2025-12-3017:381 reply

                                  > I think it’s obvious who lacks critical thinking.

                                  My critical thinking is sharp enough to recognize that you're the recently banned ninetyninenine user [0]. Just as unbalanced and quarrelsome as before I can see. It's probably better to draw some conclusion from a ban and adjust, or just leave.

                                  [0] https://news.ycombinator.com/item?id=45988923

                                  • By threethirtytwo 2025-12-3017:491 reply

                                    I’m not that guy lol.

                                    Why don’t you respond to my points rather than attack me.

                                    • By elktown 2025-12-3018:481 reply

                                      > Why don’t you respond to my points

                                      Because I believe you have a "flexible" relationship to the truth, so I'm not wasting any more time.

                      • By threethirtytwo 2025-12-3016:471 reply

                        Explain to me why my judgement is flawed. What I’m saying is true.

                        • By elktown 2025-12-3017:011 reply

                          Because, among other claims, "5x now or you're fired!" is completely ridiculous.

                          • By threethirtytwo 2025-12-3017:151 reply

                            Bro no one said 5x now or your fired that’s your own imagination adding flavor to it.

                            It’s obvious to anyone if your output is 5x less than everyone else you will eventually be let go. There’s no paradigm shift where the boss suddenly announced that. But the underlying unsaid expectation is obvious given what everyone is doing.

                            What happened was this, a couple new hires and some current employees started were using AI. There output was magnified and they were not only having more output but they were deploying code outside their areas of expertise doing dev ops, infra, backend and frontend.

                            This spread and within months everyone in the company was doing it. The boss can now throw a frontend job to a backend developer and now expect completion in a day or less. This isn’t every task but such output for the majority of tasks it’s normal.

                            If you’re not meeting that norm it’s blindingly obvious. The boss doesn’t need to announce anything when everyone is faster. There was no deliberate culture shift where the boss announced it. The closest equivalent is the boss hiring a 10x engineer to work alongside you and you have to scramble to catch up. The difference is now we know exactly what is making each engineer 10x and we can use that tool to also operate at that level.

                            Critical thinking my ass. You’re just labeling and assuming things with your premeditated subconscious bias. If anything it’s your perspective that is religious.

                            • By elktown 2025-12-3017:312 reply

                              > they were deploying code outside their areas of expertise doing dev ops, infra, backend and frontend.

                              > The boss can now throw a frontend job to a backend developer and now expect completion in a day or less.

                              Right. So essentially vibe coding in unknown domains, sounds great. Truly professional.

                              • By threethirtytwo 2025-12-3017:581 reply

                                Also can you please stop stalking me and just respond to my points instead of digging through my whole profile and attempting to do character assassinations based off of what I wrote in the past? Thanks.

                              • By threethirtytwo 2025-12-3017:52

                                Whether you agree with it or not is besides the point. The point is it’s happening.

                                Your initial stance was disbelief. Now you’re just looking down at it as unprofessional.

                                Bro, I fucking agree. It’s unprofessional. But the entire point initially was that you didn’t believe it and my objective was to tell you that this is what’s happening in reality. Scoff at it all you want, as AI improves less and less “professional” people will be able to enter our field and operate at the same level as us.

                • By threethirtytwo 2025-12-3014:31

                  He won’t be steam rolled. But he will eat his words.

                • By leptons 2025-12-309:20

                  meh. I'll believe it when I see it. We've been promised so many things in this space, over and over, that never seem to materialize.

                • By llmslave2 2025-12-3010:131 reply

                  I don't understand this idea that non-believers will be "steamrolled" by those who are currently adopting AI into their workflows. If their claims are validated and the new AI workflows end up achieving that claimed 10x productivity speedup, or even a 2x speedup, nobody is cursed to be steamrolled - they'll simply adopt those same workflows same as everyone else. In the meantime they aren't wasting their time trying to figure out the best way to coax and beg the LLM's into better performance.

                  • By baq 2025-12-3012:39

                    That's actually what I'm arguing for; use tools where they are applicable. I'm against blind contrarianism and the 'nothing ever happens' attitude since that IME is being proven more wrong each week.

            • By meindnoch 2025-12-3013:20

              Sure. Just hurry up bro, because Kurzweil is not getting any younger.

            • By goatlover 2025-12-3011:30

              Right, the Singularity will be here any day now. We can all just sit back and collect our UBI while plugging into the Matrix. /s

          • By aspenmartin 2025-12-2922:382 reply

            Seems fine, works, is fine, is better than if you had me go off and write it on my own. You realize you can check the results? You can use Claude to help you understand the changes as you read through them? I mean I just don’t get this weird “it makes mistakes and it’s horrible if you understand the domain that it is generating over” I mean yes definitely sometimes and definitely not other times. What happens if I DONT have someone more experienced to consult with or that will ignore me because they are busy or be wrong because they are also imperfect and not focused. It’s really hard to be convinced that this point of view is not just some knee jerk reaction justified post hoc

            • By hansmayer 2025-12-2923:181 reply

              Yes you can ask them "to check it for you". The only little problem is as you said yourself "they make mistakes", therefore : YOU CANNOT TRUST THEM. Just because you tell them to "check it" does not mean they will get it right this time. Again, however it seems "fine" to you, please, please, please / have a more senior person check that crap before you inflict serious damage somewhere.

              • By aspenmartin 2025-12-300:051 reply

                Nope, you read their code, ask them to summarize changes to guide your reading, ask it why it made certain decisions you don’t understand and if you don’t like their explanations you change it (with the agent!). Own and be responsible for the code you commit. I am the “most senior”, and at large tech companies that track, higher level IC corresponds to more AI usage, hmm almost like it’s a useful tool.

                • By bccdee 2025-12-305:022 reply

                  Ok but you understand that the fundamental nature of LLMs amplifies errors, right? A hallucination is, by definition, a series of tokens which is plausible enough to be indistinguishable from fact to the model. If you ask an LLM to explain its own hallucinations to you, it will gladly do so, and do it in a way that makes them seem utterly natural. If you ask an LLM to explain its motivations for having done something, it will extemporize whichever motivation feels the most plausible in the moment you're asking it.

                  LLMs can be handy, but they're not trustworthy. "Own and be responsible for the code you commit" is an impossible ideal to uphold if you never actually sit down and internalize the code in your code base. No "summaries," no "explanations."

                  • By dent9 2025-12-3015:35

                    So your argument is that if people don't use the tool correctly they might get incorrect results? How is that relevant? If you Google search for the wrong query you'll similarly get incorrect results

      • By davnicwil 2025-12-2922:255 reply

        I would say while LLMs do improve productivity sometimes, I have to say I flatly cannot believe a claim (at least without direct demonstration or evidence) that one person is doing the work of 20 with them in december 2025 at least.

        I mean from the off, people were claiming 10x probably mostly because it's a nice round number, but those claims quickly fell out of the mainstream as people realised it's just not that big a multiplier in practice in the real world.

        I don't think we're seeing this in the market, anywhere. Something like 1 engineer doing the job of 20, what you're talking about is basically whole departments at mid sized companies compressing to one person. Think about that, that has implications for all the additional management staff on top of the 20 engineers too.

        It'd either be a complete restructure and rethink of the way software orgs work, or we'd be seeing just incredible, crazy deltas in output of software companies this year of the type that couldn't be ignored, they'd be impossible to not notice.

        This is just plainly not happening. Look, if it happens, it happens, 26, 27, 28 or 38. It'll be a cool and interesting new world if it does. But it's just... not happened or happening in 25.

        • By jmogly 2025-12-2922:531 reply

          I would say it varies from 0x to a modest 2x. It can help you write good code quickly, but, I only spent about 20-30% of my time writing code anyway before AI. It definitely makes debugging and research tasks much easier as well. I would confidently say my job as a senior dev has gotten a lot easier and less stressful as a result of these tools.

          One other thing I have seen however is the 0x case, where you have given too much control to the llm, it codes both you and itself into pan’s labyrinth, and you end up having to take a weed wacker to the whole project or start from scratch.

          • By mattmanser 2025-12-309:343 reply

            Ok, if you're a senior dev, have you 'caught' it yet?

            Ask it a question about something you know well, and it'll give you garbage code that it's obviously copied from an answer on SO from 10 years ago.

            When you ask it for research, it's still giving you garbage out of date information it copied from SO 10 years ago, you just don't know it's garbage.

            • By theshrike79 2025-12-3016:071 reply

              That's why you dont use LLMs as a knowledge source without giving them tools.

              "Agents use tools in a loop to achieve a goal."

              If you don't give any tools, you get hallucinations and half-truths.

              But you give one a tool to do, say, web searches and it's going to be a lot smarter. That's where 90% of the innovation with "AI" today is coming from. The raw models aren't gettin that much smarter anymore, but the scaffolding and frameworks around them are.

              Tools are the main reason Claude Code is as good as it is compared to the competition.

              • By andrekandre 2025-12-3019:141 reply

                  > The raw models aren't gettin that much smarter anymore, but the scaffolding and frameworks around them are.
                
                yes, that is my understanding as well, though it gets me thinking if that is true, then what real value is the llm on the server compared to doing that locally + tools?

                • By theshrike79 2025-12-3020:04

                  You still can't beat an acre of specialized compute with any kind of home hardware. That's pretty much the power of cloud LLMs.

                  For a tool use loop local models are getting to "OK" levels, when they get to "pretty good", most of my own stuff can run locally, basically just coordinating tool calls.

            • By jmogly 2025-12-3014:031 reply

              Of course, step one is always critically think and evaluate for bad information. I think for research, I mainly use it for things that are testable/verifiable, for example I used it for a tricky proxy chain set up. I did try to use it to learn a language a few months ago which I think was counter productive for the reasons you mentioned.

              • By mattmanser 2026-01-0212:46

                How can you critically assess something in a field you're not already an expert on?

                That Python you just got might look good, but could be rewritten from 50 lines to 5, it's written in 2010-style, it's not using modern libraries, it's not using modern syntax.

                And it is 50 to 5. That is the scale we're talking about in a good 75% of AI produced code unless you challenge it constantly. Not using modern syntax to reduce boilerplate, over-guarding against impossible state, ridiculous amounts of error handling. It is basically a junior dev on steriods.

                Most of the time you have no idea that most of that code is totally unnecessary unless you're already an expert in that language AND libraries it's using. And you're rarely an expert in both or you wouldn't even be asking as it would have been quicker to write the code than even write the prompt for the AI.

            • By skydhash 2025-12-3013:14

              I use web search (DDG) and I don’t think I have ever try more than one queries in the vast majority of cases. Why because I know where the answer is, I’m using the search engine as an index to where I can find it. Like “csv python” to find that page in the doc.

        • By CuriouslyC 2025-12-305:383 reply

          It's entirely dependent on the type of code being written. For verbose, straightforward code with clear cut test scenarios, one agent can easily 24/7 the work of 20 FT engineers. This is a best case scenario.

          Your productivity boost will depend entirely on a combination of how much you can remove yourself from the loop (basically, the cost of validation per turn) and how amenable the task/your code is to agents (which determines your P(success)).

          Low P(success) isn't a problem if there's no engineer time cost to validation, the agent can just grind the problem out in the background, and obviously if P(success) is high the cost of validation isn't a big deal. The productivity killer is when P(success) is low and the cost of validation is high, these circumstances can push you into the red with agents very quickly.

          Thus the key to agents being a force multiplier is to focus on reducing validation costs, increasing P(success) and developing intuition relating to when to back off on pulling the slot machine in favor of more research. This is assuming you're speccing out what you're building so the agent doesn't make poor architectural/algorithmic choices that hamstring you down the line.

          • By qubitcoder 2025-12-307:381 reply

            Respectfully, if I may offer constructive criticism, I’d hope this isn’t how you communicate to software developers, customers, prospects, or fellow entrepreneurs.

            To be direct, this reads like a fluff comment written by AI with an emphasis on probability and metrics. P(that) || that.

            I’ve written software used by a local real estate company to the Mars Perseverance rover. AI is a phenomenally useful tool. But be weary of preposterous claims.

            • By CuriouslyC 2025-12-3014:27

              I'll take you at your word regarding respectfully. That was an off the cuff attempt to explain the real levers that control the viability of agents under particular circumstances. The target market wasn't your average business potato but someone who might care about a hand waived "order approximate" estimator kind of like big-O notation, which is equally hand waivey.

              Given that, if you want to revisit your comment in a constructive way rather than doing an empty drive by, I'll read your words with an open mind.

          • By yellow_lead 2025-12-307:25

            > It's entirely dependent on the type of code being written. For verbose, straightforward code with clear cut test scenarios, one agent can easily 24/7 the work of 20 FT engineers. This is a best case scenario.

            So the "verbose, straightforward code with clear cut test scenarios" is already written by a human?

          • By 63stack 2025-12-3023:52

            >For verbose, straightforward code with clear cut test scenarios, one agent can easily 24/7 the work of 20 FT engineers

            I have been working professionally for ~16 years in software development, and scenarios like this was about 5% of my work.

        • By EagnaIonat 2025-12-305:17

          > I mean from the off, people were claiming 10x probably mostly because it's a nice round number,

          Purely anecdotal, but I've seen that level of productivity from the vibe tools we have in my workplace.

          The main issue is that 1 engineer needs to have the skills of those 20 engineers so they can see where the vibe coding has gone wrong. Without that it falls apart.

        • By emseetech 2025-12-304:45

          Could be speed/efficiency was the wrong dimension to optimize for and its leading the industry down a bad path.

          An LLM helps most with surface area. It expands the breadth of possibilities a developer can operate on.

        • By andrekandre 2025-12-3019:10

            > one person is doing the work of 20 with them in december 2025 at least
          
          it reminds me of oop hype from the 90's, but maybe indeed it will eventually be true this time...?

      • By coderenegade 2025-12-2923:42

        My experience is that you get out what you put in. If you have a well-defined foundation, AI can populate the stubs and get it 95% correct. Getting to that point can take a bit of thought, and AI can help with that, too, but if you lean on it too much, you'll get a mess.

        And of course, getting to the point where you can write a good foundation has always been the bulk of the work. I don't see that changing anytime soon.

      • By dent9 2025-12-3014:262 reply

        This is completely wrong. Codex 5.2 and Claude Sonnet 4.5 don't have any of these issues. They will regularly tell you that you're wrong if you bother to ask them and they will explain why and what a better solution is. They don't make up anything. The code they produce is noticeably more efficient in LoC than previous models. And yes they really will do research, they will search the Internet for docs and articles as needed and cite their references inline with their answers.

        You talk as if you haven't used a LLM since 2024. It's now almost 2026 and things have changed a lot.

        • By claytongulick 2025-12-3015:401 reply

          With apologies, and not GP, but this has been the same feedback I've personally seen on every single model release.

          Whenever I discuss the problems that my peers and I have using these things, it's always something along the lines of "but model X.Y solves all that!", so I obediently try again, waste a huge amount of time, and come back to the conclusion that these things aren't great at generation, but they are fantastic at summarization and classification.

          When I use them for those tasks, they have real value. For creation? Not so much.

          I've stopped getting excited about the "but model X.Y!!" thing. Maybe they are improving? I just personally haven't seen it.

          But according to the AI hypers, just like with every other tech hype that's died over the past 30 years, "I must just be doing it wrong".

          • By dent9 2025-12-3016:26

            A lot of people are consistently getting their low expectations disproven when it comes to progress in AI tooling. If you read back in my comment history, six months ago I was posting about how AI is over hyped BS. But I kept using it and eventually new releases of models and tools solved most of the problems I had with them. If it has not happened for you yet then I expect it will eventually. Keep up with using the tools and models and follow their advancements and I think you'll eventually get to the point where your needs are met

        • By 63stack 2025-12-3023:55

          The same response (you are using model X instead of Y) have been perpetuated since 2024, and will still be perpetuated in 2026.

      • By to11mtm 2025-12-2922:452 reply

        I'd be willing to give you access to the experiment I mentioned in a separate reply (have a github repo), as far as the output that you can get for a complex app buildout.

        Will admit It's not great (probably not even good) but it definitely has throughput despite my absolute lack of caring that much [0]. Once I get past a certain stage I am thinking of doing an A-B test where I take an earlier commit and try again while paying more attention... (But I at least want to get where there is a full suite of UOW cases before I do that, for comparison's sake.)

        > Those twenty engineers must not have produced much.

        I've been considered a 'very fast' engineer at most shops (e.x. at multiple shops, stories assigned to me would have a <1 multiplier for points[1])

        20 is a bit bloated, unless we are talking about WITCH tier. I definitely can get done in 2-3 hours what could take me a day. I say it that way because at best it's 1-2 hours but other times it's longer, some folks remember the 'best' rather than median.

        [0] - It started as 'prompt only', although after a certain point I did start being more aggressive with personal edits.

        [1] - IDK why they did it that way instead of capacity, OTOH that saved me when it came to being assigned Manual Testing stories...

        • By imron 2025-12-306:10

          > Will admit It's not great (probably not even good) but it definitely has throughput

          Throughput without being good will just lead to more work down the line to correct the badness.

          It's like losing money on every sale but making up for it with volume.

        • By notpachet 2025-12-3010:21

          > Will admit It's not great (probably not even good)

          You lost me here. Come back when you're proud of it.

      • By photios 2025-12-308:152 reply

        Ok, let's say the 20 devs claim is false [1]. What if it's 2? I'd still learn and use the tech. Wouldn't you?

        [1] I actually think it might be true for certain kinds of jobs.

        • By bloppe 2025-12-308:361 reply

          It's not 20 and it's not 2. It's not a person. It's a tool. It can make a person 100x more effective at certain specific things. It can make them 50% less effective at other things. I think, for most people and most things, it might be like a 25% performance boost, amortized over all (impactful) projects and time, but nobody can hope to quantify that with any degree of credibility yet.

          • By andrekandre 2025-12-3019:17

              > but nobody can hope to quantify that with any degree of credibility yet
            
            i'd like to think if it was really good, we would see product quality improve over time; iow less reported bugs, less support incidents, increased sign-ups etc, that could easily be quantified no?

        • By BirdieNZ 2025-12-308:23

          Jevon's Paradox: more software will be produced, rather than fewer software engineers being employed.

      • By sh4rks 2025-12-308:38

        Post model

    • By andsoitis 2025-12-306:47

      > I’m basically just the conductor of all those processes.

      Orchestrating harmony is no mean feat.

    • By otabdeveloper4 2025-12-308:59

      AI is absolutely rock-bottom shit at all that.

    • By ravenstine 2025-12-2921:33

      Yeah, it makes me wonder whether I should start learning to be a carpenter or something. Those who either support AI or thinks "it's all bullshit" cite a lack of evidence for humans truly being replaced in the engineering process, but that's just the thing; the unprecedented levels of uncertainty make it very difficult to invest one's self in the present, intellectually and emotionally. With the current state of things, I don't think it's silly to wonder "what's the point" if another 5 years of this trajectory is going to mean not getting hired as a software dev again unless you have a PhD and want to work for an AI company.

      What doesn't help is that the current state of AI adoption is heavily top-down. What I mean is the buy-in is coming from the leadership class and the shareholder class, both of whom have the incentive to remove the necessary evil of human beings from their processes. Ironically, these classes are perhaps the least qualified to decide whether generative AI can replace swathes of their workforce without serious unforeseen consequences. To make matters worse, those consequences might be as distal as too many NEETs in the system such that no one can afford to buy their crap anymore; good luck getting anyone focused on making it to the next financial quarter to give a shit about that. And that's really all that matters at the end of the day; what leadership believes, whether or not they are in touch with reality.

    • By threethirtytwo 2025-12-306:50

      His logic is off and his experience is irrelevant because i doesn’t encompass scale to have been exposed to an actual paradigm shifting event. Civilizations and entire technologies have been overturned so he can’t say it won’t happen this time.

      What we do know is this. If AI keeps improving at the current rate it’s improving then it will eventually hit a point where we don’t need software engineers. That’s inevitable. The way for it to not happen is for this technology to hit an impenetrable wall.

      This wave of AI came so fast that there are still stubborn people who think it’s a stochastic parrot. They missed the boat.

HackerNews