Labor market impacts of AI: A new measure and early evidence

2026-03-0522:55333566www.anthropic.com

Anthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.

Figure 7: New job starts among workers age 22-25 in occupations with high observed exposure and no AI exposure, Current Population Survey
The top panel shows the percent of young workers starting new jobs in high vs. no exposure occupations. The bottom panel measures the gap between these two series in a difference-in-differences framework.

Apart from some large swings in 2020-2021, these series visually diverge in 2024, with young workers relatively less likely to be hired into exposed occupations. Job finding rates at the less exposed occupations remain stable at 2% per month, while entry into the most exposed jobs decreases by about half a percentage point. The averaged estimate in the post-ChatGPT era is a 14% drop in the job finding rate compared to that in 2022 in the exposed occupations, although this is just barely statistically significant. (There is no such decrease for workers older than 25.)

This may provide some signal of the early effects of AI on employment, and echoes the findings from Brynjolfsson et al. But there are several alternative interpretations. The young workers who are not hired may be remaining at their existing jobs, taking different jobs, or returning to school. A further data-related caveat is that job transitions may be more vulnerable to mismeasurement in surveys.10


Discussion

This report introduces a new measure for understanding the labor market effects of AI and studies impacts on unemployment and hiring. Jobs are more exposed to AI to the extent that their tasks are theoretically feasible with LLMs and observed on our platforms in automated, work-related use cases. We find that computer programmers, customer service representatives, and financial analysts are among the most exposed. Using survey data from the US, we find no impact on unemployment rates for workers in the most exposed occupations, although there’s tentative evidence that hiring into those professions has slowed slightly for workers aged 22-25.

Our work is a first step toward cataloging the impact of AI on the labor market. We hope that the analytical steps taken in this report, especially around coverage and counterfactuals, will be easy to update as new data on employment and AI usage emerge. An established approach may help future observers separate signal from noise.

There are several improvements to be made to the present work. Our usage data will be incorporated in future updates, forming an evolving picture of task and job coverage in the economy. The Eloundou et al. metric could also be updated, to the extent that it is linked to LLM capabilities as of early 2023. And, given the suggestive results around young workers and labor market entrants, a key next step might be to look at how recent graduates with educational credentials in exposed areas are navigating the labor market.

Appendix

Available here.


Acknowledgements

Written by Maxim Massenkoff and Peter McCrory.

With acknowledgements to: Ruth Appel, Tim Belonax, Keir Bradwell, Andy Braden, Dexter Callender III, Miriam Chaum, Madison Clark, Jake Eaton, Deep Ganguli, Kunal Handa, Ryan Heller, Lara Karadogan, Jennifer Martinez, Jared Mueller, Sarah Pollack, David Saunders, Carl De Torres, Kim Withee, and Jack Clark.

We additionally thank Martha Gimbel, Anders Humlum, Evan Rose, and Nathan Wilmers for feedback on earlier versions of this report.


Citation

@online{massenkoffmccrory2026labor,
 author = {Maxim Massenkoff and Peter McCrory},
 title = {Labor market impacts of AI: A new measure and early evidence},
 date = {2026-03-05},
 year = {2026},
 url = {https://www.anthropic.com/research/labor-market-impacts},
}

References

Acemoglu, Daron and Pascual Restrepo, "Robots and Jobs: Evidence from US Labor Markets," Journal of Political Economy, 2020, 128 (6), 2188–2244.

Acemoglu, Daron, David Autor, Jonathon Hazell, and Pascual Restrepo, "Artificial intelligence and jobs: Evidence from online vacancies," Journal of Labor Economics, 2022, 40 (S1), S293–S340.

Appel, Ruth, Maxim Massenkoff, Peter McCrory, Miles McCain, Ryan Heller, Tyler Neylon, and Alex Tamkin, "Anthropic Economic Index report: economic primitives," 2026.

Autor, David H, David Dorn, and Gordon H Hanson, "The China syndrome: Local labor market effects of import competition in the United States," American Economic Review, 2013, 103 (6), 2121–2168.

Autor, David H, & Thompson, N. (2025). Expertise. NBER Working Paper, (w33941).

Blinder, Alan S et al., "How many US jobs might be offshorable?," World Economics, 2009, 10 (2), 41.

Borusyak, Kirill, Peter Hull, and Xavier Jaravel, "Quasi-experimental shift-share research designs," The Review of Economic Studies, 2022, 89 (1), 181–213.

Brynjolfsson, Erik, Bharat Chandar, and Ruyu Chen, "Canaries in the coal mine? six facts about the recent employment effects of artificial intelligence," Digital Economy, 2025.

Eckhardt, Sarah and Nathan Goldschlag, "AI and Jobs: The Final Word (Until the Next One)," Economic Innovation Group (EIG), August 2025. Available at: https://eig.org/ai-and-jobs-the-final-word/

Eloundou, Tyna, Sam Manning, Pamela Mishkin, and Daniel Rock, "Gpts are gpts: An early look at the labor market impact potential of large language models," arXiv preprint arXiv:2303.10130, 2023, 10.

Fujita, S., Moscarini, G., & Postel-Vinay, F. (2024). Measuring employer-to-employer reallocation. American Economic Journal: Macroeconomics, 16(3), 1-51.

Gans, Joshua S. and Goldfarb, Avi, "O-Ring Automation," NBER Working Paper No. 34639, December 2025. Available at SSRN: https://ssrn.com/abstract=5962594

Gimbel, Martha, Molly Kinder, Joshua Kendall, and Maddie Lee, "Evaluating the Impact of AI on the Labor Market: Current State of Affairs," Research Report, The Budget Lab at Yale, New Haven, CT October 2025. Available at: https://budgetlab.yale.edu.

Graetz, Georg and Guy Michaels, "Robots at Work," Review of Economics and Statistics, 2018, 100 (5), 753–768.

Hampole, Menaka, Dimitris Papanikolaou, Lawrence DW Schmidt, and Bryan Seegmiller, "Artificial intelligence and the labor market," Technical Report, National Bureau of Economic Research 2025.

Handa, Kunal, Alex Tamkin, Miles McCain, Saffron Huang, Esin Durmus, Sarah Heck, Jared Mueller, Jerry Hong, Stuart Ritchie, Tim Belonax, Kevin K. Troy, Dario Amodei, Jared Kaplan, Jack Clark, and Deep Ganguli, "Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations," 2025.

Hui, Xiang, Oren Reshef, and Luofeng Zhou, "The short-term effects of generative artificial intelligence on employment: Evidence from an online labor market," Organization Science, 2024, 35 (6), 1977–1989.

Johnston, Andrew and Christos Makridis, "The labor market effects of generative AI: A difference-in-differences analysis of AI exposure," Available at SSRN 5375017, 2025.

Massenkoff, Maxim, "How predictable is job destruction? Evidence from the Occupational Outlook," 2025. Working Paper.

Ozimek, Adam, "Overboard on Offshore Fears," 2019. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3777307

Tamkin, Alex and Peter McCrory, "Estimating AI productivity gains from Claude conversations," 2025.

Tomlinson, K., Jaffe, S., Wang, W., Counts, S., & Suri, S. (2025). Working with AI: measuring the applicability of generative AI to occupations. arXiv preprint arXiv:2507.07935.

  1. Job offshorability: Blinder et al. (2009) and Ozimek (2019); Government growth forecasts: Massenkoff (2025); Robots: Graetz and Michaels (2018) and Acemoglu and Restrepo (2020); China shock: Autor et al. (2013) and Borusyak et al. (2022).

  2. Brynjolfsson et al. (2025) compare employment trends for workers in more versus less AI-exposed occupations, using the task exposure measures from Eloundou et al. (2023) and payroll data from ADP. Johnston and Makridis (2025) do a similar task-based analysis using US administrative data, but they aggregate treatment to the industry level. Hui et al. (2024) study how freelance jobs on Upwork responded to the release of ChatGPT and advanced image generation tools, comparing workers in directly affected categories to those in unaffected categories before and after each tool's release date. Hampole et al. (2025) instrument for firm-level AI adoption using historical university hiring networks: firms that historically recruited from universities whose graduates later entered AI-related roles faced lower adoption costs.

  3. Our task- and occupation-level exposure measures can readily incorporate other usage data, and be extended to different countries. We intend to apply this methodology to new settings over time.

  4. In their framework, “Directly exposed'” tasks were those that could be completed in half the time with an LLM (with a 2,000-word input limit and no access to recent facts). Tasks that were “exposed with tools” were those subject to the same speedup with an LLM that had access to software for, e.g., information retrieval and image processing. Tasks that were not exposed could not have their duration reduced by 50% or more using an LLM.

  5. We use the previous two Anthropic Economic Index datasets, covering usage from August and November 2025. For ONET tasks that are highly semantically similar, we split the counts across them.

  6. There are judgment calls involved at every step. Should the Eloundou et al. (2023) measure enter as {0, 0.5, 1} or something else? What determines "significant" use? How do we handle tasks which seem very similar to those with high usage, but are too rare to have been picked up specifically in the sampling for the Economic Index? How much more should automation workflows count compared to augmentation? A reassuring finding which we expand on in the Appendix is that the Spearman (rank-rank) correlation of job exposure across many resolutions to these questions is exceedingly high.

  7. To match O*NET-SOC codes to occ1990 codes in the CPS, we use the crosswalk provided by Eckhart and Goldschlag (2025).

  8. We explore this further in three ways in the Appendix. First, we ask whether the percentile cutoff that we use to define treatment matters, varying it from the median to the 95th percentile. In all cases, the impact is flat or negative (meaning that unemployment decreases for the exposed group). Next, we focus on young workers in particular, those aged 22 to 25 as in Brynjolfsson et al. (2025). Finally, we use data on unemployment insurance claimants from the Department of Labor to measure the unemployment, rather than CPS survey responses. In no extension do we find clear impacts on exposed jobs.

  9. This range is wide because the authors provide estimates against multiple counterfactuals. The 6 percentage point drop compares to a counterfactual of flat employment growth. The 16 percentage point estimate comes from a design comparing similar workers in the same firm with different occupations.

  10. See Fujita, et al. (2024).


Read the original article

Comments

  • By vb7132 2026-03-0612:368 reply

    I was at a big tech for last 10 years, quit my job last month - I feel 50x more productive outside than inside.

    Here is my take on AI's impact on productivity:

    First let's review what are LLMs objectively good at: 1. Writing boiler plate code 2. Translating between two different coding languages (migration) 3. Learning new things: Summarizing knowledge, explaining concepts 4. Documentation, menial tasks

    At a big tech product company #1 #2 #3 are not as frequent as one would think - most of the time is spent in meetings and meetings about meetings. Things move slowly - it's designed to be like that. Majority devs are working on integrating systems - whatever their manager sold to their manager and so on. The only time AI really helped me at my job was when I did a one-week hackathon. Outside of that, integrations of AI felt like more work rather than less - without much productivity boost.

    Outside, it has proven to be a real productivity boost for me. It checks all the four boxes. Plus, I don't have to worry about legal, integrations, production bugs (eventually those will come).

    So, depends who you are asking -- it is a huge game changer (or not).

    • By neya 2026-03-0618:111 reply

      I think a lot of people will not like to hear this but we use AI almost for everything internally. The noob way to go about this is just give it a couple of tasks and just give it complete root access to your life. That's always going to end up in disappointment. Instead, I realised, AI always needs an architect. Opinionated. Strategic. Authoritative.

      It is quite good at following most orders. Hence why you must ALWAYS be in the loop. AI can augment, but not replace. Maybe some day it might. But it's not now, even with the latest SOTA models.

      I let AI write my emails for me. But never the ability to hit send. I let AI access to my data to make informed decisions, but never let it make the final decision.

      You may think I'm being paranoid, but I'm a very cautious person. I don't jump into new technology fresh out of the oven and this has served me well for the last 15 years. (I learned my lesson courtesy of MongoDb).

      With AI, I am taking the same approach. Experiment, understand the limits and only then implement. Working really well so far and have managed to automate tons of tedious tasks from emails to sales to even meetings.

      I don't use Clawdbot, not any library. I wrote my own wrappers for everything using Elixir. I used Instructor and Ash framework with Phoenix and a bunch of generators to automate tedious tasks. I control the endpoints the models are loaded from (Open router) and use a multi-model flow so no one company has enough data about me. Only bits and pieces of random user IDs.

      Privacy is the real challenge with AI.

      • By sdf2df 2026-03-0618:502 reply

        "I think a lot of people will not like to hear this..."

        Lol why? You've been suckered in and will eventually crash and burn. But carry on.

        Just remember when things go wrong - it's your ass on the line.

        • By neya 2026-03-072:02

          Look, whether you like it or not AI is here and it is decent at some tasks and the world is using it to automate stuff. You saw how Clawdbot exploded, right? Despite users getting hacked left and right didn't stop the adoption. Yesterday there was again a hack incident. It's a burning pain that AI solves to the point where people don't care even if they get hacked.

          Will I crash and burn? Maybe, you're right. But, that's why I'm taking things at a very slow pace. Only automating internal tasks. Only things I trust AI to do. Very very limited scope. What's really my alternative here?

          Just sit back and watch the world move on? My alternative is not changing with the times and being stagnant. That's not really a solution. Even if I'm doing that, I want to have data points that AI is really a dead end instead of just assumptions. My alternative reality isn't a bed or roses - a lot of people at the top do believe they can replace me and my work (CTO) with AI, thanks to the hype. I'm just trying to evolve so I don't become a meme down the line. Can they actually replace me or my job with AI? Absolutely not from what I'm seeing. But hypes of cutting cost is always attractive to people at the top. Just trying to stay alive man, lol.

        • By toteroni 2026-03-0620:391 reply

          how so?

    • By m_ke 2026-03-0613:275 reply

      having been an early employee and founder of a few startups and then working at a few larger companies, most people who only ever worked at FAANG have no idea how much more productive tiny teams with ownership are.

      • By sghiassy 2026-03-0615:081 reply

        Been a startup founder - work at Meta currently.

        AI is making everyone faster that I’ve seen. I’d say 30% of the tickets I’ve seen in the last month have been solved by just clicking the delegate to AI button

        • By bubblegumcrisis 2026-03-0615:335 reply

          How did you decide to work at Meta?

          I'll be honest, just the idea of working there makes me feel like vomiting. For me, they are bizarrely evil. They're not evil like, "we're going to destroy our competition through anti competitive practices," (which they do), but "let's destroy a whole generation of minds."

          And now with the glasses. I mean, jeeze. Can there be a stronger signal of not caring for others?

          It's as if Meta sees people as cattle. Though I think a lot of techies see humans as cattle, truthfully.

          What was your rationale?

          I guess this question is out-of-the-blue, and I don't mean for you to justify your existence, but I've never understood why people choose to work for Meta.

          • By qweiopqweiop 2026-03-0615:45

            I feel the same - would I like a meta paycheck, sure, but I couldn't look at myself in the mirror knowing what the company I'm giving my work to does to people's brains (not just the young, though that is the most reprehensible).

          • By bobbane 2026-03-0618:541 reply

            I told my son I would disown him if he worked for Facebook, for the reasons stated above.

            Then he took a contracting gig for Meta. His rationalization was that the project was an ill-specified prototype that would never see the light of day - if they wanted to throw money at him for stuff like that, he would accept it.

            That gig is finished, and he's now thoroughly disillusioned with working for big tech.

          • By m_ke 2026-03-0616:26

            Guess who is running product and other related functions at OpenAI and Anthropic now

          • By matthest 2026-03-0619:02

            From this angle, what's the difference between Meta and a junk food company?

            Both sell things that are bad for you, but that the consumer has complete control over whether or not to consume.

            And not all of what Meta is selling is bad. There's a lot of information exchanged on Facebook, Instagram, etc. that are good for society. Like health/nutrition advice, etc.

          • By rybosworld 2026-03-0616:25

            I've always attributed it to people being very good at convincing themselves they aren't one of the bad guys. A big paycheck makes it even easier to ignore to what you are a part of.

            Where livelihood is concerned, rational individuals with strong morals can do irrational, and immoral things (e.g., work at the Palantir's of the world).

            TLDR: incentives don't just shape perception, they form it

      • By dirkc 2026-03-0614:472 reply

        I have a theory that when you have 2 developers working in synergy, you're at something like 1.8x what 1 person can do. As you add more people you approach 2x until some point after which you start to decrease productive. I don't think that point is far beyond 5.

        • By rhogar 2026-03-0616:33

          This is very close to the thesis, or at least theme, of the essays in The Mythical Man-Month, Fred Brooks. Some elements are dated (1975), but many feel timeless.

          Brooks law “Adding manpower to a late software project makes it later” is just the surface of some of the metaphorical language that has most stuck with me: large systems and teams quickening entanglement in tar pits through their struggle against coordination scaling pains, conceptual integrity in design akin to preserving architectural unity of Reims cathedral, roles and limitations attempting to expand surgical teams, etc.

          Love a good metaphor, even when its foundation is overextended or out of date. Highly recommend.

        • By BurningFrog 2026-03-0617:152 reply

          My experience of pair programming is the opposite. In a pair I get maybe 4x as much done as when working alone.

          Mostly it's because when we hit a point where one person would get stuck, the other usually knows what to do, and we sail through almost anything with little friction.

          • By dirkc 2026-03-0619:24

            Maybe the multiplier is 4x and by the time you have a team of ten you're back down to 2x? My theory is a bit of a hyperbole and I don't know what the multipliers would be? But I know that many times you can move quick when you're small.

            And to your point, a single person can easily get stuck, I know that applies to me many times.

          • By sdf2df 2026-03-0618:52

            There's that but youre missing a lot of variables. E.g. if one of you had perfect sleep and the other didn't the individual with perfect sleep will perform better for longer.

            I don't get why people try to simplify - you're removing important details that determine performance and therefore output. This leads to false conclusions.

      • By spacecadet 2026-03-0613:39

        This. Hell even a company that is 100 people or more. Ive seen companies just grind to snails pace around 80-90 people and then still scale to 400-500 and then it's impossible to really do anything meaningful. I have tried to test for this in interviews over the years but ultimately I just end up disappointed. At this point I don't even look, just work in small independently organized groups or coops.

      • By wek 2026-03-091:58

        I'm excited about Agents helping many tiny teams succeed. There has been hype around the "who will be the first solo founder to a billion" but I am hoping for many small teams to succeed and I think this is the more interesting story.

        I agree its in the 2-7 person range.

        The challenge for those teams is distribution. They will crush at building, but I'm not sure how they can crack distribution. Some will, but maybe there is a way to help thousands of small teams distribute.

      • By baxtr 2026-03-0614:201 reply

        I love tiny teams. I hate big corp.

        Big corporations are full with people who love to entertain 20+ people in video calls. 1-2 people speak, the other nod their heads while browsing Amazon.

        I wouldn’t be sad if those jobs vanished.

        • By ffsm8 2026-03-0615:141 reply

          Well, you should be terrified of those jobs vanishing I think.

          All of these people will consequently be on the job market competing for your opportunities.

          Yes you may feel superior to their capabilities - and may even be justified in your opinion (I know nothing about you beyond this comment)... But it'll still significantly impact your professional future if this actually happens. It would massively impact wages at the very least

          Your viewpoint is incredibly short-sighted and not actually realizing the broad effect on the industry as a whole such a change would bring.

          • By baxtr 2026-03-0615:573 reply

            Maybe I’m naive but I’m not terrified about the future at all.

            Every efficiency wave made life better for humans. Why should this one be different?

            Assume many people lose their jobs. This in turn means companies will have higher margins. Higher margins attract more competition. More competition means lower margins since some will use the lower costs to offer lower prices.

            Lower prices increase quality of life for everyone.

            People who lost their job might be able to pick up doing something they actually enjoy…

            • By jlarocco 2026-03-0617:21

              > People who lost their job might be able to pick up doing something they actually enjoy…

              That's so out of touch.

              First, you're conveniently ignoring the possibility that people actually like the job they are about to lose.

              And believe it or not most people aren't toiling away at jobs they hate because it never occurred to them to do something they like more. They work jobs they dislike because it's the only choice they have because they have to pay their bills so they can survive and so that their dependents can have an acceptable life.

            • By heavyset_go 2026-03-0616:512 reply

              Throughout history, what were once middle class and artisan professions were increasingly automated and tons of people and their families ended up in poverty until they died.

              We just gloss over them and villify the ones who tried to do anything about it (the ones that weren't executed also died in poverty).

              • By rybosworld 2026-03-0618:051 reply

                Yeah this always get's completely glossed over in these conversations.

                People always say: "Things ended up working out in the end"

                Things only worked out in the sense that society carried on without all the people who lost their jobs.

                The U.S. has recent examples of large scale job destruction.

                Michigan: From 2000-2009. Massive job destruction. 330,000 auto workers in 2000. Down to 109,000 in 2009. Estimates are that 1/3-1/2 of all those affected never achieved equal/similar employment. That is, somewhere around ~70k-120k workers never earned as much as they previously did. Since this was msotly contained within one city (Detroit), it's pretty easy for the country to ignore it and go on with their lives.

                (Detroit was in decline since the 50's really. 2000-2009 is just a particularly bad snapshot.)

                Coal mining towns have experienced the same phenomenon but more gradually. The poverty left behind by the destruction of those jobs has never been addressed.

                With AI, we are heading into a situation where potentially a much larger amount of people will be affected. So maybe that changes the calculus on the government stepping in and fixing the problem. But I wouldn't count on it.

                Sources for Michigan numbers:

                https://lehd.ces.census.gov/doc/workshop/2010/LEDautopres031...

                https://research.upjohn.org/cgi/viewcontent.cgi?article=1205...

                • By Zarathruster 2026-03-0622:21

                  > Since this was mostly contained within one city (Detroit)

                  It's concentrated in Detroit but also distributed throughout the state, as you can observe in the census.gov slides.

                  The devastation is regional. It's been a wild experience, watching it all fall apart over the last 40+ years. The decay is immense and impossible to convey to someone from a rich state. Someone from the Eastern Bloc might get it, but I've never been able to communicate it to a Californian. Hop in a car and drive from town to town. Once-prosperous communities are boarded up and gradually reclaimed by nature. Department stores are converted into soup kitchens or marijuana dispensaries.

                  "Things will work themselves out" is not a law of nature, unless we broaden our definition of "things working out" to include outcomes like "everyone young enough flees, everyone else clutches their savings until they eventually die impoverished."

                  But with AI, even outcomes like that might be overly optimistic. Where will young people flee to? Where can they go, what trade can they learn, to be safe enough to eventually die in comfort?

                  When I look at Michigan I see both the past and the future, and I am planning accordingly.

              • By baxtr 2026-03-0618:011 reply

                You need to be careful with these things. Such exaggerated narratives are the reason people are afraid.

                during the Industrial Revolution many artisan and skilled trades lost their livelihoods.

                And yet, while many people did suffer serious short-term hardship and wage collapse, most did not simply remain in lifelong poverty, because over time industrialization created new types of employment and average wages eventually rose.

                You don’t want to go back to before the Industrial Revolution. Do you?

                • By izacus 2026-03-0622:551 reply

                  [flagged]

                  • By tomhow 2026-03-070:031 reply

                    > Because your ignorance is painful.

                    It's not acceptable to attack a fellow community member like this on HN. The guidelines make it clear we're aiming for better than this:

                    Be kind. Don't be snarky. Converse curiously; don't cross-examine. Edit out swipes.

                    Comments should get more thoughtful and substantive, not less, as a topic gets more divisive.

                    When disagreeing, please reply to the argument instead of calling names. "That is idiotic; 1 + 1 is 2, not 3" can be shortened to "1 + 1 is 2, not 3."

                    Please don't use Hacker News for political or ideological battle. It tramples curiosity.

                    https://news.ycombinator.com/newsguidelines.html

                    • By izacus 2026-03-078:231 reply

                      Being kind to people who blatantly lie about history and hide suffering of thousands is how we got into the current mess.

                      • By tomhow 2026-03-079:12

                        We don't need a defiant mini-sermon and it's very poor conduct to use the term "blatantly lie" for a fellow community member who is just expressing their understanding of a topic. It is never morally necessary to abuse people on this site. This is a community not a battleground.

                        If you have a different understanding of the topic, share it, so all can benefit. That's what people do when they are sincere about contributing positively here.

                        If instead you insist on continuing to use abusive terms towards others here, we'll have to ban the account.

            • By SkyeCA 2026-03-0616:45

              > People who lost their job might be able to pick up doing something they actually enjoy…

              It's more probable they lose everything before ending up with a worse job that pays less.

    • By wnmurphy 2026-03-0616:131 reply

      I agree with your categories. The majority of the usage for me is (1) and (3).

      (1) LLMs are basically Stack Overflow on steroids. No need to go look up examples or read the documentation in most cases, spit out a mostly working starting point.

      (3) Learning. Ramping up on an unfamiliar project by asking Antigravity questions is really useful.

      I do think it makes devs faster, in that it takes less time to do these two things. But you're running into the 80% of the job that does not involve writing code, especially at a larger company.

      In theory, this should allow a company to do more with fewer devs, but in reality it just means that these two activities become easier, and the 80% is still the bottleneck.

      • By StableAlkyne 2026-03-0618:53

        > LLMs are basically Stack Overflow on steroids

        That, and I've never had to beg an LLM for an answer, or waste 5 minutes of my life typing up a paragraph to pre-empt the XY Problem Problem. Also never had it close my question as a duplicate of an unrelated question.

        The accuracy tends to be somewhat lower than SO, but IMO this is a fair tradeoff to avoid having to potentially fight for an answer.

    • By svara 2026-03-0612:52

      Yes, throughput is determined by the bottleneck and above a certain organization size, the bottleneck often is coordination costs.

    • By random3 2026-03-0617:091 reply

      Interesting.

      Are you generating revenue or, otherwise, what productivity are you measuring?

      Without generating revenue (which to be clear is a very good proxy to measure impact) everyone can be indeed very prolific in their hobbies. But labor market is about making money for a living and unless you can directly impact your day-to-day needs from your work, it can't be called productive.

      • By vb7132 2026-03-075:16

        Very valid point. I will lay down the facts for you:

        At my previous employer, I was generating $2.5million per year (revenue per employee). I didn't ship a single line of code. All the time was spent trying to convince various stake holders.

        Now, I have already built a couple of apps that help me better manage my tech news (keeps me sane) plus I am writing a blog that generates $0. It's only been a month.

        If you measure the immediate dollar value, you are right. But in life, pay-offs are not always realized immediately. Just my opinion anyway.

    • By onlyrealcuzzo 2026-03-0617:16

      Can mostly second this.

      Working on a side project, and it's truly incredible how good AI has been for MOST of it.

      Also, bewildering how truly awful it was at some seemingly random things - like writing not terribly difficult Assembly that mostly exists already to do Go-style hot splitting (to even get it to understand what older versions of Go did).

      I suspect it'll still be 3 years before AI is as good at the FAANGs as it is outside, just due to the ungodly huge context and the amount of proprietary stuff it would need to learn to use effectively, plus getting all the access to it, etc.

      But, even when it does all that, that's maybe 33% of the job.

      I just don't see mass layoffs at the really big tech companies, unless it's more focused on just cutting and cutting than actually because people have been made redundant.

      Even at the management level, I'm not sure we're going to see managers managing teams of 30 instead of teams of 10.

      At the end of the day, a manager needs to know what you're doing and if you're any good at it, and there's only so many people a person can do that effectively with.

      Maybe low-level managers go away, and it's just TLMs, but someone still needs to do your 1-on-1s and babysit those that need babysat.

    • By Madmallard 2026-03-0612:415 reply

      2. Translating between two different coding languages (migration)

      I have a game written in XNA

      100% of the code is there, including all the physics that I hand-wrote.

      All the assets are there.

      I tried to get Gemini and Claude to do it numerous times, always with utter failure of epic proportions with anything that's actually detailed. 1 - my transition from the lobby screen into gameplay? 0% replicated on all attempts 2 - the actual physics in gameplay? 0% replicated none of it works 3 - the lobby screen itself? non-functional

      Okay so what did it even do? Well it put together sort of a boilerplate main menu and barebones options with weird looking text that isn't what I provided (given that I provided a font file), a lobby that I had to manually adjust numerous times before it could get into gameplay, and then nonfunctional gameplay that only handles directional movement and nothing else with sort of half-working fish traveling behavior.

      I've tried this a dozen times since 2023 with AI and as late as late last year.

      ALL of the source code is there every single thing that could be translated to be a functional game in another language is there. It NEVER once works or even comes remotely close.

      The entire codebase is about 20,000 lines, with maybe 3,000 of it being really important stuff.

      So yeah I don't really think AI is "really good" at anything complex. I haven't really been proven wrong in my 4 years of using it now.

      • By KronisLV 2026-03-0612:462 reply

        I crave to see people saying "Here's the repo btw: ..." and others trying out porting it over, just so we see all of the ways how AI fails (and how each model does) and maybe in the middle there a few ways to improve its odds. Until it eventually gets included in training data, a bit like how LLMs are oddly good at making SVGs of pelicans on bicycles nowadays.

        And then, maybe someone slightly crazy comes along and tries seeing how much they can do with regular codegen approaches, without any LLMs in the mix, but also not manual porting.

        • By jb_hn 2026-03-0615:01

          Agreed -- coding agents / LLMs are definitely imperfect, but it's always hard to contextualize "it failed at X" without knowing exactly what X was (or how the agent was instructed to perform X)

        • By Madmallard 2026-03-0613:35

          I'm sure someone who regularly programs games in the destination language I want who also has worked with XNA in the past as a game developer could port it in a week or something yeah

      • By KellyCriterion 2026-03-0614:10

        - Split it in different modules / tasks

        - Do not say: "just convert this"

        - On critical sections you do a method-per-method-translation

        - Dont forget: your 20.000 lines source at a whole will make any model to be distracted on longer tasks (and sessions, for sure)

        - Do dedicated projects within Claude per each sub-module

      • By PacificSpecific 2026-03-0613:531 reply

        This matches my experience. Unless it's been done to death online (crud etc) it falls on its face every time.

        • By Madmallard 2026-03-0614:02

          It's okay shills and those with stock in the AI copies brainwashed themselves inside out and will spam forever on this website that if you just introduce the right test conditions agents can do anything. Never mind engineering considerations if it passes the tests it's good homie! Just spent an extra few hundred or thousand a month on it! Especially on the company I have stock in! Give me money!

      • By vb7132 2026-03-075:22

        Yes, you are right: amongst the four points, migration is the most contentious one. You need to be fairly prudent about migration and depending on the project complexity, it may or may not work.

        But I do feel this is a solvable problem long term.

      • By sheeshkebab 2026-03-0613:211 reply

        In those situations you basically need to guide llm to do it properly. It rarely one shots complex problems like this, especially in non web dev, but could make it faster than doing it manually.

        • By Madmallard 2026-03-0613:342 reply

          Oh believe me I broke it down super finely, down to single files and even single functions in some places

          It still is completely and utterly hopeless

          • By sigseg1v 2026-03-0614:492 reply

            I've done this multiple times in various codebases, both medium sized personal ones (approx 50k lines for one project, and a smaller 20k line one earlier) and am currently in the process of doing a similar migration at work (~1.4 million lines, but we didn't migrate the whole thing, more like 300k of it).

            I found success with it pretty easily for those smaller projects. They were gamedev projects, and the process was basically to generate a source of truth AST and diff it vs a target language AST, and then do some more verifier steps of comparing log output, screenshot output, and getting it to write integration tests. I wrote up a bit of a blog on it. I'm not sure if this will be of any use to you, maybe your case is more difficult, but anyway here you go: https://sigsegv.land/blog/migrating-typescript-to-csharp-acc...

            For me it worked great, and I would (and am) using a similar method for more projects.

            • By Madmallard 2026-03-0618:121 reply

              "I also wanted to build a LOT of unit tests, integration tests, and static validation. From a bit of prior experience I found that this is where AI tooling really shines, and it can write tests with far more patience that I ever could. This lets it build up a large hoard of regression and correctness tests that help when I want to implement more things later and the codebase grows."

              The tests it writes in my experience are extremely terrible, even with verbose descriptions of what they should do. Every single test I've ever written with an LLM I've had to modify manually to adjust it or straight up redo it. This was as recent as a couple months ago for a C# MAUI project, doing playwright-style UI-based functionality testing.

              I'm not sure your AST idea would work for my scenario. I'd be wanting to convert XNA game-play code to PhaserJS. It wouldn't even be close to 95% similar. Several things done manually in XNA would just be automated away with PhaserJS built-ins.

              • By sigseg1v 2026-03-0622:42

                Ya I could see where framework patterns and stuff will need a lot of corrections in post after that type of migration. For mine it was the other direction and only the server portion (Express server written in typescript for a Phaser game, and porting to Kestrel on C#, which was able to use pretty much identical code and at the end after it was done I just switch and refactor ba few things to make it more idiomatic C#).

                For the tests, I'm not sure why we have such different results but essentially it took a codebase I had no tests in, and in the port it one shot a ton of tests that have already helped me in adding new features. My game server for it runs in kubernetes and has a "auto-distribute" system that matches players to servers and redistributes them if one server is taken offline. The integration tests it wrote for testing that auto-distribute system found a legit race condition that was there in both the old and new code (it migrated it accurately enough that it had the same bugs) and as part of implementing that test it fixed the bug.

                Of course I wouldn't use it if it wasn't a good tool but for me the difference between doing this port via this method versus doing it manually in prior massive projects was such an insane time save that I would have been crazy to do it any other way. I'm super happy with the new code and after also getting the test infra and stuff like that up it's honestly a huge upgrade from my original code that I thought I had so painstakingly crafted.

            • By mettamage 2026-03-0615:40

              super cool, don't have the time to read it right now but to think in terms of ASTs is pretty handy!

          • By sheeshkebab 2026-03-0716:05

            The only model that works well for complex things is Opus, and even then barely (but it does and you need to use api/token pricing if you want guarantee it’s the real thing).

    • By wvlia5 2026-03-0613:08

      This is a bot comment.

  • By throwaw12 2026-03-065:3135 reply

    People who are saying they're not seeing productivity boost, can you please share where is it failing?

    Because, I am terrified by the output I am getting while working on huge legacy codebases, it works. I described one of my workflow changes here: https://news.ycombinator.com/item?id=47271168 but in general compared to old way of working I am saving half of the steps consistently, whether its researching the codebase, or integrating new things, or even making fixes. I have stopped writing code, occasionally I jump into the changes proposed by LLM and make manual edits if it is feasible, otherwise I revert changes and ask it to generate again but based on my learnings from the past rejected output

    I am terrified about what's coming

    • By yoyohello13 2026-03-066:0616 reply

      The companies laying off people have no vision. My company is a successful not for profit and we are hiring like crazy. It’s not a software company, but we have always effectively unlimited work. Why would anyone downsize because work is getting done faster? Just do more work, get more done, get better than the competition, get better at delivering your vision. We put profits back in the community and actually make life better for people. What a crazy fucking concept right?

      • By tkgally 2026-03-067:511 reply

        I suspect it depends partly on how locked each individual is into a particular type of work, both skill-wise and temperamentally.

        To give an example from a field where LLMs started causing employment worries earlier than software development: translation. Some translators made their living doing the equivalent of routine, repetitive coding tasks: translating patents, manuals, text strings for localized software, etc. Some of that work was already threatened by pre-LLM machine translation, despite its poor quality; context-aware LLMs have pretty much taken over the rest. Translators who were specialized in that type of work and too old or inflexible to move into other areas were hurt badly.

        The potential demand for translation between languages has always been immense, and until the past few years only a tiny portion of that demand was being met. Now that translation is practically free, much more of that demand is being met, though not always well. Few people using an app or browser extension to translate between languages have much sense of what makes a good translation or of how translation can go bad. Professional translators who are able to apply their higher-level knowledge and language skills to facilitate intercultural communication in various ways can still make good money. But it requires a mindset change that can be difficult.

        • By adelie 2026-03-069:112 reply

          I'm not in translation, but a number of close friends are in the industry. Two trends I've noticed in the industry, which I think we're seeing mirrored in tech:

          1. No one cares about quality. Even in fields you'd expect to require the 'human touch' (e.g. novel translation), publishers are replacing translators with AI. It doesn't matter if you have higher-level knowledge or skills if the company gains more from cutting your contract than it loses in sales.

          2. Translation jobs have been replaced with jobs proofreading machine translations, which pays peanuts (since AI is 'doing most of the work') but in fact takes almost as much effort as translating from scratch (since AI is often wrong in very subtle ways). The comparison to PR reviews makes itself.

          • By thbb123 2026-03-0611:352 reply

            It is not entirely true that no one cares about quality. I'd like to stay optimistic and believe that those who are demanding on the quality of their production will acquire sufficient market differentiation to prevail.

            After all, this has been Apple strategy since the 80's, and, even though there were some up's and down's, overall it's a success.

            • By palmotea 2026-03-0614:21

              > It is not entirely true that no one cares about quality. I'd like to stay optimistic and believe that those who are demanding on the quality of their production will acquire sufficient market differentiation to prevail.

              Maybe, but it probably requires a very strong and opinionated leader to pull off. The conventional wisdom in American business leadership seems to be to pursue the lowest level of quality you can get away with, and focus on cutting costs. And you'll have to fight that every second.

              I don't think that's true at the individual-contributor level (pursing quality is very motivating), but they people who move up are the ones who sound "smart" by aping conventional wisdom.

              > After all, this has been Apple strategy since the 80's, and, even though there were some up's and down's, overall it's a success.

              I might give you that "since the late 90s," but there have been significant periods where that wasn't true (e.g. the early mid-90s Mac OS was buggy and had poor foundations).

            • By kavalg 2026-03-0614:04

              someone still will, but quality will become really expensive

          • By izacus 2026-03-0622:58

            In other words, AI was used to massively depress wages and lower quality of life of employees while outputting worse results. Which is what is now happening in software.

      • By afro88 2026-03-066:162 reply

        This is exactly right IMO. I have never worked for a company where the bottleneck was "we've run out of things to do". That said, plenty of companies run out of actual software engineering work when their product isn't competitive. But it usually isn't competitive because they haven't been able to move fast enough

        • By weatherlite 2026-03-0612:261 reply

          I think it depends on:

          A) how old the product is: Twitter during its first 5 years probaby had more work to do compared to Twitter after 15 years. I suspect that is why they were able to get rid of so many developers.

          B) The industry: many b2c / ecommerce businesses are straightforward and don't have an endless need for new features. This is different than more deep tech companies

          • By thewebguyd 2026-03-0615:48

            There’s a third one, and it’s non-tech companies or companies for whom software is not a core product. They only make in-house tooling, ERP extensions, etc. Similar to your Twitter example, once the ERP or whatever is “done” there’s not much more work to do outside of updating for tax & legal changes, or if the business launches new products, opens a new location, etc.

            I’ve built several of such tools where I work. We don’t even have a dev team, it’s just IT Ops, and all of what I’ve built is effectively “done” software unless the business changes.

            I suspect there’s a lot of that out there in the world.

        • By sdf2df 2026-03-0623:32

          Not moving fast enough.. sure. But to what direction? The direction and clarity of it is the hardest part.

      • By ehnto 2026-03-066:343 reply

        That was my insight also. As a manager, you already have the headcount approved, and your people just allegedly got some significant percentage more productive. The first thought shouldn't be, great let's cut costs, it should be great now we finally have the bandwidth to deliver faster.

        On a macro level, if you were in a rising economic tide, you would still be hiring, and turning those productivity gains into more business.

        I wonder what the parallels are to past automations. When part producing companies moved from manual mills to CNC mills, did they fire a bunch of people or did they make more parts?

        • By superfrank 2026-03-068:171 reply

          I'm an EM as well and I've been telling my teams for a while now that I think they really only need to start worrying once our backlog starts going down instead of up. Generally, I still agree with that (and your) sentiment when you look at the long term, but in the short term, I think all of the following arguments can be made in favor of layoffs:

          - AI tools are expensive so until the increased productivity translates to increased revenue we need to make room in the budget

          - We expect the bottlenecks in our org to move from writing code to something else (PM or design or something) so we're cutting SWEs in anticipation of needing to move that budget elsewhere.

          - We anticipate the skillsets needed by developers in the AI world to be fundamentally different from what they are now that it's cheaper to just lay people off, run as lean as possible, and rehire people with the skills we want in a year or two than it is to try and retrain.

          I don't necessarily agree with those arguments (especially the last one), but I think they're somewhat valid arguments

          • By throwaw12 2026-03-068:481 reply

            I see similar arguments and I don't agree as well, here is why:

            > rehire people with the skills we want in a year or two than it is to try and retrain.

            before that future comes your company might become obsolete already, because you have lost your market share to new entrants

            > We expect the bottlenecks in our org to move from writing code to something else

            I would love to tell them, hey lets leverage current momentum and build, when those times come, we offer existing people with accumulated knowledge to retrain to a new type of work, if they think they're not good fit, they can leave, if they're willing, give them a chance, invest in people, make them feel safe and earn trust and loyalty from them

            > AI tools are expensive so until the increased productivity translates to increased revenue we need to make room in the budget

            1. Its not that expensive: 150$/seat/month -> 5 lunches? or maybe squeeze it from Sales personnel traveling with Business class?

            2. By the time increased productivity is realized by others, company who resisted could be so far behind, that they won't be able to afford hiring engineers with those skillsets, if they think 150$ is expensive now, I am sure they will say "What??? 350k$ for this engineer?, no way, I will instead hire contractors"

            • By ehnto 2026-03-076:03

              Your last point I think is the crux for now, if you can get so much more value out of talent then the market will eventually price that into wages. I think the idea some have instead is that I can use cheap, unskilled people to get the same value as before, which is probably true for some aspects of the industry. My experience of that so far for boutique software is AI propels unskilled employees at light speed into dead ends, similar to how a junior would spend two weeks following ideas from stack overflow and being unable to execute on it.

              However AI definitely is capable of lower end software tasks and really well trodden ground, especially when managed by a developer, so perhaps what we will see is a bigger gap in pay and talent not too different from the off-shore vs on-shore market comparisons.

              The key for me though for me is that, if AI makes your employees 20% more valuable, that will either get priced into their wage or captured by the business, but it still doesn't replace the need for good talent (software engineer, agent handler, whatever it will get called).

        • By NathanielK 2026-03-0611:44

          CNC machines drove down operator wages. Its similar to the translator example where the machine code is written by someone else, but the person running the machine still needs to understand. Simple pushing the go button is dangerous, being able to adapt is critical.

          Jobs where a machinist is in charge of large chunks of the process are rarer. Large shop will have one person setting up many machines to maximize throughput.

        • By anthonypasq 2026-03-0616:411 reply

          business success does not scale at the speed of increased profits from layoffs.

          • By ehnto 2026-03-074:07

            It's true but it's also not real growth. It will look good the first time you do it. It also relies on there being no negative growth induced by AI not meeting the same quality of output that thousands of workers were once doing.

            If it is truly because of AI, then it's still a losing strategy long term in my opinion.

      • By throw3847r7 2026-03-068:141 reply

        You need certain company culture, to be able to scale up, and to capture this value. Most companies can not just add new developers.

        AI needs documentation, automation, integration tests... It works very well for remote first company, but not for in-face informal grinding approach.

        Just year ago, client told me to delete integration tests, because "they ran too long"!

        • By joe_mamba 2026-03-069:291 reply

          >Just year ago, client told me to delete integration tests, because "they ran too long"!

          Why are you surprised customers don't like spending money on the items that don't add business value. Add to that QA, documentation, security audits, etc.

          They want to ship stuff that brings in customers and revenue day one, everything else is a cost.

          • By SideburnsOfDoom 2026-03-0613:151 reply

            > integration tests, QA etc ... the items that don't add business value

            They absolutely do add value / prevent loss, but you need some understanding in order to see that. Not seeing it is a marker of not understanding.

            • By joe_mamba 2026-03-0613:49

              >They absolutely do add value

              Not to the non-technical bean counters. When they allocate money they want to see you prove how that extra money translates to an immediate ROI, and it's difficult to prove that in an Excel sheet exactly what the ROI will be without making stuff up on vibes and feels.

              Like at one German company i was at ~15 years ago, all the devs wanted a second 19" monitor on our workstations for increased productivity, and the bean counters wouldn't approve that because they wanted proof of how that expense across hundreds of people will increase our productivity and by how much %, to see if that would offset the cost.

              This is how these people think. If you don't bring hard numbers on how much their "line will go up", they won't give you money.

              I know this is difficult to understand from the PoV of SV Americans where gazillions of dollars just fall from the sky at their tech companies.

      • By anthonypasq 2026-03-0616:40

        most businesses dont actually have an infinite amount of work that has extremely high ROI. every new project at google for example has to justify the engineering spend of developing a product that has comparable margin to the ad business. Why spend 10 million a year of engineering resources on a new product that might 1. completely fail or 2. be a decent product with 20% margins when they could do nothing and keep raking in 90% margins from the ads business.

      • By pllbnk 2026-03-0821:34

        I have a hope that many of today's engineers working at various companies will start realizing that instead of employers receiving tools to get them laid off, it is they, the engineers, who received the tools to compete with said employers and outcompete them.

        If, and it's a big if, AI models really boost productivity by an order of magnitude (I personally, while being skeptical a year or two ago, am leaning towards this idea) then engineers have a chance to realize their ideas, improve current system design patterns and build successful companies, which will inevitably (hopefully) require hiring personnel to keep competing, bringing entire software engineering market to a newly balanced state.

      • By RA_Fisher 2026-03-069:521 reply

        Does that extra work bring in more revenue? I think that’s the key question.

        • By raphaelj 2026-03-0610:131 reply

          Companies that do not reduce their workforce might outcompete you.

          It might not be about bringing more revenues but retaining market share.

          • By Esophagus4 2026-03-0613:08

            If your barrier to being competitive is a slow, bureaucratic org, restructuring and laying off might actually help long term.

      • By crocowhile 2026-03-067:56

        Because hiring less while getting more done increases margins. Your company is not for profit so doesnt care about margins. Others do.

      • By apercu 2026-03-0611:45

        I think a lot of companies have ineffective ways to measure productivity, poor management (e.g., people who were IC's then promoted to management but have no management training or experience), incentives aren't necessarily aligned between orgs and staff, so people end up with a perverse "more headcount" means I'm better than Sandy over there. Leadership and vision have been rare in my professional life (though the corporate-owned media celebrates mediocrity in leadership all the time with puff pieces).

        Once you get to a certain size company, this means a lot of bloat. Heck, I've seen small(ish) companies that had as many managers and administrators as ICs.

        But You're not wrong, I'm just pointing out how an org that has 4k people can lay off a few hundred with modest impact of the financials (though extensive impact on morale).

      • By MattGaiser 2026-03-067:46

        You would need to expand your capacity to find and define the work. I imagine that would be a major challenge.

      • By threatofrain 2026-03-066:121 reply

        These are words without weights. At some point the put money into software option will max out. Perhaps what we should all be doing is hiring more lawyers, there's always more legal work to be done. When you don't have weights then you can reason like this.

        • By yoyohello13 2026-03-066:141 reply

          I don’t know what kind of software your used to but software is pretty much universally dog shit these days. I could probably count on one hand the number of programs that I actually like using. There is an astronomical room for improvement. I don’t think we are hitting diminishing returns any time soon.

      • By arwhatever 2026-03-067:05

        I’ve been screaming this too https://news.ycombinator.com/item?id=47212237

        It’s refreshing to see the same sentiment from so many other people independently here.

      • By zipy124 2026-03-069:48

        The problem becomes if you are a service like Youtube, where you already have capture almost the entire customer base.

      • By svara 2026-03-069:43

        Yes, it's the lump of labor fallacy.

        Doesn't exclude the possibility of short term distribution, though.

      • By throwaw12 2026-03-066:122 reply

        > Just do more work, get more done

        That's one of the reasons why I am terrified, because it can lead to burn out, and I personally don't like to babysit bunch of agents, because the output doesn't feel "mine", when its not "mine" I don't feel ownership.

        And I am deliberately hitting the brake from time to time not to increase expectations, because I feel like driving someone else's car while not understanding fully how they tuned their car (even though I did those tunings by prompting)

        • By QuercusMax 2026-03-066:431 reply

          It feels very much like leading a team of junior engineers or even interns who are very fast but have no idea about why we're doing anything. You have to understand the problems you're trying to solve and describe the solutions in a way they can be implemented.

          It's not going to be written exactly like you would do it, but that's ok - because you care about the results of the solution and not its precise implementation. At some point you have to make an engineering decision whether to write it yourself for critical bits or allow the agent/junior to get a good enough result.

          You're reviewing the code and hand editing anyway, right? You understand the specs even if your agent/junior doesn't, so you can take credit even if you didn't physically write the code. It's the same thing.

          • By throwaw12 2026-03-066:591 reply

            > It feels very much like leading a team of junior engineers or even interns who are very fast but have no idea about why we're doing anything

            Yes, yes!

            And this is problem for me, because of the pace, my brain muscles are not developing enough compared to when I was doing those things myself.

            before, I was changing my mind while implementing the code, because I see more things while typing, and digging deeper, but now, because juniors are doing things they don't offer me a refactoring or improvements while typing the code quickly, because they obey my command instead of having "aha" moment to suggest better ways

            • By layer8 2026-03-0611:53

              There’s some hope that the industry will realize that managing clueless LLMs at high pace isn’t sustainable and leads to worse results, and some middle ground has to be found. Or we will reach AGI, so AI won’t be clueless anymore and really take your engineering job.

        • By ako 2026-03-068:251 reply

          I'm currently a product manager (was a software engineer and technical architect before), so i already lost the feeling of ownership of code. But just like when you're doing product management with a team of software engineers, testers, and UXers, with AI you can still feel ownership of the feature or capability you're shipping. So from my perspective, nothing changes regarding ownership.

          • By discreteevent 2026-03-069:061 reply

            > So from my perspective, nothing changes regarding ownership.

            The engineer who worked with you took ownership of the code! Have you forgotten this?

            • By ako 2026-03-069:38

              No, that’s why I wrote “from my perspective”. I started long ago writing 6502 and 68000 assembly, later c and even later Java. Every step you lose ownership of the underlying layer. This is just another step. “But it’s non deterministic!”, yes so are developers. We need QA regardless who or what write the lines of code.

      • By sdf2df 2026-03-0623:31

        Stop talking sense bro, you'll get downvoted.

        If you look at my post history I'm essentially saying the same stuff lol.

    • By rich_sasha 2026-03-0612:402 reply

      I find LLMs are good at essentially boilerplate code. It's clear what to do and it needs to be typed in. Or areas where I really have no idea where to start, because I'm not familiar with the codebase.

      I find anything else, I spend more time coaxing them into doing 85% of what I need that I'm better off doing it myself.

      So they're not useless but there's only so many times in a week that I need a function to pretty-print a table in some fashion. And the code they write on anything more complex than a snippet is usually written poorly enough that it's a write-once-never-touch-again situation. If the code needs to be solid, maintainable, testable, correct (and these are kind of minimal requirements in my book) then LLMs make little impact on my productivity.

      They're still an improvement on Google and Stack exchange, but again - only gets you so far.

      YMMV

      • By vividfrier 2026-03-0613:212 reply

        > I find anything else, I spend more time coaxing them into doing 85% of what I need that I'm better off doing it myself.

        You must be working in a very niche field with very niche functionality if that's the case? I work at a company just outside of FAANG and I work in compliance. Not a terribly complex domain but very complicated scale and data integrity requirements.

        I haven't written a single line of code manually in 2 weeks. Opus 4.6 just... works. Even if I don't give it all the context it just seems to figure things out. Occasionally it'll make an architectural error because it doesn't quite understand how the microservices interact. But these are non-trivial errors (i.e. humans could have made them as well) and when we identify such an error, we update the team-shared CLAUDE.md to make sure future agents don't repeat the error.

        • By rich_sasha 2026-03-0613:564 reply

          I often wonder what I am missing. Recently I wanted to wrap a low level vendor API with a callback API (make a request struct and request id, submit, provide a callback fn, which gets called with request IDs and messages received from vendor) to async Python (await make_request(...)). Kinda straightforward - lots of careful code of registering and unregistering callbacks, some careful thread synchronisation (callbacks get called in another thread), thinking about sane exception handling in async code. Fiddly but not rocket science.

          What I got sort of works, as in tests pass - this with Opus 4.5. It is usable, though it doesn't exist cleanly on errors despite working to death with Claude about this. On exception it exits dirtily and crashes, which is good enough for now. I had some fancy ideas about logging messages from the vendor to be able to replay them, to be able to then reproduce errors. Opus made a real hash of it, lots of "fuck it comment out the assert so the test passes". This part is unusable and worse, pollutes the working part of the project. It made a valiant effort at mocking the vendor API for testing but really badly, instead of writing 30 lines of general code, it wrote 200 lines of inconsistent special cases that don't even work altogether. Asked to fix it it just shuffles around the special cases and gets stuck.

          It's written messily enough that I wouldn't touch this even to remove the dead code paths. I could block a few days for it to fix but frankly in that time I can redo it all and better. So while it works I'm not gonna touch it.

          I did everything LLM proponents say. I discussed requirements. Agent had access to the API docs and vendor samples. I said think hard many times. Based on this we wrote a detailed spec, then detailed inplementation plan. I hand checked a lot of the high level definitions. And yet here I am. By the time Opus went away and started coding, we had the user facing API hammered out, key implementation details (callback -> queue -> async task in source thread routing messages etc), constraints (clean handling of exceptions, threadsafe etc). Tests it has to write. Any minor detail we didn't discuss to death was coded up like a bored junior.

          And this also wasn't my first attempt, this was attempt #3. First attempt was like, here's the docs and samples, make me a Python async API. That was a disaster. Second was more like, let's discuss, make a spec, then off you go. No good. Even just taking the last attempt time, I would have spent less time doing this by hand myself from scratch.

          • By Izkata 2026-03-0615:581 reply

            Based on what I've seen and heard, you have the happy path working and that's what the pro-AI people are describing with huge speedups. Figuring out and fixing the edge cases and failure modes is getting pushed into the review stage or even users, so it doesn't count towards the development time. It can even count as more speed if it generates more cases that get handled quickly.

            • By rich_sasha 2026-03-0616:49

              I'm not sure I agree with this approach, or at least it doesn't work in my area. It's like self driving cars. Having 90% reliability is almost as good as 0%. I have to be confident the thing is gonna work, correctly, or at worst fail predictably.

              I can see that there's a lot of applications where things can just randomly fail and you retry / restart, that helps with crashes.

              But the AI can't make it not crash, what's to say it does the right thing when it succeeds? Again, depends on the relative cost of errors etc.

          • By Bewelge 2026-03-0616:131 reply

            Just a guess, but to me it sounds like you're trying to do too much at once. When trying something like this:

            > lots of careful code of registering and unregistering callbacks, some careful thread synchronisation (callbacks get called in another thread), thinking about sane exception handling in async code. Fiddly but not rocket science.

            I'd expect CC to fail this when just given requirements. The way I use it is to explicitly tell it things like: "Make sure to do Y when callback X gets fired" and not "you have to be careful about thread synchronisation". "Do X, so that Exceptions are always thrown when Y happens" instead of "Make sure to implement sane Exception handling". I think you have to get a feeling for how explicit you have to get because it definitely can figure out some complexity by itself.

            But honestly it's also requires a different way of thinking and working. It reminds me of my dad reminiscing that the skill of dictating isn't used at all anymore nowadays. Since computers, typing, or more specifically correcting what has been typed has become cheap. And the skill of being able to formulate a sentence "on the first try" is less valuable. I see some (inverse) parallel to working with AI vs writing the code yourself. When coding yourself you don't have to explicitly formulate everything you are doing. Even if you are writing code with great documentation, there's no way that it could contain all of the tacit knowledge you as the author have. At least that's how I feel working with it. I just got really started with Claude Code 2 months ago and for a greenfield project I am amazed how much I could get done. For existing, sometimes messy side projects it works a lot worse. But that's also because it's more difficult to describe explicitly what you want.

            • By rich_sasha 2026-03-0617:002 reply

              > The way I use it is to explicitly tell it things like: "Make sure to do Y when callback X gets fired" and not "you have to be careful about thread synchronisation". "Do X, so that Exceptions are always thrown when Y happens" instead of "Make sure to implement sane Exception handling".

              At this point I'm basically programming in English, no? Trying to squeeze exact instructions into an inherently ambiguous representation. I might as well write code at this point, if this is the level of detail required. For this to work, I need to be able to say "make this thread-safe", maybe "by using a queue". Not explaining which synchronisation primitive to use in every last piece of the code.

              This is my point actually. If I describe the task to accuracy level X, it still doesn't seem to work. To make it work, perhaps I need to describe it to level Y>X, but that for now takes me more time than to do it myself.

              There's lots of variables here, how fast I am at writing code or planning structure, how close to spec the things needs to be, etc. My first "vibe code" was a personal productivity app in Claude Code, in Flutter (task timing). I have 0 idea about Dash or Flutter or any web stuff, and yet it made a complete app that did some stuff, worked on my phone, with a nice GUI, all from just a spec. From scratch, it would take me weeks.

              ...though in the end, even after 3 attempts, the final thing still didn't actually work well enough to be useful. The timer would sometimes get stuck or crash back down to 0, and froze when the app was minimised.

              • By krackers 2026-03-074:45

                I'm reminded of the notion of "komolgorov complexity" here. There might be some tasks for which a short natural language description is sufficient, and others for which a sufficiently formal description is needed to the point that it's easier to actually write the code than describe it in English.

              • By Bewelge 2026-03-0618:35

                > At this point I'm basically programming in English, no?

                Yea, except they can handle some degree of complexity. Its usefulness obviously really depends on that degree. And I'm sure there are still a lot of domains and types of software where that tradeoff between doing it yourself or spelling it out isn't worth it.

          • By 8note 2026-03-0921:30

            Consistently over time, ive gotten really bad results trying to make anything with python using claude, so that might be it.

            i think ive done some python on each major model release, not necessarily woth claude code, but it feels like if it touches python, it gets stupid, quickly.

            and not just for the python, it stays stupid when it wants to say, write html. for a tiny data labeling tool, it turned off user select, for instance

          • By stiiv 2026-03-0614:14

            > On exception it exits dirtily and crashes, which is good enough for now

            Silent failures and unexplained crashes are high on my list of things to avoid, but many teams just take them for granted in spite of the practical impact.

            I think that a lot of orgs have a culture of "ship it and move on," accompanied by expectations like: QA will catch it, high turnover/lower-skill programmers commit stuff like this all the time anyway, or production code is expected to have some rough edges. I've been on teams like that, mostly in bigger orgs with high turnover and/or low engineering standards.

        • By mbbutler 2026-03-0616:07

          Two use-cases recently where Claude sucked for me:

          1. Performance-critical code to featurize byte slices for use in a ML model. Claude kept trying to take multiple passes over the slice when the featurization can obviously be done in one. After I finally got it to do the featurization in one pass it was double-counting some bytes but not others (double counting all of them would have been fine since the feature vector gets normalized). Overall it was just very frustrating because this should have been straight-forward and instead it was dogshit.

          2. Performance-critical code that iterates over lines of text and possibly applies transformations, similar to sed. Claude kept trying to allocate new Strings inside of the hot-loop for lines that were not transformed. When I told it to use Cow<'a, str> instead so that the untransformed lines, which make up the majority of processed lines, would not need a new allocation, Claude completely fucked up the named lifetimes. Importantly, my CLAUDE.md already tells Claude to use copy-on-write types to reduce allocations whenever possible. The agent just ignored it, which is _the_ issue with LLMs: they're non-deterministic and any guidance you provide is ultimately just a suggestion.

      • By joenot443 2026-03-0613:17

        > I spend more time coaxing them into doing 85% of what I need that I'm better off doing it myself

        What was the last thing you built in which you felt this was the case?

    • By mirsadm 2026-03-0611:026 reply

      I have an app which is fairly popular. This release cycle I used Claude Code and codex to implement all the changes / features. It definitely let me move much quicker than before.

      However now that it's in the beta stage the amount of issues and bugs is insane. I reviewed a lot of the code that went in as well. I suspect the bug fixing stage is going to take longer than the initial implementation. There are so many issues and my mental model of the codebase has severely degraded.

      It was an interesting experiment but I don't think I would do it again this way.

      • By aurareturn 2026-03-0613:13

        The last 10% takes up 90% of the time. Usually, the time is spent finding issues you didn't even know about. This was true before LLMs.

      • By johsole 2026-03-0618:29

        I make mistakes when writing code, but I know what types of mistakes I make. With AI it's like a coworker who makes mistakes, sometimes they're obvious to me and sometimes they're not, because I make different mistakes.

      • By truetraveller 2026-03-0611:03

        Thanks for the insight!

      • By sdf2df 2026-03-0623:43

        "There are so many issues and my mental model of the codebase has severely degraded."

        Not only that, the less coding you do in general? Guess what, fixing issues that in the past wouldve been a doddle (muscle memory) become less harder due to atrophy.

        Swear most people dont think straight and cant see the obvious.

      • By sdf2df 2026-03-0623:36

        Congrats. Now post this more often so the bozo's who downvote posts that push-against pro-LLM stuff f-off.

        I came to the same conclusion when producing a video with Grok. Did the job but utterly painful and it was definitely very costly - I used 50 free-trial accounts and maxed them out each day for a month.

        Im pretty sure these conclusions hold across all models and therefore the technology by extension.

      • By maplethorpe 2026-03-0611:141 reply

        Rather than trying to fix the bugs yourself, have you tried asking Claude to fix them for you?

        • By mirsadm 2026-03-0611:19

          I have already been doing this. I could keep doing it but I'm not going to. I want to be able to understand my own code because that is what let's me make sound higher level decisions.

    • By jpollock 2026-03-068:164 reply

      The last time I tried AI, I tested it with a stopwatch.

      The group used feature flags...

          if (a) {
             // new code
          } else {
             // old code
          }
      
          void testOff() {
             disableFlag(a);
             // test it still works
          }
          
          void testOn() {
              enableFlag(a);
              // test it still works
          }
      
      However, as with any cleanup, it doesn't happen. We have thousands of these things lying around taking up space. I thought "I can give this to the AI, it won't get bored or complain."

      I can do one flag in ~3minutes. Code edit, pr prepped and sent.

      The AI can do one in 10mins, but I couldn't look away. It kept trying to use find/grep to search through a huge repo to find symbols (instead of the MCP service).

      Then it ignored instructions and didn't clean up one or the other test, left unused fields or parameters and generally made a mess.

      Finally, I needed to review and fix the results, taking another 3-5 minutes, with no guarantee that it compiled.

      At that point, a task that takes me 3 minutes has taken me 15.

      Sure, it made code changes, and felt "cool", but it cost the company 5x the cost of not using the AI (before considering the token cost).

      Even worse, the CI/CD system couldn't keep up the my individual velocity of cleaning these up, using an automated tool? Yeah, not going to be pleasant.

      However, I need to try again, everyone's saying there was a step change in December.

      • By laserlight 2026-03-069:422 reply

        I did my own experiment with Claude Code vs Cursor tab completion. The task was to convert an Excel file to a structured format. Nothing fancy at all.

        Claude Code took 4 hours, with multiple prompts. At the end, it started to break the previous fixes in favor of new features. The code was spaghetti. There was no way I could fix it myself or steer Claude Code into fixing it the right way. Either it was a dead-end or a dice roll with every prompt.

        Then I implemented my own version with Cursor tab completion. It took the same amount of time, 4 hours. The code had a clear object-oriented architecture, with a structure for evolution. Adding a new feature didn't require any prompts at all.

        As a result, Claude Code was worse in terms of productivity: the same amount of time, worse quality output, no possibility of (or at best very high cost of) code evolution.

        • By thesamethrowawa 2026-03-069:491 reply

          Are you able to share your prompts to Claude Code? I assume not, they are probably not saved - but this genuinely surprised me, it seems like exactly the type of task an LLM would excel at (no pun intended!). What model were you using OOI?

          • By laserlight 2026-03-0610:03

            > this genuinely surprised me

            Me too. After listening to all the claims about Claude Code's productivity benefits, I was surprised to get the result I got.

            I'm not able to share details of my work. I was using Claude Opus 4.5, if I recall correctly.

        • By shinycode 2026-03-069:51

          The exact same prompt ? Everything depends on the prompt and it’s different tools. These days the quality and what’s build around the prompt matters as much as the code. We can’t feed generic query.

      • By sensanaty 2026-03-0612:26

        Similar happened to me just now. Claude whatever-is-the-latest-and-greatest, in Claude Code. I also tried out Windsurf's Arena Mode, with the same failure. To intercept the inevitable "holding it wrong" comments, we have all the AGENTS.md and RULES.md files and all the other snake oil you're told to include in the project. It has full context of the code, and even the ticket. It has very clear instructions on what to do (the kind of instructions I would trust an unpaid intern with, yet alone a tool marketed as the next coming of Cyber Jesus that we're paying for), in a chat with minimal context used up already. I manually review every command it runs, because I don't trust it running shell scripts unsupervised.

        I wanted it to finish up some tests that I had already prefilled, basically all the AI had to do was convert my comments into the final assertions. A few minutes later of looping, I see it finishes and all tests are green.

        A third of the tests were still unfilled, I guess left as an exercise for the reader. Another third was modified beyond what I told it to do, including hardcoding some things which made the test quite literally useless and the last third was fine, but because of all the miscellaneous changes it made I had to double check those anyways. This is about the bare minimum where I would expect these things to do good work, a simple take comment -> spit out the `assert()` block.

        I ended up wasting more time arguing with it than if I had just done the menial task of filling out the tests myself. It sure did generate a shit ton of code though, and ran in an impressive looking loop for 5-10 minutes! And sure, the majority of the test cases were either not implemented or hardcoded so that they wouldn't actually catch a breakage, but it was all green!!

        That's ultimately where this hype is leading us. It's a genuinely useful tool in some circumstances, but we've collectively lost the plot because untold billions have poured into these systems and we now have clueless managers and executives seeing "tests green -> code good" and making decisions based on that.

      • By embedding-shape 2026-03-0610:50

        What model, what harness and about how long was your prompt to fire off this piece of work? All three matters a lot, but importantly missing from your experience.

      • By sdf2df 2026-03-0623:48

        [dead]

    • By wasmainiac 2026-03-066:292 reply

      Because its failure rate is too high. Beyond boilerplate code and CRUD apps, if I let AI run freely on the projects I maintain, I spend more time fixing its changes than if I just did it myself. It hallucinates functionally, it designs itself into corners, it does not follow my instructions, it writes too much code for simple features.

      It’s fine at replacing what stack overflow did nearly a decade ago, but that isn’t really an improvement from my baseline.

      • By leptons 2026-03-067:561 reply

        That's my experience too. It's okay at a few things that save me some typing, but it isn't really going to do the hard work for me. I also still need to spend significant amounts of time figuring out what it did wrong and correcting it. And that's frustrating. I don't make those mistakes, and I really dislike being led down bad paths. If "code smells" are bad, then "AI" is a rotting corpse.

        • By thewebguyd 2026-03-0616:131 reply

          > If "code smells" are bad, then "AI" is a rotting corpse.

          This is what's so frustrating about the hype bros for me. In most cases, everything AI spits out are code smells.

          We're all just supposed to toss out every engineering principle we've learned all so the owner class can hire less developers and suppress wages?

          I'm sure it's working great for everyone working on SaaS CRUD or web apps, but it's still not anywhere close to solving problems outside that sphere. Native? It's very hit and miss. It has very little design sense (because, why would it? It's a language model) so it chokes on SwiftUI, it also can't stop using deprecated stuff.

          And that's not even that specialized. It still hallucinates cmdlets if you try to do anything with PowerShell, and has near zero knowledge about the industry I work in, a historically not tech-forward industry where things are still shared in handcrafted PDF reports emailed out to subscribers.

          I'm going to leave this field entirely if the answer just becomes "just make everything in React/React Native because it's what the AI does best."

          • By fragmede 2026-03-089:47

            That's interesting because I've been having pretty good luck with AI on SwiftUI. It didn't use to be but at some point they got good at that.

      • By qudat 2026-03-0612:361 reply

        It’s not that it just makes mistakes but it also implements things in ways I don’t like or are not relevant to the business requirements or scope of the feature / project.

        I end up replacing any saved time with QA and code review and I really don’t see how that’s going to change.

        In my mind I see Claude as a better search engine that understands code well enough to find answers and gain understanding faster. That’s about it.

        • By anthonypasq 2026-03-0616:441 reply

          can you imagine two years in the future and still believe this will be true? You are just dragging your feet. You will give in sooner or later, and i would suggest sooner.

          • By qudat 2026-03-072:57

            Nah I’m using it extensively, I know the limits. I do not think scaling is going to magically fix the fundamental limits of attention LLMs

    • By apsurd 2026-03-066:241 reply

      AI dramatically increases velocity. But is velocity productivity? Productivity relative to which scope: you, the team, the department, the company?

      The question is really, velocity _of what_?

      I got this from a HN comment. It really hit for me because the default mentality for engineers is to build. The more you build the better. That's not "wrong" but in a business setting it is very much necessary but not sufficient. And so whenever we think about productivity, impact, velocity, whatever measure of output, the real question is _of what_? More code? More product surface area? That was never really the problem. In fact it makes life worse majority of the time.

      • By mattmanser 2026-03-067:05

        The real question is, is it increasing their velocity?

        They've already admitted they just 'throw the code away and start again'.

        I think we've got another victim of perceived productivity gains vs actual productivity drop.

        People sitting around watching Claude churn out poor code at a slower rate than if they just wrote it themselves.

        Don't get me wrong, great for getting you started or writing a little prototype.

        But the code is bad, riddled with subtle bugs and if you're not rewriting it and shoving large amounts of AI code into your codebase, good luck in 6-12 months time.

    • By tripledry 2026-03-069:221 reply

      Something I've been thinking about lately is if there is value in understanding the systems we produce and if we expected to?

      If I can just vibe and shrug when someone asks why production is down globally then I'm sure the amount of features I can push out increases, but if I am still expected to understand and fix the systems I generate, I'm not convinced it's actually faster to vibe and then try to understand what's going on rather than thinking and writing.

      In my experience the more I delegate to AI, the less I understand the results. The "slowness and thinking" might just be a feature not a bug, at times I feel that AI was simply the final straw that finally gave the nudge to lower standards.

      • By joe_mamba 2026-03-069:271 reply

        >if I can just vibe and shrug when someone asks why production is down globally

        You're pretty high up in the development, decision and value-addition chain, if YOU are the responsible go-to person for these questions. AI has no impact on your position.

        • By tripledry 2026-03-0610:23

          Naa, I'm just a programmer. Experience may vary depending on company and country, for me this has been true from tiny startups to global corporations.

          Tangential, I don't even know what "responsible" in the corporate world means anymore, it seems to me no one is really responsible for anything. But the one thing that's almost certain is that I will fix the damn thing if I made it go boom.

    • By kodablah 2026-03-065:57

      > People who are saying they're not seeing productivity boost, can you please share where is it failing?

      At review time.

      There are simply too many software industries that can't delegate both authorship _and_ review to non-humans because the maintenance/use of such software, especially in libraries and backwards-compat-concerning environments, cannot justify an "ends justifies the means" approach (yet).

    • By msvana 2026-03-067:131 reply

      I work as an ML engineer/researcher. When I implement a change in an experiment it usually takes at least an hour to get the results. I can use this time to implement a different experiment. Doesn't matter if I do it by hand or if I let an agent do it for me, I have enough time. Code isn't the bottleneck.

      I also heard an opinion that since writing code is cheap, people implement things that have no economic value without really thinking it through.

      • By apsurd 2026-03-067:201 reply

        +1 on the economic value line. Not everything needs to be about money but if you get paid to ship code it's about money. And now we have coworkers shipping insane amounts of "features" because it's all free to ship and being an engineer, it ends there.

        Only it doesn't, there's product positioning, UX, information architecture, onboarding and training, support, QA, change management, analytics, reporting… sigh

        • By embedding-shape 2026-03-0610:52

          > but if you get paid to ship code it's about money.

          Tip to budding software engineers: try to not work in these sort of places, as they're about "looking busy" rather than engineering software, where the latter is where real long-lasting things are built, and the former is where startup founders spend most their money.

          The last paragraph is where the tricky and valuable parts are, and also where AI isn't super helpful today, and where you as a human can actually help out a lot if you're just 10% better than the rest of the "engineers" who only want to ship as fast as possible.

    • By kranke155 2026-03-068:401 reply

      I work in commercials.

      We can now make 1$ million dollar commercials with 100,000$ or less. So a 90% reduction in costs - if we use AI.

      The issue is they don’t look great. AI isn’t that great at some key details.

      But the agencies are really trying to push for it.

      They think this is the way back to the big flashy commercials of old. Budgets are lower than ever, and shrinking.

      Big issue here is really the misunderstanding of cause - budgets are lower, because advertising has changed in general (TV is less and less important ) and a lot of studies showed that advertising is actually not all that effective.

      So they are grabbing onto a lifeboat. But I’m worried there’s no land.

      I’ve planned my exit.

      • By uxcolumbo 2026-03-068:452 reply

        Advertisement is not that effective in general or just for certain channels, i.e. TV?

        Also what are you existing to?

        • By kranke155 2026-03-069:041 reply

          So my understanding - from a friend at WPP who told me the same and from a freakonomics episode - is that advertising was wildly oversold before digital.

          When the metrics arrived with digital, they saw that advertising, in some ways, was just not as effective as they’d hoped. In some ways the ROI wasn’t there. Seth Godin agrees. He says that advertising in the digital era could be as simple as just having a good product. I think this is Tesla’s position on it - make the best product and the internet takes care of it.

          Legacy companies have kept large ad budgets but those are diminishing. From what I spoke with my friend at WPP, he said their data science team showed that outside of a new product or a product that is not recognised by consumers, the actual outcomes from ads are marginal or incremental. Thats what he told me. If your product is already known to consumers, the ROI is questionable.

          • By mikkupikku 2026-03-0611:001 reply

            Advertising's foremost job is to sell the premise of advertising to business management. Selling the business's product is always secondary to that.

            • By ambicapter 2026-03-0613:091 reply

              Always felt suspicious to me that so much of company dynamics are basically about selling yourself to management...and there's one team in the company who's full-time job is selling? Wonder how that will turn out.

              • By gzread 2026-03-0614:02

                None of my coworkers could figure out why I was laid off, and were shocked because I was important to getting the work done, but management made it clear I hadn't been selling myself to management.

        • By kranke155 2026-03-069:08

          My exit is storytelling. I think that’s the only thing that will remain. I suspect humans will still want to hear stories about and from other humans.

          There’s something about AIs that feels wrong for storytelling. I just don’t think people will want AIs to tell them stories. And if they do… Well, I believe in human storytelling.

    • By exfalso 2026-03-0612:55

      It's failing when there is no data in the training set, and there are no patterns to replicate in the existing code base.

      I can give you many, many examples of where it failed for me:

      1. Efficient implementation of Union-Find: complete garbage result 2. Spark pipelines: mostly garbage 3. Fuzzer for testing something: half success, non-replicateable ("creative") part was garbage. 4. Confidential Computing (niche): complete garbage if starting from scratch, good at extracting existing abstractions and replicating existing code.

      Where it succeeds: 1. SQL queries 2. Following more precise descriptions of what to do 3. Replicating existing code patterns

      The pattern is very clear. Novel things, things that require deeper domain knowledge, coming up with the to-be-replicated patterns themselves, problems with little data don't work. Everything else works.

      I believe the reason why there is a big split in the reception is because senior engineers work on problems that don't have existing solutions - LLMs are terrible at those. What they are missing is that the software and the methodology must be modified in order to make the LLM work. There are methodical ways to do this, but this shift in the industry is still in baby shoes, and we don't yet have a shared understanding of what this methodology is.

      Personally I have very strong opinions on how this should be done. But I'm urging everyone to start thinking about it, perhaps even going as far as quitting if this isn't something people can pursue at their current job. The carnage is coming:/

    • By belZaah 2026-03-067:10

      I don’t think the objections are not necessarily in terms of lack of productivity although my personal experience is not that of massive productivity increases. The fact that you are producing code much faster is likely just to push the bottleneck somewhere else. Software value cycles are long and complicated. What if you run into an issue in 5 years the LLM fails to diagnose or fix due to complex system interactions? How often would that happen? Would it be feasible to just generate the whole thing anew matching functionality precisely? Are you making the right architecture choices from the perspective of what the preferred modus operandi of an llm is in 5 years? We don’t know. The more experienced folks tend to be conservative as they have experienced how badly things can age. Maybe this time it’ll be different?

    • By dumfries 2026-03-069:44

      "it works" is a very low standard when it comes to software engineering. Why are we not holding AI generated code to the same standards as we hold our peers during code reviews?

      I have never heard anyone say "it works" as a positive thing when reviewing code..

      Yes, there is a productivity boost but you can't tell me there is no decrease in quality

    • By oytis 2026-03-068:452 reply

      > I have stopped writing code, occasionally I jump into the changes proposed by LLM and make manual edits if it is feasible, otherwise I revert changes and ask it to generate again but based on my learnings from the past rejected output

      Isn't it a very inefficient way to learn things? Like, normally, you would learn how things work and then write the code, refining your knowledge while you are writing. Now you don't learn anything in advance, and only do so reluctantly when things break? In the end there is a codebase that no one knows how it works.

      • By throwaw12 2026-03-069:221 reply

        > Isn't it a very inefficient way to learn things?

        It is. But there are 2 things:

        1. Do I want to learn that? (if I am coming back to this topic again in 5 months, knowledge accumulates, but there is a temptation to finish the thing quickly, because it is so boring to swim in huge legacy codebase)

        2. How long it takes to grasp it and implement the solution? If I can complete it with AI in 2 days vs on my own in 2 weeks, I probably do not want to spend too much time on this thing

        as I mentioned in other comments, this is exactly makes me worried about future of the work I will be doing, because there is no attachment to the product in my brain, no mental models being built, no muscles trained, it feels someone else's "work", because it explores the code, it writes the code. I just judge it when I get a task

        • By oytis 2026-03-069:351 reply

          I don't know where it goes, but it sounds pretty dumb for the companies involved too. Tech companies are in the business of nurturing teams knowledgeable in things so they can build something that gives them an advantage over competition. If there is no knowledge being built, there is no advantage and no tech business.

          • By hobofan 2026-03-069:49

            > Tech companies are in the business of nurturing teams knowledgeable in things

            It pains the anti-capitalist fibers in my body to say this, but no they are not. At the maximum the value is in organizational knowledge and existing assets (= source code, documentation), so that people with the least knowledge possible can make changes. In software companies in general, technical excellence and knowledge is not strongly correlated with economic success as long as you clear a certain bar (that's not that high). In comparison, in hardware/engineering companies, that's a lot more correlated.

            In the concrete example of a legacy codebase we have here, there is even less value in trying to build up knowledge in the company, as it has already been decided that the system is to be discarded anyways.

      • By hobofan 2026-03-068:481 reply

        > you would learn how things work and then write the code

        In a legacy codebase this may require learning a lot of things about how things work just to make small changes, which may be much less efficient.

        • By oytis 2026-03-069:19

          I might still be naive about the industry, but if you don't know how the legacy codebase works, you might either delegate the change to someone else in the company who does, or, if there is no one left, use this opportunity to become the person who knows at least something about it.

    • By pinkmuffinere 2026-03-066:141 reply

      I asked opus 4.6 how to administer an A/B test when data is sparse. My options are to look at conversion rate, look at revenue per customer, or something else. I will get about 10-20k samples, less than that will add to cart, less than that will begin checkout, and even less than that will convert. Opus says I should look at revenue per customers. I don't know the right answer, but I know it is not to look at revenue per customers -- that will have high variance due to outlier customers who put in a large order. To be fair, I do use opus frequently, and it often gives good enough answers. But you do have to be suspicious of its responses for important decisions.

      Edit: Ha, and the report claims it's relatively good at business and finance...

      Edit 2: After discussion in this thread, I went back to opus and asked it to link to articles about how to handle non-normally distributed data, and it actually did link to some useful articles, and an online calculator that I believe works for my data. So I'll eat some humble pie and say my initial take was at least partially wrong here. At the same time, it was important to know the correct question to ask, and honestly if it wasn't for this thread I'm not sure I would have gotten there.

      • By onion2k 2026-03-066:561 reply

        A/B tests are a statistical tool, and outliers will mess with any statistical measure. If your data is especially prone to that you should be using something that accounts for them, and your prompt to Opus should tell it to account for that.

        A good way to use AI is to treat it like a brilliant junior. It knows a lot about how things work in general but very little about your specific domain. If your data has a particular shape (e.g lots of orders with a few large orders as outliers) you have to tell it that to improve the results you get back.

        • By pinkmuffinere 2026-03-066:591 reply

          I did tell it that I expect to see something like a power-law distribution in order value, so I think I pretty much followed your instructions here. Btw, if you do know the right thing to do in my scenario, I'd love to figure it out. This is not my area of expertise, and just figuring it out through articles so far.

          • By Karrot_Kream 2026-03-067:161 reply

            I recommend reading Wikipedia and talking to LLMs to get this one. Order values do follow power-law distributions (you're probably looking for an exponential or a Zipf distribution.) You want to ask how to perform a statistical test using these distributions. I'm a fan of Bayesian techniques here, but it's up to you if you want to use a frequentist approach. If you can follow some basic calculus you can follow the math for constructing these statistical tests, if not some searching will help you find the formulas you need.

            • By pinkmuffinere 2026-03-067:381 reply

              Thanks for the suggestions! I didn't want to do the math myself, but I did take your suggestion and found some articles discussing ways to make it work even with a non-normal distribution:

              - https://cxl.com/blog/outliers/

              - https://www.blastx.com/insights/the-best-revenue-significanc...

              - (online tool to calculate significance) https://www.blastx.com/rpv-calculator

              I'm not checking their math, but the articles make sense to me, and I trust they did implement it correctly. In the end the LLM did get me to the correct answer by suggesting the articles, so I guess I should eat some humble pie and say it _did_ help me. At the same time, if I didn't have the intuition that using rpv as-is in a t-test would be noisy, and the suggestions from this comment thread, I think I could have gone down the wrong path. So I'm not sure what my conclusion is -- maybe something like LLMs are helpful once you ask the right question.

              • By Karrot_Kream 2026-03-068:371 reply

                One heuristic I like to use when thinking about this question (and I honestly wish the answer space here were less emotionally charged, so we could all learn from each other) is that: LLMs need a human to understand the shape of the solution to check the LLM's work. In fields that I have confirmed expertise in, I can easily nudge and steer the LLM and only skim its output quickly to know if it's right or wrong. In fields I don't, I first ask the LLM for resources (papers, textbooks, articles, etc) and familiarize myself with some initial literature first. I then work with the LLMs slowly to make a solution. I've found that to work well so far.

                (I also just love statistics and think it's some of the most applicable math to everyday life in everything from bus arrival times to road traffic to order values to financial markets.)

                • By pinkmuffinere 2026-03-0617:44

                  I think this is a _really_ insightful answer about effectively working with LLMs. And you’re winning me over on statistics too :)

    • By gurghet 2026-03-0611:481 reply

      Basically it tricks you into making the code less maintainable, so while it seems to boost productivity, it's just delaying huge failures.

      • By rootusrootus 2026-03-0617:01

        Exactly this, IMO. We are in a race to see whether the mountain of technical debt that AI is creating grows faster than AI's ability to whittle it down later.

    • By iugtmkbdfil834 2026-03-068:47

      I don't want to generalize from my specific situation too much, but I want to offer an anecdote from my neck of the woods. On my personal sub, I agree it is kinda crazy the kind of projects I can get into now with little to no prior knowledge.

      On the other hand, our corporate AI is.. not great atm. It was briefly kinda decent and then suddenly it kinda degraded. Worst case is, no one is communicating with us so we don't know what was changed. It is possible companies are already trying to 'optimize'.

      I know it is not exactly what you are asking. You are saying capability is there, but I am personally starting to see a crack in corporate willingness to spend.

    • By staticassertion 2026-03-068:283 reply

      When it comes to novel work, LLMs become "fast typers" for me and little more. They accelerate testing phases but that's it. The bar for novelty isn't very high either - "make this specific system scale in a way that others won't" isn't a thing an LLM can ever do on its own, though it can be an aid.

      LLMs also are quite bad for security. They can find simple bugs, but they don't find the really interesting ones that leverage "gap between mental model and implementation" or "combination of features and bugs" etc, which is where most of the interesting security work is imo.

      • By asadm 2026-03-068:341 reply

        I think your analysis is a bit outdated these days or you may be holding it wrong.

        I am doing novel work with codex but it does need some prompting ie. exploring possibilities from current codebase, adding papers to prompt etc.

        For security, I think I generally start a new thread before committing to review from security pov.

        • By staticassertion 2026-03-068:361 reply

          You can do novel work with an LLM. You can. The LLM can't. It can be an aid - exploring papers, gathering information, helping to validate, etc. It can't do the actual novel part, fundamentally it is limited to what it is trained on.

          If you are relying on the LLM and context, then unless your context is a secret your competitor is only ever one prompt behind you. If you're willing to pursue true novelty, you need a human and you can leap beyond your competition.

          • By bdangubic 2026-03-0611:091 reply

            of course you need a human but do not need nearly as many humans as there are currently in the labor force

            • By staticassertion 2026-03-0612:481 reply

              Maybe, but I'm not really convinced. LLMs make some aspects of the job faster, mainly I don't have to type anymore. But... that was always a relatively small portion of the job. Design, understanding constraints, maintaining and operating code, deciding what to do, what not to do, when to do it, gaining consensus across product, eng, support, and customers, etc. I do all of those things as an engineer. Coding faster is really awesome, it's so nice, and I can whip up POCs for the frontend etc now, and that's accelerating development... but that's it.

              The reality is that a huge portion of my time is spent doing similar work and what LLMs largely do is pick up the smaller tasks or features that I may not have prioritized otherwise. Revolutionary in one sense, completely banal and a really minor part of my job in many others.

              • By bdangubic 2026-03-0614:221 reply

                I think the core issue (evidenced by constant stream of debates on HN) is the everyone’s experience with LLMs is different. I think we can all agree that some experiences are like yours while there are others that are vastly different than yours. Sometimes I hear “you just don’t know how to use them etc…” as if there is some magic setup that makes them do shit but the reality is that our actual jobs are drastically different even though we all technically have same titles. I have been a contractor for a decade now and have been on projects that require real “engineers” doing real hardcore shit. I have also been on projects where tens of people are doing work I can train my 12-year old daughter to be proficient in a month. I would gauge that percentage of the former is much smaller than later

                • By staticassertion 2026-03-0712:59

                  I don't think this is an issue of experience here. I don't know that anyone has claimed that LLMs can create truly novel solutions to complex problems given that the technology is token prediction.

      • By truetraveller 2026-03-0611:04

        This is basically my take as well!

    • By girvo 2026-03-0613:56

      Sometimes I’m scared.

      Sometimes I realise that this particular task has been slower than if I’d done it myself when I take in to account full wall clock time.

      I can’t tell what type of task is going to work ahead of time yet.

    • By vividfrier 2026-03-0613:18

      Same. Whenever an article like this one pops up the comments seem to be filled with confirmation bias. People who don't see a productivity boost agree with the article.

      I work at tech company just outside of big tech and I feel fairly confident that we won't have a need for the amount of developers we currently have within 3-4 years.

      The bottleneck right now is reviewing and I think it's just a matter of time before our leadership removes the requirement for human code reviews (I am already seeing signs of this ("Maybe for code behind feature flags we don't need code reviews?").

      Whenever there's an incident, there is a pagerduty trigger to an agent looking at the metrics, logs, software component graphs, and gives you an hypothesis on what the incident is due to. When I push a branch with test failures, I get one-click buttons in my PR to append commits fixing those tests failures (i.e. an agent analyses the code, the jira ticket, the tests, etc. and suggests a fix for the tests failing). We have a Slack agent we can ping in trivial feature requests (or bugs) in our support channels.

      The agents are being integrated at every step. And it's not like the agents will stop improving. The difference between GPT3.5 and Opus 4.6 is so massive. So what will the models look like in 5 years from now?

      We're cooked and the easiest way to tell someone works at a company who hasn't come very far in their AI journey is that they're not worried.

    • By dataflow 2026-03-066:00

      I feel like this might be heavily dependent on both your task and the AI you're using? What language do you code in and what AI do you use? And are your tasks pretty typical/boilerplate-y with prior art to go off of, or novel/at-the-edge-of-tech?

    • By sivanmz 2026-03-067:41

      It’s been my experience as of recently. I point it at an issue tracker and ask it to investigate, write a test to reproduce the problem and plan a fix together. There’s lots of hand holding from me but it saves me a lot of work and I’ve been surprised by its comfort with legacy code bases. For now I feel empowered, and I’m actually working more intensively, but I was wondering to myself if I’m going run out of work this year. Interestingly, our metrics show that output is slowed by increased workload on reviewers.

    • By matheusmoreira 2026-03-0623:54

      Terrifying doesn't quite say it. The situation we're in is either we achieve a post-scarcity society or we'll all die.

    • By aurareturn 2026-03-065:523 reply

        People who are saying they're not seeing productivity boost, can you please share where is it failing?
      
      Believe it or not, I still know many devs who do not use any agents. They're still using free ChatGPT copy and paste.

      I'm going to guess that many people on HN are also on the "free ChatGPT isn't that good at programming" train.

      • By dataflow 2026-03-065:541 reply

        Which one would you recommend as the best right now? Claude Code?

        • By redhed 2026-03-0618:46

          I have been having a lot of success with Cursor. I like being able to switch between Anthropic and OpenAI models. Claude Code does gives way more tokens/$ than Cursor right now though.

      • By throwaw12 2026-03-065:551 reply

        > They're still using free ChatGPT copy and paste

        Probably that's the reason why some people are sure their job is still safe.

        Nature of job is changing rapidly

        • By aurareturn 2026-03-065:572 reply

          I totally get tech CEOs who threaten to fire their devs who do not embrace AI tools.

          I'm not a tech CEO but people who are anti-LLM for programming have no place on my team.

          • By cherrycherry98 2026-03-073:24

            Anecdotally, the person on my team who is producing the most output and the highest quality happens to explicitly shun LLMs and is also relatively young. Other team members embracing AI spend their day being misled by chat bots, trying to get their agents to work properly by toying with contexts (just needs a little more knowledge!), and producing verbose code that's hard to review and with obvious bugs when you actually think about what it's doing. My favorite is how everything is excessively commented but more than once I've caught the code not matching what the comment said it would do!

            FWIW I find it useful if I know exactly what I want and it's quicker to prompt it than type it myself. Also for research and building understanding it's generally good. I still catch it being wrong on details of you're really paying attention or literally contradicting itself between prompts. That gives me a lot of pause about trusting things it told me that I just accepted as fact without having enough knowledge myself to question it.

          • By salawat 2026-03-066:063 reply

            And you are paying for their tokens on top of their salary, right? Right?

            • By aurareturn 2026-03-066:103 reply

              You can do a lot with just $20 Codex CLI subscription. Tokens are cheap compared to the $20k we're paying for a dev each month.

              • By ido 2026-03-066:421 reply

                Even the $200 claude max monthly subscription is peanuts compared to salary cost.

                • By monksy 2026-03-067:111 reply

                  Tell that to the company that I was just at that cut Intelij licenses as cost cutting measures.

                  • By aurareturn 2026-03-067:351 reply

                    If they really want to cut cost, fire the worst dev on the team and use that money to give everyone a Codex subscription.

                    • By KronisLV 2026-03-068:422 reply

                      Or better yet, fire the managers or bean counters that think decreasing everyone’s productivity is good for long term savings.

                      I’m reminded of https://www.joelonsoftware.com/2000/08/09/the-joel-test-12-s...

                      • By mikkupikku 2026-03-0611:071 reply

                        Fire the middle management, HR, and etc that have been enthusiastically using AI to do their jobs for the past two or three years already. 90% of them can be replaced by an agent with access to an email account.

                        • By thewebguyd 2026-03-0616:35

                          Tbh, if companies want to use AI to lay off and cut costs, that's exactly where they should be doing it, not engineering.

                          How much bloat and bureaucracy bottleneck is sitting in middle management whose favorite past time is wasting everyone's time on meetings that could have been an email? HR? Not the execs, but the HR drones that do nothing but answer employee questions about policy, could have already been replaced with not even an AI, just an old school chatbot, a long time ago.

                          Instead of cutting engineers, cut the non-tech jobs, flatten the structure.

                      • By monksy 2026-03-070:59

                        Oh they laid people off so they could outsource more to India. Despite the managers reminding them, the cost is 1/3 the cost but 3x slower. American devs were 5x more productive overall.

              • By hdgvhicv 2026-03-067:413 reply

                Amazes me that people pay 20k a month for a dev rather than paying 2k a month for one in Poland or 1k a month for one from India

                There’s obviously a benefit of paying higher rates for US programmers, but does that benefit change when llms are thrown into the mix

                • By apercu 2026-03-0611:491 reply

                  My experience with outsourcing over 20+ years (Russia, Romania, India, South America) is that you just move money around when you do it.

                  It takes more planning, more specification, more coordination, more QA. The quality is almost always worse, and remediation takes forever. So your BA, QA and PM time goes way up and absorbs any cost savings.

                  YMMV.

                  • By thewebguyd 2026-03-0616:36

                    Sounds like an LLM, tbh. Using Claude also takes more planning, more explicit specification, prompting, more manual review, more QA.

                • By aurareturn 2026-03-0612:51

                  Makes no sense because LLMs makes it far less worth it to outsource developers.

                • By forgotlastlogin 2026-03-0611:12

                  2k in Poland you say...

              • By baq 2026-03-066:36

                Exactly, the $20 codex is so good value it’s irresponsible to not give it to everyone. Claude code $20 is otoh pointless, the limits are good enough for 10 mins of work twice per business day.

            • By onion2k 2026-03-067:00

              Every business that's taking AI seriously is giving their team enterprise accounts to AI services. Otherwise you have no control over where your code, data, company info, etc is going.

              Someone deciding to drop a spreadsheet of customer data into their personal AI account to increase their productivity would be catastrophic for business, so you need rules. And rules means paying for enterprise AI tooling.

            • By mikkupikku 2026-03-0611:041 reply

              "Bring your own tools" is not exactly novel in the workplace. Maybe so for office workers, but not more generally. Anyway, these particular tools are cheap enough that it hardly even matters who is expected to pay for them.

              The $20 a month tier in particular is a trivial expense, on par with businesses that expect their workers to wear steel toed shoes. Some may give workers a little stipend to buy those boots, some not. Either way, it doesn't really matter.

              • By thewebguyd 2026-03-0616:42

                Just because it's not novel, doesn't mean it's right. I also don't agree with, for example, many mechanics being forced to buy their own tools (especially what little they get paid).

                I don't do tech outside of 9-5, so either my employer pays for it all, or I don't use it. Simple as that. Thankfully, they do pay for it, but I couldn't imagine working somewhere that says "You need to use AI" and then not providing it on their dime.

                Quite frankly it should be regulation that if a W2 employee needs something to perform their job duties, the employer must provide it.

      • By salawat 2026-03-066:051 reply

        Not everyone has the capability to rent out data center tier hardware to just do their job. These things require so much damn compute you need some serious heft to actually daisy chain enough stages either in parallel or deep to get enough tokens/sec for the experience to go ham. If you're making bags o' coke money, and deciding to fund Altman's, Zuckernut's or Amazon/Google's/Microsoft's datacenter build out, that's on you. Rest of us are just trying to get by on bits and bobs we've kept limping over the years. If opencode is anything to judge the vibecoded scene by, I'm fairly sure at some point the vibe crowd will learn the lesson of isolating the most expensive computation ever from the hot loop, then maybe find one day all they needed was maybe something to let the model build a context, and a text editor.

        Til then wtf_are_these_abstractions.jpg

        • By ijk 2026-03-0618:23

          This is my current problem: I can get work to pay for a Claude Max subscription, but for personal use or to learn how to use it that's a big price tag.

          I worry that we're returning to an era of renting core development tools. After the huge benefits from free and open source tools, that's a bitter pill to swallow.

    • By boxedemp 2026-03-065:43

      I'm with you. The project I'm working on is moving at phenomenal velocity. I'm basically spending my time writing specs and performing code reviews. As long as my code review comments and design docs are clear I get a secure, scalable, and resilient system.

      Tests were always important, but now they are the gatekeepers to velocity.

    • By motbus3 2026-03-0612:48

      I think you can more stuff done earlier but the quality is not good or it doesn't work as expected if you tinker with it enough. Fixing the issues from the generated code usually doesn't work at all

    • By RandomLensman 2026-03-066:42

      Outside of coding/non-physical areas, the impact can be quite muted. I haven't seen much impact on surgical procedures, for example (but maybe others have?).

    • By KronisLV 2026-03-067:042 reply

      I’m currently working across like 5 projects (was 4 last week but you know how it is). I now do more in days than others might in a week.

      Yesterday a colleague didn’t quite manage to implement a loading container with a Vue directive instead of DOM hacks, it was easier for me to just throw AI at the problem and produced a working and tested solution and developer docs than to have a similarly long meeting and have them iterate for hours.

      Then I got back to training a CNN to recognize crops from space (ploughing and mowing will need to be estimated alongside inference, since no markers in training data but can look at BSI changes for example), deployed a new version of an Ollama/OpenAI/Anthropic proxy that can work with AWS Bedrock and updated the docs site instructions, deployed a new app that will have a standup bot and on-demand AI code review (LiteLLM and Django) and am working on codegen to migrate some Oracle forms that have been stagnating otherwise.

      It’s not funny how overworked I am and sure I still have to babysit parallel Claude Code sessions and sometimes test things manually and write out changes, but this is a completely different work compared to two or three years ago.

      Maybe the problem spaces I’m dealing with are nothing novel, but I assume most devs are like that - and I’d be surprised at people’s productivity not increasing.

      When people nag in meetings about needing to change something in a codebase, or not knowing how to implement something and its value add, I’ll often have something working shortly after the meeting is over (due to starting during it).

      Instead of sending adding Vitest to the backlog graveyard, I had it integrated and running in one or two evenings with about 1200 tests (and fixed some bugs). Instead of talking about hypothetical Oxlint and Oxfmt performance improvements, I had both benchmarked against ESLint and Prettier within the hour.

      Same for making server config changes with Ansible that I previously didn’t due to additional friction - it is mostly just gone (as long as I allow some free time planned in case things vet fucked up and I need to fix them).

      Edit: oh and in my free time I built a Whisper + VLM + LLM pipeline based on OpenVINO so that I can feed it hours long stream VODs and get an EDL cut to desired length that I can then import in DaVinci Resolve and work on video editing after the first basic editing prepass is done (also PyScene detect and some audio alignment to prevent bad cuts). And then I integrated it with subscription Claude Code, not just LiteLLM and cloud providers with per-token costs for the actual cuts making part (scene description and audio transcriptions stay local since those don't need a complex LLM, but can use cloud for cuts).

      Oh and I'm moving from my Contabo VPSes to running stuff inside of a Hetzner Server Auction server that now has Proxmox and VMs in that, except this time around I'm moving over to Ansible for managing it instead of manual scripts as well, and also I'm migrating over from Docker Swarm to regular Docker Compose + Tailscale networks (maybe Headscale later) and also using more upstream containers where needed instead of trying to build all of mine myself, since storage isn't a problem and consistency isn't that important. At the same time I also migrated from Drone CI to Woodpecker CI and from Nexus to Gitea Packages, since I'm already using Gitea and since Nexus is a maintenance burden.

      If this becomes the new “normal” in regards to everyone’s productivity though, there will be an insane amount of burnout and devaluation of work.

      • By Karrot_Kream 2026-03-067:10

        > When people nag in meetings about needing to change something in a codebase, or not knowing how to implement something and its value add, I’ll often have something working shortly after the meeting is over (due to starting during it).

        We've started building harnesses to allow people who don't understand code to create PRs to implement their little nags. We rely on an engineer to review, merge, and steward the change but it means that non-eng folks do not rely on us as a gate. (We're a startup and can't really afford "teams" to do this hand-holding and triage for us.)

        As you say we're all a bit overworked and burned out. I've been context switching so much that on days when I'm very productive I've started just getting headaches. I'm achieving a lot more than before but holding the various threads in my head and context switching is just a lot.

      • By leptons 2026-03-067:571 reply

        >I now do more in days than others might in a week.

        I've always done more in days than others might in a week. YMMV.

        • By sph 2026-03-0614:321 reply

          So do I, this is why I work 15 hours a week [1] and laugh at those that use this new productivity tool to work themselves even harder for the same pay. Wasn’t the point of automation to work less?

          1: pre-AI. Not keen on becoming a manager of an idiot savant, so I’m planning my exit.

          • By leptons 2026-03-0718:10

            I doubt you're as productive as someone competent not using AI.

    • By drekipus 2026-03-069:44

      > my job is easier now, I do less. > I am terrified about what's coming.

      God I hope I never ever have to work with you

    • By fulafel 2026-03-066:00

      A terminology tangent because it's an econ publication: Notice that the article doesn't talk about productivity.

      Productivity is a term of art in economics and means you generate more units of output (for example per person, per input, per wages paid) but doesn't take quality or otherwise desireability into account. It's best suited for commodities and industrial outputs (and maybe slop?).

    • By randusername 2026-03-0617:37

      I see an individual productivity boost, but not necessarily a collective one.

      I don't think features per hour is really what is holding back most established businesses.

      My experiences suggest that we still have some time before the people that understand the plumbing of the business _and_ AI bubble up to positions of authority through wielding it practically and successfully at increasingly greater scale.

    • By lm28469 2026-03-069:47

      Meanwhile gemini tells me my go code doesn't compile (it does)

      Gaslight me by telling me I must be a time traveler because I use go 1.26 but the latest version actually is 1.24

      And tell me I can't use wg.Go() because this function does not exist (it does)

    • By truetraveller 2026-03-065:493 reply

      You were probably deficient in RESEARCH skills before. No offense to you, since I was also like this once. LLMs research and put the results on the plate. Yes, for people who were deficient in research skills, I can see 2-3x improvements.

      Note1: I have "expert" level research skills. But LLMs still help me in research, but the boost is probably 1.2x max. But

      Note2: By research, I mean googling, github search, forum search, etc. And quickly testing using jsfiddle/codepen, etc.

      • By throwaw12 2026-03-065:541 reply

        no worries, I do not get offended quickly.

        But I also think you are overestimating your RESEARCH skills, even if you are very good at research, I am sure you can't read 25 files in parallel, summarize them (even if its missing some details) in 1 minute and then come up with somewhat working solution in the next 2 minutes.

        I am pretty sure, humans can't comprehend reading 25 code files with each having at least 400 lines of non-boilerplate code in 2 minutes. LLM can do it and its very very good at summarizing.

        I can even steer its summarizing skills by prompting where to focus on when its reading files (because now I can iterate 2-3 times for each RESEARCH task and improve my next attempt based on shortcomings in the previous attempt)

        • By truetraveller 2026-03-0610:54

          OK, it's not just RESEARCH, but "RESEARCHability" of the source content [in this case code], and also critical analysis ability [not saying you are deficient in anything, speaking in general terms].

          In this example, if the 25 files are organized nicely, and I had I nice IDE that listed class/namespace members of each file neatly, I might take 30 minutes to understand the overall structure.

          Morever, If I critically analyzed this, I would ask "how many times does this event of summarizing 25 files happen"? I mean, are we changing codebases every day? No, it's a one time cost. Moreover, manually going through will provide insight not returned by LLM.

          Obviously, every case is different, and perhaps you do need to RESEARCH new codebases often, I dunno!

      • By siva7 2026-03-068:381 reply

        Ok Mr. Expert Level Researcher, go back and read the comment of parent again to find out that it has nothing to do with deficiency in research skills.

        • By truetraveller 2026-03-0610:56

          Lol! Didn't mean any harm, just giving my 2cents!

      • By throwaw12 2026-03-067:021 reply

        please don't change your comment constantly (or at maybe add UPDATE 1/2/3), because you had different words before, like you were saying something in this fashion:

        * you probably lack good RESEARCH skills

        * I can see at most 1.25x improvements - now it is 2-3x

        By updating your comment you are making my reply irrelevant to your past response

        • By truetraveller 2026-03-0610:28

          Apologies, I changed this within a ~10 minute period. Never knew you would actually see it so fast.

    • By therealdrag0 2026-03-065:491 reply

      I can only explain it by people not having used Agentic tools and or only having tried it 9 months ago for a day before giving up or having such strict coding style preferences they burn time adjusting generated code to their preferences and blaming the AI even though they’re non-functional changes and they didn’t bother to encode them into rules.

      The productivity gains are blatantly obvious at this point. Even in large distributed code bases. From jr to senior engineer.

      • By MattGaiser 2026-03-067:011 reply

        I can see someone who is very particular about their way being the right way having issues with it. I’m very much the kind of person who believes that if I can’t write a failing test, I don’t have a very serious case. A lot of devs aren’t like that.

        • By layer8 2026-03-0612:22

          Sometimes you’re unable to write a failing test because the code is such that you can’t reliably reason about it, and hence have a hard time finding the cases where it will do the wrong thing. Being able to reason about code in that way is as important as for the code to be testable.

    • By zozbot234 2026-03-0611:47

      > I am terrified about what's coming

      Why? This is great. AI fixing up huge legacy codebases is just taking the jobs humans would never want to do.

  • By bandrami 2026-03-060:1515 reply

    I don't write code for a living but I administer and maintain it.

    Every time I say this people get really angry, but: so far AI has had almost no impact on my job. Neither my dev team nor my vendors are getting me software faster than they were two years ago. Docker had a bigger impact on the pipeline to me than AI has.

    Maybe this will change, but until it does I'm mostly watching bemusedly.

    • By kdheiwns 2026-03-063:337 reply

      Yep. All AI has done for me is give me the power of how good search engines were 10+ years ago, where I could search for something and find actually relevant and helpful info quickly.

      I've seen lots of people say AI can basically code a project for them. Maybe it can, but that seems to heavily depend on the field. Other than boilerplate code or very generic projects, it's a step above useless imo when it comes to gamedev. It's about as useful as a guy who read some documentation for an engine a couple years ago and kind of remembers it but not quite and makes lots of mistakes. The best it can do is point me in the general direction I need to go, but it'll hallucinate basic functions and mess up any sort of logic.

      • By kranner 2026-03-064:551 reply

        My experience is the same. There are modest gains compensating for lack of good documentation and the like, but the human bottlenecks in the process aren't useless bureaucracy. Whether or not a feature or a particular UX implementation of it makes sense, these things can't be skipped, sped up or handed off to any AI.

        • By freddref 2026-03-066:591 reply

          What are these bottlenecks specifically that you feel are essential?

          Am trying to compare this to reports that people are not reviewing code any more.

          • By kranner 2026-03-069:15

            When features and their exact UI implementations are being developed, feedback and discussions around those things.

      • By bee_rider 2026-03-064:031 reply

        Thinking of it, I haven’t seen as many “copy paste from StackOverflow” memes lately. Maybe LLMs have given people the ability to

        1) Do that inside their IDEs, which is less funny

        2) Generate blog post about it instead of memes

        • By Izkata 2026-03-0619:551 reply

          Python one-upped that a long time ago:

            from stackoverflow import quick_sort
          
          https://github.com/drathier/stack-overflow-import

          • By bandrami 2026-03-0711:28

            30 years ago (well, 28 years ago) we had the Perl Cookbook which was basically Stack Overflow but printed and without annoying moderators closing your questions. That and the camel book never even made it back to my bookshelf because they just always sat open on my desk.

      • By mchaver 2026-03-0612:08

        > All AI has done for me is give me the power of how good search engines were 10+ years ago

        So the good old days before search engines were drowning with ads and dark patterns. My assumption is big LLMs will go in the same direction after market capture is complete and they need to start turning a profit. If we are lucky the open source models can keep up.

      • By throwaw12 2026-03-065:171 reply

        > how good search engines were 10+ years ago

        For me this is a huge boost in productivity. If I remember how I was working in the past (example of Google integration):

        Before:

            * go through docs to understand how to start (quick start) and things to know
            * start boilerplate (e.g. install the scripts/libs)
            * figure out configs to enable in GCP console
            * integrate basic API and test
            * of course it fails, because its Google API, so difficult to work with
            * along the way figure out why Python lib is failing to install, oh version mismatch, ohh gcc not installed, ohh libffmpeg is required,...
            * somehow copy paste and integrate first basic API
            * prepare for production, ohhh production requires different type of Auth flow
            * deploy, redeploy, fix, deploy, redeploy
            * 3 days later -> finally hello world is working
        
        Now:

            * Hey my LLM buddy, I want to integrate Google API, where do I start, come up with a plan
            * Enable things which requires manual intervention
            * In the meantime LLM integrates the code, install lib, asks me to approve installation of libpg, libffmpeg,....
            * test, if fails, feed the error back to LLM + prompt to fix it
            * deploy

        • By noosphr 2026-03-065:44

          This is what you'd use a search engine for 10 years ago.

          The docs used to be good enough that there would be an example which did exactly what you needed more often than the llm gets it right today.

      • By redhed 2026-03-0618:56

        What language/engine did you try it with for gamedev? Just curious if it was weak in a popular engine.

      • By demorro 2026-03-0612:511 reply

        It makes me wonder if the majority of all-in on AI folks are quite young and never experienced pre-enshittification search.

        • By bandrami 2026-03-0614:131 reply

          Also I see so much talk about "boilerplate" I can't help but wonder if people just never had decent text editors, or never bothered to customize them at all?

          • By demorro 2026-03-0614:341 reply

            Aye, I know. Don't get me wrong, I knew that the majority of devs have always been worse than useless, but it's been disconcerting to see quite how much value folk are getting out of agents for problems that have been solved for decades.

            Arguably this solution is "better" because you don't even really need to understand that you have specific problems to have the agent solve them for you, but I fail to see the point of keeping these people employed in that case. If you haven't been able to solve your own workflow issues up until now I have zero trust in you being able to solve business problems.

            • By joquarky 2026-03-0623:43

              > the majority of devs have always been worse than useless

              I disagree with "always".

              This is only the recent wave of brogrammers who care nothing about the quality of the tech and are only in this industry for the gold rush.

              They aren't inherently technically minded, they just know how to schmooze their way around and convince decision makers to follow capricious trends over solid practices.

      • By rhubarbtree 2026-03-0612:031 reply

        Are you using Claude Opus 4.5/6?

        If not, then you’re not close to the cutting edge.

        • By bandrami 2026-03-0612:47

          Until two weeks from now, at which point you'll be hopelessly obsolete. I've seen this treadmill before and am happy to let it settle down first.

    • By thewebguyd 2026-03-060:215 reply

      Same here, more or less, in the ops world. Yeah, I use AI but I can't honestly say it's massively improved my productivity or drastically changed my job in any way other than the emails I get from the other managers at my work are now clearly written by AI.

      I can turn out some scripts a little bit quicker, or find an answer to something a little quicker than googling, but I'm still waiting on others most of the time, the overall company processes haven't improved or gotten more efficient. The same blockers as always still exist.

      Like you said, there has been other tech that has changed my job over time more than AI has. The move to the cloud, Docker, Terraform, Ansible, etc. have all had far more of an impact on my job. I see literally zero change in the output of others, both internally and externally.

      So either this is a massively overblown bubble, or I'm just missing something.

      • By linsomniac 2026-03-064:411 reply

        You're missing something.

        I've been in ops for 30 years, Claude Code has changed how I work. Ops-related scripting seems to be a real sweet spot for the LLMs, especially as they tend to be smaller tools working together. It can convert a few sentences into working code in 15-30 minutes while you do something else. I've given it access to my apache logs Elastic cluster, and it does a great job at analyzing them ("We suspect this user has been compromised, can you find evidence of that?"). It's quite startling, actually, what it's able to do.

        • By thewebguyd 2026-03-065:062 reply

          Yeah, it's useful for scripting, but it's still only marginally faster. It certainly hasn't been "groundbreaking productivity" like it's being sold.

          The problem with analyzing logs is determinism. If I ask Claude to look for evidence of compromise, I can't trust the output without also going and verifying myself. It's now an extra step, for what? I still have to go into Elastic and run the actual queries to verify what Claude said. A saved Kibana search is faster, and more importantly, deterministic. I'm not going to leave something like finding evidence of compromise up to an LLM that can, and does, hallucinate especially when you fill the context up with a ton of logs.

          An auditor isn't going to buy "But Claude said everything was fine."

          Is AI actually finding things your SIEM rules were missing? Because otherwise, I just don't see the value in having a natural language interface for queries I already know how to run, it's less intuitive for me and non deterministic.

          It's certainly a useful tool, there's no arguing that. I wouldn't want to go back to working with out it. But, I don't buy that it's already this huge labor market transformation force that's magically 100x everyone's productivity. That part is 100% pure hype, not reality.

          • By bandrami 2026-03-065:102 reply

            The tolerance for indeterminacy is I think a generational marker; people ~20 years younger than me just kind of think of all software as indeterminate to begin with (because it's always been ridiculously complicated and event-driven for them), and it makes talking about this difficult.

            • By sebmellen 2026-03-065:29

              I shudder to think of how many layers of dependency we will one day sit upon. But when you think about it, aren’t biological systems kind of like this too? Fallible, indeterminable, massive, labyrinthine, and capable of immensely complex and awe inspiring things at the same time…

            • By kiba 2026-03-066:37

              People younger than me are not even adults. I grew up during the dial up era and then the transition to broadband. I don't think software is indeterminate.

          • By linsomniac 2026-03-065:321 reply

            >still only marginally faster.

            Is it? A couple days ago I had it build tooling for a one-off task I need to run, it wrote ~800 lines of Python to accomplish this, in <30m. I found it was too slow, so I got it to convert it to run multiple tasks in parallel in another prompt. Would have taken a couple days for me to build from hand, given the number of interruptions I have in the average day. This isn't a one-off, it's happening all the time.

            • By bandrami 2026-03-072:361 reply

              Did that need to be 800 lines of Python, though, is the question

              • By linsomniac 2026-03-075:36

                NEED to be? No.

                But, to be robust you want a signal handler with clean shutdown, a circuit breaker, argument processing (100 lines right there), logging, reporting progress to our dashboard (it's going to run 10-15 days), checking errors and exceptions, retrying on temp fail, documentation... It adds up.

                So it could be shorter, but it's not like there is anything superfluous in it.

      • By keeda 2026-03-063:122 reply

        > ... but I'm still waiting on others most of the time, the overall company processes haven't improved or gotten more efficient. The same blockers as always still exist.

        And that's the key problem, isn't it? I maintain current organizations have the "wrong shape" to fully leverage AI. Imagine instead of the scope of your current ownership, you own everything your team or your whole department owns. Consider what that would do to the meetings and dependencies and processes and tickets and blockers and other bureaucracy, something I call "Conway Overhead."

        Now imagine that playing out across multiple roles, i.e. you also take on product and design. Imagine what that would do to your company org chart.

        I added a much more detailed comment here: https://news.ycombinator.com/item?id=47270142

        • By applfanboysbgon 2026-03-063:343 reply

          > Imagine instead of

          > Now imagine

          > Imagine what that would do

          Imagine if your grandma had wheels! She'd be a bicycle. Now imagine she had an engine. She could be a motorcycle! Unfortunately for grandma, she lives in reality and is not actually a motorcycle, which would be cool as hell. Our imagination can only take us so far.

          To more substantively reply to your longer linked comment: your hypothesis is that people spend as little as 10% of time coding and the other 90% of time in meetings, but that if they could code more, they wouldn't need to meet other people because they could do all the work of an entire team themselves[1]. The problem with your hypothesis is that you take for granted that LLMs actually allow people to do the work of an entire team themselves, and that it is merely bureacracy holding them back. There have been absolutely zero indicators that this is true. No productivity studies of individual developers tackling tasks show a 10x speedup; results tend to be anywhere from +20% to minus 20%. We aren't seeing amazing software being built by individual developers using LLMs. There is still only one Fabrice Bellard in the world, even though if your premise could escape the containment zone of imagination anyone should be able to be a Bellard on their own time with the help of LLMs.

          [1] Also, this is basically already true without LLMs. It is the reason startups are able to disrupt corporate behemoths. If you have just a small handful of people who spend the majority of their work time writing code (by hand! No LLMs required!), they can build amazing new products that outcompete products funded by trillion-dollar entities. Your observation of more coding = less meetings required in the first place has an element of truth to it, but not because LLMs are related to it in any particular way.

          • By pishpash 2026-03-063:56

            This isn't the counter you think it is. It's too much to expect existing behemoths to reshape their orgs substantially on a quick enough timeline. The gains will be first seen in new companies and new organizations, and they will be able to stay flat a longer and outcompete the behemoths.

          • By sgc 2026-03-063:571 reply

                 >  Imagine if your grandma had wheels! She'd be a bicycle.
            
            I always took this to be a sharp jab saying the entire village is riding your grandma, giving it a very aggressive undertone. It's pretty funny nonetheless.

            Too early to say what AI brings to the efficiency table I think. In some major things I do it's a 1000x speed up. In others it is more a different way of approaching a problem than a speed up. In yet others, it is a bit of an impediment. It works best when you learn to quickly recognize patterns and whether it will help. I don't know how people who are raised with ai will navigate and leverage it, which is the real long-term question (just as the difference between pre- and post-smartphone generations is a thing).

            • By demorro 2026-03-0613:031 reply

              1000x is ridiculous. What are you doing where that level of improvement is measurable. That means you are doing things that would have taken you a year of full-time work in less than half a day now.

              EDIT: Retracted, I think the example given below is reasonably valid.

              • By sgc 2026-03-0613:481 reply

                I understand, but the improvement is actually more than that. It is not directly programming, but look at this page [1] for example. I spent years handcrafting parallel texts of English and Greek and had managed to put just under 400 books online. With AI, I managed to translate and put in parallel 1500 more books very quickly. At least 2/3 of those have never been translated into English, ever. That means I have done what the entire history of English-speaking scholars has never managed to do. And the quality is good enough that I have already had publishers contacting me to use the translations. There are a couple other areas where I am getting similar speed ups, but of course this is not the norm.

                [1] https://catholiclibrary.org/library/browse/

                • By demorro 2026-03-0614:072 reply

                  ... you know what. Whilst I suspect the quality of these translations is probably not great. Fair play this is a valid example.

                  • By duncangh 2026-03-0814:28

                    This is my favorite type of interaction to see when browsing HN. I feel like this respectful mosh pit of ideas and practitioners is uniquely rad. Thanks for the serendipity dopamine this Sunday

                  • By sgc 2026-03-0614:331 reply

                    Of course they are not perfect, but no translation is even close to perfect. The floor is actually incredibly low. All I can say is that many doctoral-level scholars, including myself and some academic publishers, find them to be somewhere between serviceable and better than average.

                    • By applfanboysbgon 2026-03-0616:591 reply

                      Knowing the quality of LLM translations between the two languages I speak, hearing it used like this by supposed academics invokes a deep despair in me. "Serviceable" is a flimsy excuse for mass-producing and publishing slop. Particularly given that slop will displace efforts to produce human translations, putting a ceiling on humanity's future output - no one will ever aspire to do better than slop, so instead of a few great translations, we'll get more slop than we would ever even want to read.

                      I guess it does depend on the languages involved; one study suggests that it's even worse than Google Translate for some languages, but maybe actually okay at English<-->Spanish?

                      > There were 132 sentences between the two documents. In Spanish, ChatGPT incorrectly translated 3.8% of all sentences, while GT incorrectly translated 18.1% of sentences. In Russian, ChatGPT and GT incorrectly translated 35.6% and 41.6% of all sentences, respectively. In Vietnamese, ChatGPT and GT incorrectly translated 24.2% and 10.6% of sentences, respectively.

                      https://jmai.amegroups.org/article/view/9019/html

                      • By sgc 2026-03-0618:24

                        I wouldn't have put it online if I didn't think it was a major improvement over nothing. Realistically, if we haven't translated it in the last 500 years, there is no point for the next several hundred years of history to stick with nothing as well. It takes a bit more than pasting sentences in chatGPT to get a serviceable translation of course, but significantly better results than that are possible. I have not tried translating into other languages, but I am sure having English as the target language is a help.

                        It's all right there on my website in parallel text, everybody can check and come to their own conclusion rather than driving by with unhelpful generalizations. And really, that is the primary scope of these translations: as aids in reading an original text.

          • By keeda 2026-03-064:392 reply

            > No productivity studies of individual developers tackling tasks show a 10x speedup; results tend to be anywhere from +20% to minus 20%.

            The only study showing a -20% came back and said, "we now think it's +9% - +38%, but we can't prove rigorously because developers don't want to work without AI anymore": https://news.ycombinator.com/item?id=47142078

            Even at the time of the original study, most other rigorous studies showed -5% (for legacy projects, obsolete languages) to 30% (more typical greenfield AND brownfield projects) way back in 2024. Today I hear numbers up to 60% from reports like DX.

            But this is exactly missing the point. Most of them are still doing things the old way, including the very process of writing code. Which brings me to this point:

            > There have been absolutely zero indicators that this is true.

            I could tell you my personal experience, or link various comments on HN, or point you to blogs like https://ghuntley.com/real/ (which also talks about the origanizational impedance mismatch for AI), but actual code would be a better data point.

            So there are some open-source projects worth looking at, but they are typically dismissed because they look so weird to us. Here's two mostly vibe-coded (as in, minimal code review, apparently) projects that people shredded for having weird code, but is already used by 10s of 1000s of people, up to 11 - 18K stars now. Look at the commit volume and patterns for O(300K) LoC in a couple of months, mostly from one guy and his agent:

            https://github.com/steveyegge/beads/graphs/commit-activity

            https://github.com/steveyegge/gastown/graphs/commit-activity

            It's like nothing we've seen before, almost equal number of LoC additions and deletions, in the 100s of Ks! It's still not clear how this will pan out long term, but the volume of code and apparent utility (based purely on popularity) is undeniable.

            • By laserlight 2026-03-0612:022 reply

              > we now think it's +9% - +38%

              If you are referring to the following quote [0], you are off by a sign:

              > we now estimate a speedup of -18% with a confidence interval between -38% and +9%.

              [0] https://metr.org/blog/2026-02-24-uplift-update/

              • By demorro 2026-03-0613:10

                That update blog is funny. The only data they can get at reports slowdowns, but they struggle to believe it because developers self-report amazing speedups.

                You'd get the same sort of results if you were studying the benefits of substance abuse.

                "It is difficult to study the downsides of opiates because none of our participants were willing to go a day without opiates. For this reason, opiates must be really good and we're just missing something."

              • By keeda 2026-03-0617:00

                My bad, I messed up by being lazy while switching from decreases in time taken (that they report) to increased in throughput. (Yes, it's not just flipping the sign, but as I said, I was being lazy!) The broad point still holds, their initial findings have been reversed, and they expect selection effects masked a higher speedup.

                The language is confusing, but the chart helps: https://metr.org/assets/images/uplift-2026-post/uplift_timel...

            • By applfanboysbgon 2026-03-065:181 reply

              > they are typically dismissed because they look so weird to us.

              I dismiss them because Yegge's work (if it can even be called his work, given that he doesn't look at the code) is steaming garbage with zero real-world utility, not "because they look weird". You suggest the apparent utility is undeniable, while saying "based purely on popularity" -- but popularity is in no way a measure of utility. Yegge is a conman who profited hundreds of thousands of dollars shilling a memecoin rugpull tied to these projects. The actual thousands of users are people joining the hypetrain, looking to get in on the promised pyramid scheme of free money where AI will build the next million dollar software for you, if only you have the right combination of .md files to make it work. None of these software are actually materialising, so all the people in this bubble can do is make more AI wrappers that promise to make other AI wrappers that will totally make them money.

              I am completely open to being proven wrong by a vibe-coded open source application that is actually useful, but I haven't seen a single one. Literally not even one. I would count literally anything where the end-product is not an AI wrapper itself, which has tens to hundreds of thousands of users, and which was written entirely by agents. One example of that would be great. Just one. There have been a couple of attempts at a web browser, and Claude's C compiler, but neither are actually useful or have any real users; they are just proofs of concept and I have seen nothing that convinces me they are a solid foundation from which you could actually build useful software from, or that models will ever be on a trajectory to make them actually useful.

              • By keeda 2026-03-0617:431 reply

                The memecoin thing was stupid, totally. Yegge should never have touched it, because well, crypto, but also because that's a distraction from the actual project.

                > popularity is in no way a measure of utility

                Why would it be popular if it's not useful? Yegge is not like some superstar whose products are popular just because he made them. And while some people may be chasing dollars, most of them are building software that scratches an itch. (Search for Beads on GitHub, you'll find thousands of public repos, and lord knows how many private repos.)

                Beads has certainly made my agents much more effective, even the older models. To understand its utility you have to do agentic coding for a while, see the stupid mistakes agents make because they forget everything, and then introduce Beads and see almost all those issues melt away.

                > None of these software are actually materialising

                They are if you look for them. There are many indications (often discussed here) showing spikes in apps on app stores, number of GitHub projects, and Show HN entries. Now, you may dismiss these as "not actually useful", and at this volume that's undoubtedly true for a lot of them.

                But there is already early data showing growth not only in mobile app downloads, but also time spent per user and revenue -- which are pretty clear indications of utility: https://sensortower.com/blog/state-of-mobile-2026

                Edit: it occurs to me that by "vibe-coding" we may be talking about two different things -- I tend to mean "heavily AI-assisted coding" whereas you likely mean "never look at the code YOLO coding." I'll totally agree that YOLO vibe-coded apps by non-experts will be crap. Other than Beads and Gastown I don't know of any such app that is non-trivial. But then those were steered by a highly experienced engineer, and my original point was, vibe-coding correctly could look very weird by today's best practices.

                • By thewebguyd 2026-03-0620:311 reply

                  > I tend to mean "heavily AI-assisted coding" whereas you likely mean "never look at the code YOLO coding."

                  The original point that sparked this sub-thread though is that AI is being overhyped. If actual vibe coding (YOLO it, never look at or understand the code, thus truly enabling non-technical folk to have revolutionary power and ability) doesn't work, then AI is yet just another tool in the toolbelt like any other developer life enhancing tech we've had so far, it's just a new form of IDE.

                  Being a new form of IDE, while very useful, isn't exactly entire economy transforming revolutionary tech. If it can't be used by someone with zero computer/eng experience to build something useful and revenue generating, the amount of investment we've seen into it is way overblown and is well overdue for a pretty severe correction.

                  I buy AI as a "developer enhancing tool" just like any other devtools that we've seen over my career. I don't currently buy it as a "total labor economy transformation force."

                  • By keeda 2026-03-0719:21

                    But that's the thing... I don't YOLO it, I review all the code it generates, yet I personally am doing something -- using computer vision, a field I had negligible prior background in -- that would have taken a team of 2 - 3, AND about a semester or year of an advanced degree to catch up on the field! I am effectively doing the work of a team.

                    So when I see people online say they feel 10x productive, I tend to believe them. But I'm working solo, and have none of the encumbrances and "Conway Overhead" of coordinating with a lot of other people, so I also understand why the overall effect is so limited. Which is why I think current companies are "shaped wrong" for AI.

                    When companies eventually adapt, it will be a "labor economy transformation force" because the same dynamics will play out across all knowledge work. And I am not talking as an AI booster, but as a parent whose kids are interested in software engineering; I have every incentive to hope my prognostications do not come true, but I prefer being prepared for the worst.

        • By sdf2df 2026-03-063:161 reply

          What a load of fluff lmao. Are you Nadella?

          • By keeda 2026-03-065:08

            Hah! I would say I'm flattered, but I find his style of speaking rather stilted.

      • By tayo42 2026-03-065:471 reply

        Ops hasn't been in the crosshairs of Ai yet.

        Imo it's only a matter of time as companies start to figure out how to use ai. Companies don't seem to have real plans yet and everyone is figuring out ai in general out.

        Soon though I will think agents start popping up, things like first line response to pages, executing automation

        • By bandrami 2026-03-066:051 reply

          We've had deterministic automation of tier one response for over a decade now. What value would indeterminacy add to that?

          • By tayo42 2026-03-066:131 reply

            To deal with the problems where there is ambiguity in the problem and the approach to solving it. Not everything is a basic decision tree. Humans aren't deterministic either, the way we woukd approach a problem is probably different. Is one of us right or wrong? We're generally just focused on end results.

            Maybe 2 years ago Ai was doing random stuff and we got all those funny screenshots of dumb gemini answers. The indeterminism leading to random stuff isn't really an issue any more.

            The way it thinks keeps it on track.

            • By bandrami 2026-03-067:481 reply

              Two weeks ago I asked a frontier model to list five mammals without "e" in their name and number four was "otter"

              • By tayo42 2026-03-0615:431 reply

                Is identifying mammals without the letter E part of your ops work flow?

                Opus 4.6 didn't have an issue with this question though.

                • By thewebguyd 2026-03-0617:181 reply

                  > Is identifying mammals without the letter E part of your ops work flow?

                  No, but it can show unreliability for adjacent tasks. Identifying a CIDR block in traffic logs is a normal part of an ops work flow. It means it's more likely to fail if you need to generate a complex Regex to filter PII from a terabyte of logs. If the model has a blind spot for specific characters because it tokenizes words instead of seeing individual characters, then it can miss a critical path of failure because the service name didn't fit its probabilistic training.

                  Maybe you need to boilerplate Terraform. If the model can't reliably (reliably, as in, 100% deterministic, does this without fail) parse constraints, it's not just a funny mistake it's a potential 5 figure billing error.

                  Ops can't run on "mostly accurate." That's just simply not good enough. We need deterministic precision.

                  For AI to be useful in this world to the extent others have claimed it is for software eng, we'll likely need more advanced world models, not just something that can predict the next most likely token.

                  • By tayo42 2026-03-0618:402 reply

                    Your terraform written by a person already doesn't have deterministic precision. Ai isn't messing these things up either.

                    If your Ai work flow is still dumping logs into a chat and saying search it for some pattern, then you should see what something like Claude code approaches problems. These agents aren't building scripts to solve problems. Which is your deterministic solution.

                    • By thewebguyd 2026-03-0619:161 reply

                      That still only just makes it a force multiplier for engineers, like any other tech, not a replacement as it's being hyped and sold as.

                      Claude resorting to writing code for everything, because that's all the model can do without too many hallucinations and context poisoning, is just a higher speed REPL. Great, that's useful.

                      But that's not what is being hyped and sold. What's being hyped and sold is "You don't need an Ops guy anymore, just talk to the computer." Well, what happens when the AI decides the "fix" is to just open up 0.0.0.0/0 to the world to make the errors go away? The non technical minimum wage person now just talking to the computer has no idea they just pwned the company.

                      If AI's answer is "Just write a script to solve the prompt" then you still need technical people, and it's vasly over hyped.

                      I'll be interested when you actually can just dump logs in a chat and analyze it without the model having to resort to writing code to solve the problem. That will be revolutionary. Imagine all the time I'd save by not having to make business reports, I can just tell the business people to point AI at terabytes of CSV exports and just ask it questions. That is when it will stop just being labor compression for existing engineers, and start being a world changing paradigm shift.

                      For now, it's just yet another tool in my toolbelt.

                      • By tayo42 2026-03-0622:121 reply

                        Not sure why the implementation is important or not. The point is the system will be triggered by some text input and complete the task asynchronously on its own.

                        • By bandrami 2026-03-072:29

                          Right but our company takes on interns not to help the company but to help the intern

                    • By bandrami 2026-03-0621:341 reply

                      > Your terraform written by a person already doesn't have deterministic precision

                      Can you expand on that? Because it sure seems to me like it is in fact deterministic unless the person deliberately made it otherwise

                      • By tayo42 2026-03-0622:11

                        If i give you a task to write terraform or any code, you won't write what I write, you probably won't even write the same thing twice. You can introduce a bug too, we're not perfect. The output of the task "write some terraform" already isn't deterministic when dealing with people.

      • By sdf2df 2026-03-060:23

        Youre not missing anything.

        Humans are funny. But most cant seem to understand that the tool is a mirage and they are putting false expectations on it. E.g. management of firms cutting back on hiring under the expectation that LLMs will do magic - with many cheering 'this is the worst itll be bro!!".

        I just hope more people realise before Anthropic and OAI can IPO. I would wager they are in the process of cleaning up their financials for it.

    • By httpz 2026-03-064:003 reply

      This is a classic case of Productivity Paradox when personal computers were first introduced into workplaces in the 80s.

      A famous economist once said, "You can see the computer age everywhere but in the productivity statistics."

      There are many reasons for the lag in productivity gain but it certainly will come.

      https://en.wikipedia.org/wiki/Productivity_paradox

      • By bandrami 2026-03-064:072 reply

        That's only certain if investments in tech infrastructure always led to productivity increases. But sometimes they just don't. Lots of firms spent a lot of money on blockchain five years ago, for instance, and that money is just gone now.

        • By 20k 2026-03-064:332 reply

          I find it odd the universal assumption that AI is going to be good for productivity

          The loss of skills, complete loss of visibility and experience with the codebase, and the complete lack of software architecture design, seems like a massive killer in the long term

          I have a feeling that we're going to see productivity with AI drop through the floor

          • By hombre_fatal 2026-03-065:082 reply

            I'd claim the opposite. Better models design better software, and quickly better software than what most software developers were writing.

            Just yesterday I asked Opus 4.6 what I could do to make an old macOS AppKit project more testable, too lazy to even encumber the question with my own preferences like I usually do, and it pitched a refactor into Elm architecture. And then it did the refactor while I took a piss.

            The idea that AI writes bad software or can't improve existing software in substantial ways is really outdated. Just consider how most human-written software is untested despite everyone agreeing testing is a good idea simply because test-friendly arch takes a lot of thought and test maintenance slow you down. AI will do all of that, just mention something about 'testability' in AGENTS.md.

            • By bandrami 2026-03-065:152 reply

              OK so this comes back to the question I started this subthread with: where is this better software? Why isn't someone selling it to me? I've been told for a year it's coming any day now (though invariably the next month I'm told last month's tools were in fact crap and useless compared to the new generation so I just have to wait for this round to kick in) and at some point I do have to actually see it if you expect me to believe it's real.

              • By hombre_fatal 2026-03-065:312 reply

                How would you know if all software written in the last six months shipped X% faster and was Y% better?

                Why would you think you have your finger on the pulse of general software trends like that when you use the same, what, dozen apps every week?

                Just looking at my own productivity, as mere sideprojects this month, I've shipped my own terminal app (replaced iTerm2), btrfs+luks NAS system manager, overhauled my macOS gamepad mapper for the app store, and more. All fully tested and really polished, yet I didn't write any code by hand. I would have done none of that this month without AI.

                You'd need some real empirics to pick up productivity stories like mine across the software world, not vibes.

                • By Tanjreeve 2026-03-065:46

                  It's on the people pushing AI as the panacea that has changed things to show workings. Not someone saying "I've not seen evidence of it". Otherwise it's "vibes" as you put it.

                • By bandrami 2026-03-065:472 reply

                  Right, I'm sympathetic to the idea that LLMs facilitate the creation of software that people previously weren't willing to pay for, but then kind of by definition that's not going to have a big topline economic impact.

                  • By littlexsparkee 2026-03-0622:07

                    Well, we don't know - that's capturing 2 scenarios: software that whose impact is low as reflected by lack of investment and legitimately useful improvements that just weren't valued (fix slow code, reduce errors and increase uptime, address security concerns) because the cost was not appreciated / papered over by patches / company hasn't been bitten yet

                  • By hombre_fatal 2026-03-0613:47

                    Why did you add that "weren't willing to pay for" condition?

                    Most of the software I replaced was software I was paying for (iStat Menus, Wispr Flow, Synology/Unraid). That I was paying for a project I could trivially take on with AI was one of the main incentives to do it.

              • By eucyclos 2026-03-066:251 reply

                Here's an example: https://eudaimonia-project.netlify.app/

                I'm happy to sell it to you, though it is also free. I guided Claude to write this in three weeks, after never having written a line of JavaScript or set up a server before. I'm sure a better JavaScript programmer than I could do this in three weeks, but there's no way I could. I just had a cool idea for making advertising a force for good, and now I have a working version in beta.

                I'd say it is better software, but better is doing a lot of heavy lifting there. Claude's execution is average and always will be, that's a function of being a prediction engine. But I genuinely think the idea is better than how advertising works today, and this product would not exist at all if I had to write it myself. And I'm someone who has written code before, enough that I was probably a somewhat early adopter to this whole thing. Multiply that by all the people whose ideas get to live now, and I'm sure some ideas will prove to be better even with average execution. Like an llm, that's a function of statistics.

                • By bandrami 2026-03-066:341 reply

                  In glad you made something with it you wanted to make, and as a fan of Aristotle I'm always happy to see the word eudaimonia out there. Best of luck. That said I don't understand what this does or why I would want the tokens it mentions.

                  • By eucyclos 2026-03-066:402 reply

                    Yeah, I gotta make a video walkthrough. Its basically a goal tracker combined with an ad filter - write what you want out of life and block ads, it replaces them with ads that actually align with your long term goals instead of distracting from them. The tokens let you add ads to the network, though you also get some for using the goal tracker.

                    • By bandrami 2026-03-067:38

                      Though this does suggest one possible answer to me: the new software is largely web applications, and the web is just a space I don't spend much time anymore other than a few retro sites like this

                    • By tasuki 2026-03-0613:013 reply

                      No, you don't need a video walkthrough. You need that damn web page to explain – in plain language – what this is and what it's good for.

                      • By demorro 2026-03-0613:40

                        They can't, they never did the work to discover what it's good for because they skipped over implementation and concept validation.

                        This concept will never work outside of their own head. People continue to think producing something is the hard part my word.

                      • By eucyclos 2026-03-0613:451 reply

                        Would the above explanation be better? The website is there because stripe needs a landing page and the text is there because I'm trying to communicate the aspiration the instantiation I can always explain in detail if someone wants to hear how that would work.

                        • By tasuki 2026-03-0620:451 reply

                          > Would the above explanation be better?

                          No idea. I certainly didn't get it. Goal tracker is one thing, ad blocker is another thing. Why would I want to combine them? And why would I want to see any ads at all? Perhaps I'm just not the target audience...

                          • By eucyclos 2026-03-0623:53

                            Maybe not, but you might want to see ads because 1) they fund a huge part of the free internet so you would at least want other people to see them and 2) if they were targeted not at what you're most likely to buy today but at what would most help you achieve goals you'r struggling with, they'd be a constant source of useful information and motivation as you go about your day. Aligning incentives between you and advertisers turns ads from friction to tailwind, and advertisers already want to align with what incentivises you if the alternative is having their ads blocked.

                            That second point is the part that seems obvious to me but I have a hard time communicating.

                      • By KellyCriterion 2026-03-0615:331 reply

                        ++1

                        I didnt get it either on first glance when scrolling down the whole page

                        • By eucyclos 2026-03-0619:40

                          Wow, that us useful feedback, thanks! I'll update that this weekend.

            • By 20k 2026-03-068:272 reply

              And now you have no idea how any of the code works

              AI writes bad software by virtue of it being written by the AI, not you. No actual team member understands what's going on with the code. You can't interrogate the AI for its decision making. It doesn't understand the architecture its built. There's nobody you can ask about why anything is built the way it is - it just exists

              Its interesting watching people forget that the #1 most important thing is developers who understand a codebase thoroughly. Institutional knowledge is absolutely key to maintaining a codebase, and making good decisions in the long term

              Its always been possible to trade long term productivity for short term gains like this. But now you simply have no idea what's going on in your code, which is an absolute nightmare for long term productivity

              • By hombre_fatal 2026-03-0617:291 reply

                You can read as much or as little of the code as you want.

                The status quo was that I have no better understanding of code I haven't touched in a year, or code built by other people. Now I have the option to query the code with AI to bootstrap my understanding to exactly the level necessary.

                But you're wrong on every claim about LLM capabilities. You can ask the AI exactly why it decided on a given design. You can ask it what the best options were and why it chose that option. You can ask it for the trade-offs.

                In fact, this should be part of your Plan feedback loop before you move to Implementation.

                • By 20k 2026-03-0618:39

                  You can ask the AI why, but its answer doesn't come from any kind of genuine reasoning. It doesn't know why it did anything, because it doesn't exist as a sentient being. It just makes something up that sounds good

                  If you choose to take AI reasoning at face value, you're choosing to accept pretty strong technical debt

              • By mirsadm 2026-03-0611:121 reply

                My own observation is that the initial boost to productivity results in massive crippling technical debt.

                • By davidvartanian 2026-03-0823:19

                  That's just because everyone is misusing AI. If you ask AI to do a job and you have no idea what it did, you lost ownership, which means you're asking to be replaced. You need to own the task. If you fully delegate your task to anyone else or to AI, you no longer know what's going on. AI does not necessarily produce more tech debt, but AI might do things you don't expect because it lacks context and specificity to perform accurately.

          • By nikkwong 2026-03-064:471 reply

            Having the productivity "drop through the floor" is a bit hyperbolic, no? Humans are still reviewing the PRs before code merge at least at my company (for the most part, for now).

            • By bandrami 2026-03-065:002 reply

              I don't know that it's likely but it's certainly a plausible outcome. If tooling keeps getting built for this and the financial music stops it's going to take a while for everybody to get back up to speed

              Remember this famously happened before, in the 1970s

              • By Tanjreeve 2026-03-065:501 reply

                There's an actual working product now, albeit one which is currently loss leading. In software world at least there is definitely enough value for it to be used even if it's just better search engine. I'm not sure why it would disappear if the financial music stops as opposed to being commoditised.

                • By bandrami 2026-03-066:04

                  Because there's cheaper ways to get an equally good search engine? But yes I imagine some amount of inference will continue even in an AI Winter 3.0 scenario.

        • By salawat 2026-03-066:111 reply

          Ironically, abstraction bloat eats away any infra gains. We trade more compute to allow people less in tune with the machine to get things done, usually at the cost of the implementation being eh... Suboptimal, shall we say.

          • By bandrami 2026-03-067:40

            I think there's a broad category error where people see that every gain has been an abstraction (true) but conclude from that that every abstraction will be a gain (dubious)

      • By danbolt 2026-03-069:01

        My unfounded hunch for the computing bit is that home computers became more and more commonplace in the home as we approached the 21st century.

        A Commodore 64 was a cool gadget, but “the family computer” became a device that commoditized the productivity. The opportunity cost of applying a computer to try something new went to near zero.

        It might have been harder for someone to improve the productivity of an old factory in Shreveport, Louisiana with a computer than it was for the upstarts at id to make Doom.

      • By kranner 2026-03-064:441 reply

        > There are many reasons for the lag in productivity gain but it certainly will come.

        Predictions without a deadline are unfalsifiable.

        • By KellyCriterion 2026-03-0615:34

          Well the thing with predictions is that they are in genral difficult - esp. when it comes to those in future :-D

    • By fnordpiglet 2026-03-064:402 reply

      My employer is pretty advanced in its use of these tools for development and it’s absolutely accelerated everything we do to the point we are exhausting roadmaps for six months in a few weeks. However I think very few companies are operating like this yet. It takes time for tools and techniques to make it out and Claude code alone isn’t enough. They are basically planning to let go of most of the product managers and Eng managers, and I expect they’re measuring who is using the AI tools most effectively and everyone else will be let go, likely before years end. Unlike prior iterations I saw at Salesforce this time I am convinced they’re actually going to do it and pull it off. This is the biggest change I’ve seen in my 35 year career, and I have to say I’m pretty excited to be going through it even though the collateral damage will be immense to peoples lives. I plan to retire after this as well, I think this part is sort of interesting but I can see clearly what comes next is not.

      • By p1esk 2026-03-065:20

        I’m observing very similar trends at a startup I’m at. Unfortunately I’m not ready to retire yet.

      • By blackcatsec 2026-03-065:442 reply

        Why are you excited for this? They’re not going to give YOU those peoples’ salaries. You will get none of it. In fact, it will drag your salary through the floor because of all the available talent.

        • By fnordpiglet 2026-03-0616:49

          I’m excited as a computer scientist to see it happening in my life time. I am not excited for the consequences once it’s played out. Hence my comment about retiring, and empathy for everyone who is still around once I do. I never got into this for the money - when I started engineers made about as much as accountants. It’s only post 1997 or so that it became “cool” and well paid. I am doing this because I love technology and what it can do and the science of computing. So in that regard it’s an amazing time to be here. But I am also sad to see the black box cover the beauty of it all.

        • By Karrot_Kream 2026-03-069:212 reply

          I'm very confused about this. Salary is only one portion of your total compensation. The vast majority of tech companies offer equity in a company. The two ways to increase the FMV of your equity is: increase your equity stake or increase the value of the total equity available. Hitting the same goals with fewer people means your run rate is lower, which increases the value of your equity (the FMV prices in lower COGS for the same revenue.) Also, keeping on staff often means you want to offer them increased equity stakes as an employment package. Letting staff go means more of that available equity pool is available to distribute to remaining employees.

          We aren't fungible workers in a low skill industry. And if you find yourself working in a tech company without equity: just don't, leave. Either find a new tech company or do something else altogether.

          • By camdenreslink 2026-03-0617:262 reply

            Equity is negotiable just like salary, and if supply of developer labor increases with the same or less demand, you'll get less equity just like you will get less salary.

            • By blackcatsec 2026-03-0723:41

              I can't believe the person you replied to thinks that they're going to get some magical more amount of equity because you can hopefully do more with fewer people. That's assuming the entire business landscape doesn't also change with AI, disincentivizing so much investment in companies in the first place because someone else with AI can create a competitor in a shorter amount of time...

            • By fnordpiglet 2026-03-099:25

              They’re also betting they’re the P99 engineer. Most do. 98% aren’t.

          • By p1esk 2026-03-0715:44

            In the last three startups I worked at I didn’t bother exercising my vested equity - even a successful exit would at best triple the price of those shares - not worth the risk. One of those three startups already failed.

    • By bandrami 2026-03-060:192 reply

      The dev team is committing more than they used to. A lot, in fact, judging from the logs. But it's not showing up as a faster cadence of getting me software to administer. Again, maybe that will change.

      • By righthand 2026-03-060:211 reply

        In my experience it is now twice the amount of merge requests as a follow-up appears to correct any bugs no one reviewed in the first merge request.

        • By silentkat 2026-03-063:042 reply

          I’m at a big tech company. They proudly stated more productivity measures in commits (already nonsense). 47% more commits, 17% less time per commit. Meaning 128% more time spent coding. Burning us out and acting like the AI slop is “unlocking” productivity.

          There’s some neat stuff, don’t get me wrong. But every additional tool so far has started strong but then always falls over. Always.

          Right now there’s this “orchestrator” nonsense. Cool in principle, but as someone who made scripts to automate with all the time before it’s not impressive. Spent $200 to automate doing some bug finding and fixing. It found and fixed the easy stuff (still pretty neat), and then “partially verified” it fixed the other stuff.

          The “partial verification” was it justifying why it was okay it was broken.

          The company has mandated we use this technology. I have an “AI Native” rating. We’re being told to put out at least 28 commits a month. It’s nonsense.

          They’re letting me play with an expensive, super-high-level, probabilistic language. So I’m having a lot of fun. But I’m not going to lie, I’m very disappointed. Got this job a year ago. 12 years programming experience. First big tech job. Was hoping to learn a lot. Know my use of data to prioritize work could be better. Was sold on their use of data. I’m sure some teams here use data really well, but I’m just not impressed.

          And I’m not even getting into the people gaming the metrics to look good while actually making more work for everyone else.

          • By booleandilemma 2026-03-065:231 reply

            Management is just stupid sometimes. We had a similar metric at my last company and my manager's response was "well how else are we supposed to measure productivity?", and that was supposed to be a legitimate answer.

            • By sdf2df 2026-03-0618:17

              The benefits of AI either accrue toward incremental revenue-generation or cost-saving.

              Its not rocket science to measure actually. The issue is most people dont know how to think properly to invent the right proxies.

          • By sdf2df 2026-03-063:17

            Lol its gonna take longer than it should for this to play out.

            Sunk cost fallacy is very real, for all involved. Especially the model producers and their investors.

            Sunk cost fallacy is also real for dev's who are now giving up how they used to work - they've made a sunk investment in learning to use LLMs etc. Hence the 'there's no going back' comments that crop up on here.

            As I said in this thread - anyone who can think straight - Im referring to those who adhere to fundamental economic principles - can see what's going on from a mile away.

      • By whateveracct 2026-03-063:26

        I think they feel more productive but aren't actually.

    • By steve_adams_86 2026-03-0617:32

      In my org I get far more done than ever, but I also find it more exhausting.

      Because I can get so much done, I've lost my sense for what's enough. And if I can squeeze out a bit more relatively easily, why wouldn't I? When do I hit the brakes?

      There are some tasks where LLMs are not all that helpful, and I find myself kind of savoring those tasks.

      I'm surprised you don't notice a difference. Where I work it has been transformative. Perhaps it's because we're relatively small and scrappy, so the change in pace is easier with less organizational inertia. We've dramatically changed processes and increased outputs without a loss in quality. For less experienced programmers who are more interested in simple scripts for processing data, their outputs are actually far better, and they're learning faster because the Claude Code UI exposes them to so many techniques in the shell. I now see people using bash pipes for basic operations who wouldn't have known a thing about bash a couple years ago. The other day a couple less-technical people came to me to learn about what tests are. They never would have been motivated to learn this before. It's really cool.

      It doesn't reduce work at all, though. We're an under-funded NGO with high ambition. These changes allow us to do more with the same funding. Hopefully it allows us to get more funding, too. I can't see it leading to anyone being let go; we need every brain we can get.

    • By eucyclos 2026-03-064:521 reply

      A tool with a mediocre level of skill in everything looks mediocre when the backdrop is our own area of expertise and game changing when the backdrop is an unfamiliar one. But I suspect the real game changer will be that everyone is suddenly a polymath.

      • By sibeliuss 2026-03-065:13

        This ^ Exactly it. This will be the change.

    • By redhale 2026-03-0611:512 reply

      I don't doubt your sincerity. But this represents an absolutely bonkers disparity compared to the reality I'm experiencing.

      I'm not sure what to say. It's like someone claiming that automobiles don't improve personal mobility. There are a lot of logical reasons to be against the mass adoption of automobiles, but "lack of effectiveness as a form of personal mobility" is not one of them.

      Hearing things like this does give me a little hope though, as I think it means the total collapse of the software engineering industry is probably still a few years away, if so many companies are still so far behind the curve.

      • By tasuki 2026-03-0612:48

        > It's like someone claiming that automobiles don't improve personal mobility.

        I prefer walking or cycling and often walk about 8km a day around town, for both mobility and exercise. (Other people's) automobiles make my experience worse, not better.

        I'm sure there's an analogy somewhere.

        (Sure, automobiles improve the speed of mobility, if that's the only thing you care about...)

      • By bandrami 2026-03-075:04

        I don't think I'm asking for something unreasonable: I'll believe this actually speeds up software creation when one of my vendors starts getting me software faster. That's not some crazy ludditism on my part, I don't think?

    • By lovich 2026-03-062:472 reply

      > so far AI has had almost no impact on my job.

      Are you hiring?

      • By cute_boi 2026-03-063:592 reply

        My friend used to say that, and he got quietly fired and outsourced because now someone in India can use ChatGPT to produce similar code, lol.

        IMO AI will make 70-80% job obsolete for sure.

        • By bandrami 2026-03-064:41

          But, as I said above, I don't produce code; I administer it (administrate? whichever it is).

        • By leptons 2026-03-068:07

          >now someone in India can use ChatGPT to produce similar code,

          lol, that sounds like a disaster for the codebase.

      • By LPisGood 2026-03-064:38

        My company has been hiring a ton over the last year or so. Jobs are out there

    • By Kye 2026-03-060:50

      I've taken to calling LLMs processors. A "Hello World" in assembly is about 20 lines and on par with most unskilled prompting. It took a while to get from there to Rust, or Firefox, or 1T parameter transformers running on powerful vector processors. We're a notch past Hello World with this processor.

      The specific way it applies to your specific situation, if it exists, either hasn't been found or hasn't made its way to you. It really is early days.

    • By ygouzerh 2026-03-0612:47

      I feel that it differs a lot between companies. It seems like corporate are having less an impact for now, as external innovation needs tailoring to adapt to its needs (e.g a security solution that needs 3 month projects to be tailored to the company tech stack), whereas startups and smaller firms see the most of the impact so far.

    • By willmadden 2026-03-060:593 reply

      Build a new feature. If you aren't bogged down in bureaucracy it will happen much faster.

      • By YesBox 2026-03-065:012 reply

        I dont use LLMs much. When I do, the experience always feels like search 2.0. Information at your fingertips. But you need to know exactly what you're looking for to get exactly what you need. The more complicated the problem, the more fractal / divergent outcomes there are. (Im forming the opinion that this is going to be the real limitations of LLMs).

        I recently used copilot.com to help solve a tricky problem for me (which uses GPT 5.1):

           I have an arbitrary width rectangle that needs to be broken into smaller 
           random width rectangles (maintaining depth) within a given min/max range. 
        
        The first solution merged the remainder (if less than min) into the last rectangle created (regardless if it exceeded the max).

        So I poked the machine.

        The next result used dynamic programming and generated every possible output combination. With a sufficiently large (yet small) rectangle, this is a factorial explosion and stalled the software.

        So I poked the machine.

        I realized this problem was essentially finding the distinct multisets of numbers that sum to some value. The next result used dynamic programming and only calculated the distinct sets (order is ignored). That way I could choose a random width from the set and then remove that value. (The LLM did not suggest this). However, even this was slow with a large enough rectangle.

        So I poked my brain.

        I realized I could start off with a greedy solution: Choose a random width within range, subtract from remaining width. Once remaining width is small enough, use dynamic programming. Then I had to handle the edges cases (no sets, when it's okay to break the rules.. etc)

        So the LLMs are useful, but this took 2-3 hours IIRC (thinking, implementation, testing in an environment). Pretty sure I would have landed on a solution within the same time frame. Probably greedy with back tracking to force-fit the output.

        • By gilbetron 2026-03-0615:051 reply

          I just tested this with Claude Code and Opus 4.6, with the following prompt:

          "I have an arbitrary width rectangle that needs to be broken into smaller random width rectangles (maintaining depth) within a given min/max range. The solution needs to be highly performant from an algorithmic standpoint, well-tested using TDD and Red/Green testing, written in python, and not have any subtle errors."

          It got the answer you ended up with (if I'm understanding you correctly) the first time in just over 2 minutes of working, and included a solid test suite examining edge cases and with input validation.

          • By YesBox 2026-03-0615:371 reply

            How can we verify if you dont post the code?

            I appreciate you testing, even though it's not a great comparison:

            - My feedback cycle of LLM prompting forced me to be more explicit with each call, which benefited your prompt since I gave you exactly what to look for with fewer nuances.

            - Maybe GPT 5.1 is old or kneecapped for newer versions of GPT

            - Maybe Opus/Claud is just a way better model :P

            Please post the code!

            Edit: Regarding "exactly what to look for", when solving a new problem, rarely is all the nuance available for the first iteration.

            • By gilbetron 2026-03-071:501 reply

              I didn't prompt anything odd, just standard prompt "etiquette", actually I significantly prompted less than I would usually do, trying to do a simple prompt like you did.

              • By YesBox 2026-03-075:40

                I’ll believe it when I see the code

        • By redhale 2026-03-0611:351 reply

          > I don't use LLMs much

          Sorry to be so blunt, but it's not surprising that you aren't able to get much value from these tools, considering you don't use them much.

          Getting value from LLMs / agents is a skill like any other. If you don't practice it deliberately, you will likely be bad at it. It would be a mistake to confuse lack of personal skill for lack of tool capability. But I see people make this mistake all the time.

          • By YesBox 2026-03-0613:311 reply

            Would be helpful if you pointed out what I did wrong :).

            If it's "you didn't explain the problem clearly enough", then that aligns with my original comment.

            • By windward 2026-03-0616:17

              If you ask the chatbot for best practices it will tell you, including that you don't use a chatbot.

      • By bandrami 2026-03-061:011 reply

        Most of these are new features, but then they have to integrate with the existing software so it's not really greenfield. (Not to mention that our clients aren't getting any faster at approving new features, either.)

        • By willmadden 2026-03-061:531 reply

          Did you train a self-hosted/open source LLM on your existing software and documentation? That should make it far more useful. It's not claude code, but some of those models are 80% there. In 6 months they'll be today's claude code.

          • By bandrami 2026-03-062:231 reply

            What would that help us with?

            • By willmadden 2026-03-0616:121 reply

              The LLM needs to understand your existing codebase if it's going to be useful building features that integrate with said codebase seamlessly without breaking things or assuming things that don't exist. That's not something you want to give away to a private AI company, so self-host an open source model.

              • By bandrami 2026-03-072:301 reply

                I understood how that would help the LLM, I asked how that would help us. What decision would that inform for us that we're currently having trouble with?

                • By willmadden 2026-03-112:39

                  What "decision" are you talking about? You said you don't code, but you administer and maintain code. It sounds like you integrate new features into an existing codebase?

      • By sdf2df 2026-03-061:40

        Its this kind of thinking that tells me people cant be trusted with their comments on here re. "Omg I can produce code faster and it'll do this and that".

        No simply 'producing a feature' aint it bud. That's one piece of the puzzle.

    • By rhubarbtree 2026-03-0612:022 reply

      This is not a good thing. If you’re not being exposed and skilling up already, you’re likely to be in the camp that is washed away.

      If you can’t be exposed to it in your day job, start using Claude opus in the evening so you know what’s coming.

      • By matkoniecz 2026-03-0612:451 reply

        So far I have not seen much skill gain from using LLM extensively.

        Maybe I will be replaced by matrix multiplication in my job, but if I need to use LLM at some point I expect little benefit from starting now.

        Yes, I tried to use Claude Code two months ago. It was scary, but not useful.

        • By rhubarbtree 2026-03-0613:191 reply

          “Not useful” —- one of those moments where you have to be able to adjust your views in the face of new evidence. Humans are so wedded to their beliefs that it can be agonising to let go. I have nothing but respect for people who admit they were wrong, though. I remained a skeptic for a long time, but 4.5 was enough to convince me to adopt for production code.

          • By matkoniecz 2026-03-0613:29

            So far I updated from "meh, not useful" to "scary, not useful".

            If it would be useful I would continue to use it, but at this point I would not use even if it would be free, not proprietary and not funding replacing me.

      • By bandrami 2026-03-070:551 reply

        Eh, people have been warning me "you'll be left behind!" about the flavor of the week for decades now and it hasn't happened yet. If it happens it happens.

        • By rhubarbtree 2026-03-107:18

          I fear you have not yet fully appreciated the magnitude of this particular change.

    • By sdf2df 2026-03-060:221 reply

      I will personally say right now... its not gonna change lol.

      People who actually know how to think can see it a mile away.

      • By stevenhuang 2026-03-063:351 reply

        It's telling you feel the need to create a throw away to voice this opinion.

        • By sdf2df 2026-03-063:372 reply

          1) Not a throaway, can't remember what my old account is called 2) Feel free to screen shot. Stick it on your desktop and set a reminder and check the state of the world in 12 months time.

          Job done fella.

          • By jaxn 2026-03-064:482 reply

            For some of us, the world has already changed drastically. I am shipping more code, better code, less buggy code WAY faster than ever before. Big systemic changes for the better to our infra as well. There are days where I easily do 2 weeks worth of my best work ever.

            I totally understand that not everyone is having that experience. And yet until people live it, it seems they just discount the experience others are having.

            I'll take the 12 month bet.

            • By leptons 2026-03-068:111 reply

              >I am shipping more code, better code, less buggy code WAY faster than ever before.

              It's clearly relative. For all we know you're a crap coder and AI is now your crutch. We have no evidence that with AI you are as good as an average developer with a fair amount of experience. And even if you do have a fair amount of experience, that doesn't mean you're a good coder.

              • By sdf2df 2026-03-0618:14

                Exactly lol.

                The iPod project was done in months, not years. Im convinced most people aren't as good at programming / focusing on the right stuff as they claim.

            • By salawat 2026-03-066:36

              Cool, and you're doing it on top of the single largest IP hijacking in the history of the world, a massive uptick in infra spend and energy burn to "just throw more compute" at it instead of figuring out how to throw "the right compute at it", cannibalization of the onboarding graduates, and losing having enough friction to keep you from running off after what's probably a bad idea on further analysis, because you can crank this out in a weekend. Last time somewhat did that, we got fucking JS. We still haven't rid ourselves of it.

              Let us not lose sight of how we got here.

          • By stevenhuang 2026-03-063:531 reply

            12 months I won't be surprised if there's not much change. But in 5 years? 10? Anything can happen. It is presumptuous to think you can project the future capabilities of this technology and confidently state that labour markets will never be affected.

            • By sdf2df 2026-03-064:094 reply

              You prove my point.

              Guys like you dont get it. You think OAI, Amazon etc can freely put large amounts of money into this for 5-10 years? Lmao - delusional. Investors are impatient. Show huge jumps in revenue this year or you no longer have permission to put monumental amounts of money into this anymore.

              Short of that they'll just destroy the stock price by selling off; leaving employees who get paid via SBC very unhappy.

              • By stevenhuang 2026-03-0618:56

                Whether investors will see returns soon enough to service their debt loads is entirely another matter. I do agree the likely course of action is we get a crash of sorts, since the only way their investments pan out is if labour is replaced entirely which of course sounds unlikely in near term.

                My point is the cat is out of the bag. It doesn't take massive investments to achieve iterative improvements on SOTA. As long as the technology does not plateau, smaller labs have shown it's possible to advance the frontiers independent of large companies/investments. And as these frontiers advance, more and more of economical knowledge work will be subsumed by AI. I don't see a way out of this, which is why I am a strong proponent of wealth distribution eg UBI.

              • By dolebirchwood 2026-03-064:491 reply

                > You think OAI, Amazon etc can freely put large amounts of money into this for 5-10 years?

                Won't matter. The Chinese models will be running on potatoes by then and be better than ever.

                • By sdf2df 2026-03-0618:25

                  By the time that's obvious investors in the market would've priced that in. Again repeating myself here.

              • By HWR_14 2026-03-066:461 reply

                Whatever you want to say about other companies, Amazon (and Meta) is quite willing to spend many years pouring billions into technology they think will pay off later.

                • By Ekaros 2026-03-068:552 reply

                  Looking at VR and Meta. They absolutely can be wrong. So even after investing what seems to be enough, there might not be any payoff.

                  • By HWR_14 2026-03-0823:27

                    There may be no payoff in the end, but it won't be for lack of resources thrown at the problem or investor impatience.

                  • By sdf2df 2026-03-0618:101 reply

                    And the investors were correct to crush the stock price down to 90-odd dollars. Which finally forced Zuck to face the music.

                    This place is full of bozos.

                    • By HWR_14 2026-03-0823:28

                      Did Zuck change because the stock price went down or he was dissatisfied with the lack of progress/adoption?

              • By greyw 2026-03-064:431 reply

                Such are reductive and superficial way of thinking on how investments works. Makes me confident you dont really are able to make a good prediction

                • By sdf2df 2026-03-0618:11

                  Lol okay, show me your portfolio. Ive beaten the market after-adjusting for risk for years on-end.

HackerNews