It's not a bad paper, but it's also turning into a fantastic illustration of how much thirst there is out there for anything that shows that AI productivity doesn't work.
> It's not a bad paper, but it's also turning into a fantastic illustration of how much thirst there is out there for anything that shows that AI productivity doesn't work.
Maybe. I think it's a fantastic illustration of anyone doing anything to provide something other than hype around the subject. An actual RCT? Too good to be true! The thirst is for fact vs. speculation, influencer blogposts and self-promotion.
That this RCT provided evidence opposing the hype, is, of course, irresistible.
The other papers ive seen so far all used toy projects. This paper was the first ive seen that deliberately used realistic projects.
I think that part is understated, while the part about the "steep learning curve" has been considerably overstated. Ive watched plenty of good developers with hours of cursor practice, using plenty of relevant context giving clear instructions wasting chunks of 5-10 minutes at a time because 4/5 times claude produced crap that needed to be undone.
That story doesnt sell though. Nobody will pay me consulting fees to say it. Hype has more $$$ attached.
ofjcihen · 2d ago
To be fair that thirst probably comes from people who aren’t seeing the gains the hype would lead you to believe and are reaching into the void to not feel like they’re taking crazy pills.
It’s also probably not coming from a place of “I’m scared of AI so I want it to fail” but more like “my complex use case doesn’t work with AI and I’m really wondering why that is”.
There’s this desire it seems to think of people who aren’t on the hype train as “against” AI but people need to remember that these are most likely devs with a decade of experience who have been evaluating the usefulness of the tools they use for a long time.
jsbisviewtiful · 2d ago
Personal take, but I think some people also understand that the hype machine around AI is coming from the rich and C-level people. Meanwhile companies are widely and openly axing jobs and or not paying artists, citing AI as the source of their new fortune. Personally, my use of AI in my job has so far not been that fruitful and for something that has so far dramatically underdelivered on its promises of utopian ideas we are instead actively seeing it used to undercut the 99% - and that’s not even getting into the environmental impact or the hellscape it’s made of the Internet.
archagon · 2d ago
It’s telling that there seem to be no AI companies with “boring” CEOs. It’s all glassy-eyed zealots talking about how their product will upend the workforce.
Yoric · 2d ago
> To be fair that thirst probably comes from people who aren’t seeing the gains the hype would lead you to believe and are reaching into the void to not feel like they’re taking crazy pills.
Yes, exactly!
I've spent way too much time trying to get anything remotely close to an LLM writing useful code. Yeah, I'm sure it can speed up writing code that I can write in my sleep, but I want it to write code I can learn from, and so far, my success rate is ~0 (although the documentation along the bogus code is sometimes a good starting point).
Having my timelines filled by people who basically claim that I'm just an idiot for failing to achieve that? Yeah, it's craze-inducing.
Every time I see research that appears to confirm the hype, I see a huge hole in the protocol.
Now finally, some research confirming my observations? It feels so good!
Terr_ · 2d ago
Trying out an analogy:
The time is the early 2000s, and the Segway™ is being suggested as the archetype of almost all future personal transportation in cities and suburbs. I don't hate the product, there's neat technology there, they're fun to mess with, but... My bullshit sensor is still going off.
I become tired of being told that I'm just not using enough imagination, or that I would understand if only I was plugged into the correct social-groups of visionaries who've given arguments I already don't find compelling.
Then when somebody does a proper analysis of start/stop distance, road throughput, cargo capacity, etc, that's awesome! Finally, some glimmer of knowledge to push back the fog of speculation.
Sure, there's a nonzero amount of confirmation bias going on, but goshdangit at least I'm getting mine from studies with math, rather than the folks getting it from artistic renderings of self-balancing vehicles filling a street in the year 2025.
goalieca · 1d ago
For some, it is also coming from a place that their company leadership is mandating AI use.
emp17344 · 2d ago
Well, there aren’t any studies showing AI agents boost productivity, so it’s all we’ve got. It seems like a well-conducted study, so I’m inclined to trust its conclusions.
simonw · 2d ago
One of the articles linked from the OP includes links to such studies: https://theconversation.com/does-ai-actually-boost-productiv... - scroll down to the "AI and individual productivity" section, there are two papers there on the "increases productivity" side followed by two others that didn't.
bluefirebrand · 2d ago
Anyone who is an employee drawing a salary should be extremely hopeful that AI productivity doesn't work.
Why should we be eager to find out that some new tech is going to undercut us and replace us, devaluing us even more than we already are?
simonw · 2d ago
Do you benefit from open source?
Open source packages are the biggest productivity boost of my entire career, at no point did I think "wow, I wish these didn't exist, they're a threat to my livelihood".
bluefirebrand · 2d ago
If someone released an open source software that immediately replaced the software you were working on and you lost your job, what then?
What if that happened across the entire industry overnight?
What if every single time you sat down to write a piece of software, someone immediately produced the Open Source version of it?
That's what people are hoping AI is going to do to us
Why would I support that
simonw · 2d ago
> If someone released an open source software that immediately replaced the software you were working on and you lost your job, what then?
WordPress did that to the commercial CMS industry. Linux did that to the commercial Unix industry (RIP, Sun Microsystems).
> That's what people are hoping AI is going to do to us
I think those people are full of crap, and I mostly ignore them.
fragmede · 2d ago
The thing about open source, is that it came from a programmer in the halls of academia, MIT. But imagine the idea came from a Harvard Business School MBA instead. "You should do a bunch of work for free, and then give it to me for free, and uh, you can figure out a business model later." Does that sound remotely sane?
I have no leg to stand on, I owe my career to proprietary software building on top of open source code, but lately I have been wondering what that alternate world would look like. As it stands, if you want to write software and sell it, you've got at most a couple of months before someone open sources a clone of your project. Sure, there's no guarantee it'll be anywhere near as good. Photoshop is still king. But one can't help but wonder what the proprietary IBM world would have been like.
simonw · 2d ago
We lived in that world prior to ~2000. It sucked. Programmers spent all of their time reinventing the wheel - so much so that when I was at university around 2001-2004 one of the biggest topics was "software reusability". How could we get out of the trap of writing the same things over and over again?
Back then the answer was supposed to be OOP. With hindsight, the answer was open source - in particular robust library ecosystems like PyPI and NPM.
Yoric · 1d ago
Hey, I'm old enough to remember "components" being a research topic in academia, with researchers theorizing what we know today as services & APIs.
It looked more fun on paper, FWIW.
didibus · 2d ago
I've always thought that I'd probably be paid even more and have the choice of many more jobs if open source didn't exist.
Could it be there'd be less jobs because it would have prevented some businesses to bootstrap and stay afloat? That's also possible.
And this is really the thing with AI, some think it'll grow the pie, therefore increasing the demand for developers. Some think it'll just shrink the size of the piece of pie that's for developers.
If we could predict the former, and be sure of it, I doubt anyone would be against AI.
Aside that though, as a developer, you are also concerned with, is this actually making my job better/faster/more-fun. Or am I just forced into an annoying process of using a tool that doesn't help and will just cause more problems I'll have to deal with.
simonw · 2d ago
I am so confident that open source expanded the job market.
Prior to open source software development was awful. We spent all of our time re-inventing wheels - or, if we were lucky, licensing expensive badly shaped wheels from vendors and crossing our fingers that they would work (because we couldn't fix them if they didn't).
As a result, very few companies developed software - it's not worth hiring a software engineer if it takes them six months to deliver that initial login screen.
Open source started taking off about 25 years ago, and software productivity skyrocketed. The result was a multi-decade boom in jobs and salaries for us software engineers.
didibus · 2d ago
It's plausible, I just want to point out it's hard to know if that's just coincidental, as opposed to causal.
thunky · 1d ago
> We spent all of our time re-inventing wheels
Still do.
At least back then it probably felt more like we were doing something new because 8000 versions of the same thing wasn't right at your fingertips.
Yoric · 1d ago
Good point.
Open-source does replace entire companies, but typically with better products that other companies can build upon, and in the end, this expands the demand for software.
AI is expected to replace developers, and I believe that it will empower non-techies to develop highly-functional mockups, which will in turn increase the amount of (crappy) software on the market, which I believe will in turn increase the demand for people who can fix this software – if we multiply the number of apps by 1,000, even if only 1% of the apps need quality, this still means that we've increased the demand for software developers 10x.
So I actually think that, within a few years, AI will increase our job security as software developers. It's going to be a rocky few years, of course, but I think we'll be better off, at least until the rest of our dystopian timeline catches up.
Now, what I hate is the dishonesty of AI companies. I mean, as software developers, we strive on rigor: good documentation, compilers and JITs that speed up our code without changing its behavior, APIs and tools that work as claimed, etc. AI companies and their PR, on the other hand, peddle magic thinking and gaslighting. That's the exact opposite of rigor. And this keeps assaulting my sanity.
kbelder · 2d ago
Should we be wary of any productivity gains?
Should be looking for ways to work slower? I can go back to just one monitor.
bluefirebrand · 2d ago
If I personally do not benefit from productivity gains, then why should I be excited about them?
If productivity gains mean that either me or a coworker is laid off, and my boss makes more money, why should I care about the that?
Having a job is not a favor that people with money are doing for employees, it is an exchange of services for money
I provide a service. So do most people. Anything that gives the people with money a way to avoid paying for my service is bad for most of us
munksbeer · 1d ago
This is a slightly shallow way of looking at productivity gains. You benefit from productivity gains across every industry. Without consumers, there is no capital. So you, the consumer, benefit from increased productivity, even if it means that somewhere, someone, lost out. We just have to go back to the original luddite example. It is cliche for a reason. A quote:
"The tragedy is that although these groups of workers did suffer from the deployment of new technology, over time, the general population benefitted from cheaper clothes, lower prices and new jobs in other sectors of the economy."
Obviously, as a developer, I don't want to lose my career. But am I a hypocrite? Are you a hypocrite for buying anything that is now manufactured by automation, putting people out of jobs? I'd say yes.
ajaisjdh · 2d ago
There’s literally billions of dollars on the “pro AI side”.
What you’re seeing is a thirst for objective reporting. The average person only has the ability to provide anecdotes - many of which are in stark contrast to the narrative pushed by the billionaires pumping AI.
I don’t think anyone serious thinks AI isn’t useful in some capacity - but it’s more like a bloom filter than a new branch of mathematics. Magically powerful in specific use cases, but not a paradigm shift.
didibus · 2d ago
Personally, I think this is a disingenuous take. The thirst is for tangible data, the issue is that we've never been able to measure any form of productivity/quality in software development.
My team does two person PR reviews for example. We'd go a lot faster if we didn't or even just allowed a single reviewer. Similarly, we have no idea what the quality impact would be if we stopped, and what we gain by doing so. We are we not having a 3 reviewer rule for example, why not, two is an arbitrary number?
Unit tests... We'd surely go a lot faster if we didn't bother with them. Teams used to have some dedicated QA members and you'd rely entirely on manual testing. You can push a lot more code out. Was software in the 90s when unit tests and integ tests wasn't used buggier than today's software?
Now take AI, what is the impact of its use? It's not even obvious if it reduced the time it takes to launch a feature, my team isn't suddenly ahead of schedule on all our projects, even though we all use Agentic tools actively now. Ask any one of us and "I think it makes us faster" will be the answer. But ask us why we have a 2 person review rule and we'd similarly say: "I think it prevents bugs and improves the code quality".
The difference with AI now is that you pay for it, it's not free. Having unit tests or doing a 2 person review is just a process change. AI is something you pay for, so there's more desire to know for sure. And it also is something people would like to know if they can lower their headcount without impacting their competitive edge and ability to deliver fast and with good enough quality. Nobody wants to lower the headcount and find out the hard way.
thewebguyd · 2d ago
> the issue is that we've never been able to measure any form of productivity/quality in software development.
Yep. It's been a "problem" for decades at this point. Business types constantly trying, and failing, to find some way to measure dev productivity, like they can with other types of office drone work.
We've been through Lines of Code, Function Points, various agile metrics, etc. None of these have given business types their holy grail of a perfectly objective measure of productivity. But no one wants to accept an answer of "You just can't effectively measure productivity in software development" because we now live in a data-driven business culture where every little thing must be measured and quantified.
qsort · 2d ago
I don't think AI lives up to the current hype but this article is garbage.
They're obviously talking about the METR paper, but the main takeaway according to the authors themselves was that self-reporting productivity increases is unreliable, not that you should cancel your subscription.
Nothing in that paper said that AI can't speed up software engineering.
Why are we responding to hype with nonsense?
didibus · 2d ago
> Nothing in that paper said that AI can't speed up software engineering
I mean, the paper did provide tangible data that at least in their experiments, AI slowed down software engineering.
What they said is that it's not a proof that there's isn't a scenario or a mechanism where AI could result in speeding up software engineering. For that more research would be needed in measuring productivity of AI in more varied contexts.
For me at least, their experiment seem to describe the average developer's use of AI. So it's probably telling you that currently on average AI might be slowing things down.
Now the question is, can we find good data of outliers, and is it a simple matter of figuring out how to use it effectively, so we can upskill people and get the average to now be faster. Or will the outlier be conditioned on like, only for newbies, only for prototypes, only for the first X weeks on a greenfield code base, etc.
Edit: That said, the most fascinating data point of that study is how software engineers are not able to determine if AI makes them faster or slower, because they all thought they were 20% faster but were 19% slower in reality. So now you have to become really skeptical of anyone who claims they found a methodology or a workflow where their use of AI makes them faster. We need better measurement than just "I feel faster".
nayshins · 2d ago
I'm happy to let people think that AI does not yield productivity gains. There is no point engaging on this topic, so I will just outwork/outperform them.
quxbar · 2d ago
I now have the pleasure of giving exercises to candidates where they are explicitly allowed to use any AI or autocomplete that they want, but it's one of those tricky real-world problems where you'll only get yourself into trouble if you only follow the model's suggestions. It really separates the builders from the bureaucrats far more effectively than seeing who can whiteboard or leetcode.
jamil7 · 2d ago
Its kind of a trap, we allow people in interviews to do the same and some of them waste so much time accepting wrong LLM completions and then changing them than if they'd just written the code themselves.
pydry · 2d ago
Ive been doing this inadvertently for years by making tasks that were as realistic as possible - explicitly based upon the code the candidate will be working upon.
As it happens, this meant when candidates started throwing AI at the task, instead of performing that magic it usually can when you make it build a todo app or solve some done-to-death irrelevant leetcode problem it flailed and left the candidate feeling embarrassed.
I really hope AI signals the death knell of fucking stupid interview problems like leetcode. Alas many companies are instead knee jerking and "banning" AI from interview use instead (even claude, hilariously).
BoiledCabbage · 2d ago
> but it's one of those tricky real-world problems where you'll only get yourself into trouble if you only follow the model's suggestions.
What's the goal of this? What are you looking for?
AnimalMuppet · 2d ago
I presume, people who can code, as opposed to people who can only prompt an LLM.
In the real world, you hit problems that the LLM doesn't know what to do with. When that happens, are you stuck, or can you write the code?
flashgordon · 2d ago
Id be seeing if the candidate actually understanding what the llm is spitting out and pushing back when it doesn't make sense vs are they one of the "infinite monkeys on infinite typewriters"
skeeter2020 · 2d ago
IF (and a big IF) LLMs are the future of coding, this doesn't mean humans don't do anything, but the role has changed from author to editor. Maybe you don't need to create the implementation but you sure better know how to read and understand it.
yomismoaqui · 2d ago
That's really interesting... can you give more details about the problem you are using?
This sounds like in there will be a race between this kind of booby trap tests and AIs learning them.
gritzko · 2d ago
Long-tail problems are not reiterated in the dataset.
Making LLM remember that can be difficult.
elpakal · 2d ago
Some code challenge platforms allow for seeing how often someone pasted things in. That's been interesting.
arealaccount · 2d ago
Interesting, care to elaborate? Or this is a carefully guarded secret?
raydev · 2d ago
Not sharing what our coding questions are, but we also allow LLMs now. Interviewees choice to do so.
In quite a few interviews in the last year I have come away convinced that they would have performed far better if they had relied on their own knowledge/experience exclusively. Fumbling with windows/tabs, not quite reading what they are copying, if I ask why they chose something, some of them would fold immediately and opt for something way better or more sensible, implying they would have known what to do had they bothered to actually think for a moment.
I put down "no hire" for all of them of course.
Lionga · 2d ago
If you are so happy to let people think that AI does not yield productivity gains, why comment here?
How exactly did you outperform? Show, don't talk.
nayshins · 2d ago
I rolled out a migration to 60+ backends by using Claude code to manage it in the background. Simultaneously, I worked on other features while keeping my usual meeting load. I have more commits and releases per week than I have had in my whole career, which is objectively more productive.
ajaisjdh · 2d ago
> I rolled out a migration to 60+ backends
How is anyone supposed to understand what this means?
Given the ambiguity in your description and lack of actual code it’s hard to take you seriously.
seszett · 2d ago
Sometimes when I read such meaningless things my first reaction is to feel like I'm too ignorant to understand what the person says.
But then when I really think about it usually they're just bullshitting out being purposefully vague, using terms that don't mean anything precise in order to avoid actual criticism.
dkn · 2d ago
I question your assertion that more commits and releases per week is more productivity. There could be unexpected effects from your commits that create more work for you or for others and that could be hard to quantify.
Doing bad things faster might feel more productive to you, but it doesn’t mean that you are delivering more value. You might be, but the metrics you have shared to not prove that.
bluefirebrand · 2d ago
If you are producing much lower quality then no, it is not objectively more productive
nayshins · 2d ago
If you find yourself working at the same place as me, feel free to judge my work, but until then, enjoy speculating.
haswell · 2d ago
The issue I have with comments like this one is the one-dimensional notion of value described as "productivity gains" for a single person.
There are many things in this world that could be fairly described as "more productive" or "faster" than the norm, yet few people would argue that it makes those things a net benefit. You can lie and cheat your way to success, and that tends to be successful too. There are good reasons society frowns on this.
To me, focusing only on "I'm more productive" while ignoring the systemic and societal factors impacted by that "productivity" is completely missing the forest for the trees.
The fact that you further feel that there isn't even a point in engaging on the topic is disturbing considering those ignored factors.
troupo · 2d ago
> I'm happy to let people think that AI does not yield productivity gains.
vs.
--- start quote ---
In a randomised controlled trial – the first of its kind – experienced computer programmers could use AI tools to help them write code.
"Your quote is very representative of the magical wishful thinking most people have about AI"
Your comment here is very representative of how quickly people who are AI skeptics will jump on anything that supports their skepticism.
troupo · 2d ago
The person above literally pitches an unsupported belief against a study (however flawed it may be). If that doesn't support my skepticism, I don't know what does.
simonw · 2d ago
When you've lived with AI boosting your productivity for a year or more it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
kbelder · 2d ago
>it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
In my youth, I would have argued this was bad. Now, I tend to agree. Not that studies are worthless; but they are just part of the accumulation of evidence, and when they contradict a clear result you are directly seeing, you need to weight the evidence appropriately.
(Obviously, replicated studies showing clear effects should be more heavily weighted.)
Everything is just shifting odds.
hooverd · 2d ago
When you've lived with -AI- stimulants boosting your productivity for a year or more it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
we don't demand every developer pop Adderall though
NoGravitas · 2d ago
Don't give the management any ideas.
troupo · 2d ago
> When you've lived with AI boosting your productivity for a year or more it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
Me: The person above literally pitches an unsupported belief against a study
You: it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
Really? Really?!!
As for "boosting your productivity", it's also what I'm talking about in the article I linked:
--- start quote ---
For every description of how LLMs work or don't work we know only some, but not all of the following:
- Do we know which projects people work on? No
- Do we know which codebases (greenfield, mature, proprietary etc.) people work on? No
- Do we know the level of expertise the people have? No. Is the expertise in the same domain, codebase, language that they apply LLMs to? We don't know.
- How much additional work did they have reviewing, fixing, deploying, finishing etc.? We don't know.
Even if you have one person describing all of the above, you will not be able to compare their experience to anyone else's because you have no idea what others answer for any of those bullet points.
--- end quote ---
So what happens when we actually control and measure those variables?
Wait, don't answer: "no, it's easier to believe yourself over a study".
See? Skeptics don't even have to "jump on anything that supports their skepticism." Even you supply them with material.
simonw · 2d ago
I'm not disputing that different people have radically different experiences of how much productivity boost they can get out of working with LLMs.
What I'm willing to assert as fact, based not just on my own experiences (though they're a major role) but on observing this space for several years and talking to literally hundreds of people, is that LLMs can provide you a very real productivity boost in coding if you take the time to learn how to use them - or if you get lucky and chance upon the most productive patterns.
I was just going to write "see, you actually agree with me", but got hit by the reply rate limit :)
And I agree with >90% of what you write, so I was surprised that this bout took us to weird places.
gishglish · 2d ago
These fucking people sound like clickbait ads trying to sell me grift pills lmao. Go into the dietary supplement business instead, you’ll do well.
pydry · 2d ago
You remind me of the haskell hype of ~2016-2018 where the community wrote tons of blog posts and passionate comments on HN about the theory of types and the "productivity boost" of their language while simultaneously producing a paltry output of actual useful software in the hands of actual users.
Im sure they were completely genuine in how they felt, just as i am sure you are too.
refulgentis · 2d ago
I don't know if you're doing a parody of study-ism, or, if you're a bit ashamed to be doing it.
Either way, it's hard to interlocute if your misreading is an absolute conclusion that cannot be argued, then transmutated warranted skepticism.
jgalt212 · 2d ago
The converse is also true.
ofjcihen · 2d ago
I don’t think the snark is warranted here. The person you’re responding to is linking an article dealing with that issue and the title is what they mentioned.
Edit: SimonW? Really? I didn’t see the name but I didn’t expect you to be like that.
simonw · 2d ago
I stand by what I said.
I don't think the response from troupo that nayshins's personal experience is invalidated by a "randomised controlled trial" was well argued, so I imitated what I saw as their snarky wording with my own reworded version of it.
I do take the "AI isn't actually a productivity boost" thing a little bit personally these days, because the logical conclusion for that is that I've been deluding myself for the past two years and I'm effectively a victim of "magical thinking".
(That said, I did actually go to delete my comment shortly after posting it because I didn't think it added anything to the conversation, but it had already drawn a reply so I left it there.)
ofjcihen · 2d ago
That’s totally fair and I get it.
Working in security I often feel the same way and let’s be fair in the grand scheme of things it’s not that big of a deal.
troupo · 2d ago
> because the logical conclusion for that is that I've been deluding myself for the past two years and I'm effectively a victim of "magical thinking".
You may just as well have. I, for one, am absolutely ready to re-evaluate any and all approaches I have with AI to see if I am actually more productive or not.
But moreover, your own singular experience with your own code and projects may make you more productive. We don't know if it does because we don't have a baseline against which to measure.
But even moreover over that moreover is that we don't even have a question "does a single senior engineer's experience with his own code and approaches can be generalised over the entire population of programmers?" Skeptics say: no (and now have some proof of that). Optimists loudly say: yes, of course, and dismiss everyone who dares contradict out of hand.
refulgentis · 2d ago
Not OP, not sure what you mean but curious.
- Snark?
- Is "the issue" that anyone who claims any productivity gain is using magical thinking?
- How does the linked article "deal with" "the issue"?
- What title did they mention?
- What did they link to that has that title?
> Edit: SimonW? Really? I didn’t see the name but I didn’t expect you to be like that.
Like what? I think you're getting a bit emotional & personal here, I don't read anything remotely inappropriate into Simon's comment. Been here 15 years. OP's was odd for HN in that it admits 0 argument: if you think you have productivity gains, it's magical thinking.
ofjcihen · 2d ago
So I understand how people can get emotional about this topic but really quick let’s calm down and reevaluate what I said.
My comment was in response to Simon’s reply to a user who posted an article. The title of the article they posted addresses magical thinking in AI.
Now whether that’s an opinion you share or not is not the point. Simon responded as if the user was only calling any perceived gains from AI as magical thinking which is not the case.
I’ll let you come back to that when you feel like it. Altogether, though it’s just disappointing to see someone who’s work I read often jumping to an emotional response when it’s not warranted.
fragmede · 2d ago
Never meet your heros. Then again, if the worst is a bit of emotionally charged snark, of which we are all guilty of at some point, I think I'll live.
ofjcihen · 2d ago
Very true and to be fair Simon’s response with his perspective is valid and understandable.
refulgentis · 2d ago
(not op)
Gosh, I was conflicted, then you pulled out that sentence and I was convinced. :)
Alternatively: When faced with a contradiction, first, check your premises.
I don't want to belabor the point too much, there's little common ground if we're at all or nothing thinking - "the study proved AI is net-negative because of this pull quote" isn't discussion.
pydry · 2d ago
ive watched a lot of people code with cursor, etc. and i noticed that they seem to get a rush when it occasionally does something amazing that more than offsets their disappointment when it (more often) screws up.
the psychological effect reminds me a bit of slot machines, which provide you with enough intermittent wins to make you feel like you're winning while youre lose.
I think this might be linked to that study that found experienced oss devs who thought they were faster when they were in actual fact 20% slower.
hooverd · 2d ago
Crazy how productivity gains just lead to more work for you.
stronglikedan · 2d ago
Crazy how people would let their managers know they could get more done, instead of getting the same amount done quicker and having more free time.
gishglish · 2d ago
If I was genuinely getting more free time, I’d be more amenable to this line of thought, but RTO put the nail in that coffin.
nayshins · 2d ago
more work is good for the soul... until it isnt
bluefirebrand · 2d ago
This is actually worth talking about imo
There is nothing in it for me, if I am more productive but earn the same and don't get any more time off
Why should I bother at that point?
Kurtz79 · 2d ago
If we assume that AI coding actually increases productivity of a programmer without side effects (which of course is a controversial assumption, but not affecting the actual question):
1) If you are a salaried employee, if you are seen as less productive than your colleagues that use AI, at the very least you won't be valued as much. Either you will eventually earn less than your colleagues or be made redundant.
2) If you are a consultant, you'll be able to invoice more work in the same amount of time. Of course, so will your competitors, so that rates for a set amount work will probably decrease.
3) If you are an entrepreneur, you will be able to create a new product hiring less people (or on your own). Of course, so will your competitors, so that the expectations for viable MVPs will likley be raised.
In short, if AI coding assistants actually make a programmer more productive, you will likely have to learn to live with it in order to not be left behind.
danaris · 2d ago
This is only true if the degree to which they increase productivity meaningfully rises above the level of noise.
That is to say: "Productivity" is notoriously extremely hard to measure with accuracy and reliability. Other factors such as different (and often terrible) productivity measures, nepotism/cronyism, communication skills, self-marketing skills, and what your manager had for breakfast on the day of performance review are guaranteed to skew the results, and highly likely, in what I would guess is the vast majority of cases, to make any productivity increases enabled by LLMs nearly impossible to detect on a larger scale.
Many people like to operate as if the workplace were a perfectly efficient market system, responding quickly and rationally to changes like productivity increases, but in fact, it's messy and confusing and often very slow. If an idealized system is like looking through a pane of perfectly smooth, clear glass, then the reality is, all too often, like looking through smudgy, warped, clouded bullseye glass into a room half-full of smoke.
bluefirebrand · 2d ago
The problem is that it doesn't actually matter if it really makes a programmer more productive or not.
Because productivity is hard to measure, if we just assume that using AI tools is more productive we're likely to be making stupid choices
And since I strongly think that AI coding is not making me personally more productive it puts me in a situation where I have to behave irrationally in order to show employers that I'm a good worker bee
I am increasingly feeling trapped between a losers choice. I take the mental anguish of using AI tools against my vetter judgment or I take the financial insecurity (and associated mental anguish) of just being unemployed
CharlesW · 2d ago
The question is what happens when you're half as productive as everyone around you.
actually based on your own admission this is not what you're doing...
throwawayqqq11 · 2d ago
Green or naturally grown brown field projects?
People who boast about AI enhanced productivity seem to always forget to mention.
thefz · 2d ago
Until you will not have access to it and be outperformed by people used to thinking every day.
masfuerte · 2d ago
Yet here you are, engaging.
nayshins · 2d ago
the art of the bait
kubb · 2d ago
Good luck getting paid more for your improved performance :D
some_random · 2d ago
If they're right in their belief that AI usage leads to significantly more performance, their compensation is that they will keep their job.
bluefirebrand · 2d ago
"You get to keep your job" is the worst consolation prize I can imagine
stronglikedan · 2d ago
No one gets paid more to get the same job done, unless you count free time as compensation.
nayshins · 2d ago
I get paid a lot already.
kubb · 2d ago
Then enjoy demonstrating your AI-supercharged productivity advantage over your inferior peers, I think they'll love that.
nayshins · 2d ago
Turns out that I spend a lot of my spare time teaching others how to use them effectively. My goal is to help as many people become experts at this as I can.
gishglish · 2d ago
> so I will just outwork/outperform them.
At the game of producing garbage slop? Probably yeah.
archagon · 2d ago
Uh, and who are you, exactly?
spaceman_2020 · 2d ago
All AI impact studies and research papers need to be taken with a pinch of salt. The field is moving so fast that by the time you get peer reviewed, you’re already outdated
I’ve watched coding change from Cursor-esque IDEs to terminal based agentic tools within months.
djhn · 1d ago
Which also means spending a lot of time learning AI tools gets wasted as the tooling improves.
I still suspect the vast silent majority of professional software devs haven’t integrated any, even Cursor-style, AI tools in to their main gig.
And I reckon that’s completely rational, for those that have made this choice explicitly.
Early adopters of AI tools are making a speculative bet, but so far most of them seem happy with the return.
elicash · 2d ago
I'm building things at a level of complexity I wouldn't have even attempted without AI.
This piece, however, only focuses on time spent on a task that could be done both ways. Even there, it falls short. Let's assume this study is correct and a specific coding task does take me 19% more time with AI. I can still be more productive because the AI doing some of the work allows me to do other tasks during that time.
I do worry about atrophy of my mind outsourcing too many tasks, admittedly. But that's a different issue.
hattmall · 2d ago
There is really no argument that AI creates some productivity gains. Even if it's just an improved autocomplete. Because autocomplete does create some productivity gains. Pushing farther is murkier though. When it comes to the bulk of the type of work that AI is proving useful for, one of the main questions is why is there a need for speed. It's not in a sense of fear of automating jobs, it's just that generally we are reaching the bottlenecks more quickly and potentially causing more, but different problems, than we are solving.
It's similar to the story of the development of vehicles and how even though we move much faster we spend a greater amount of time in transit. My mom used to lament how annoying it was to have to drive to the grocery store because when she was younger and not everyone had cars the store came to you. Twice a day, in the morning and the evening the "rolling store" would drive through the neighborhood and if they didn't have what you needed right then, they would bring it on the next trip. We are finally coming back full circle with things like Instacart but it's taken a solid ~60 years of largely wasted inefficient travel times.
asdev · 2d ago
I think AI greatly reduces the starting costs for mundane tasks/boilerplate, for a reduction in velocity in implementation. So possibly an illusion that programmers feel more productive. It could be that RPE(Rate of perceived exertion) is lower when using AI for tasks, but raw throughput may be higher if programmers just do the jobs themselves and get into productive/flow state.
thewebguyd · 2d ago
> t could be that RPE(Rate of perceived exertion) is lower when using AI for tasks, but raw throughput may be higher if programmers just do the jobs themselves and get into productive/flow state.
I think you're onto something with this take, based on my own experience. I definitely agree that my RPE seems lower when I'm using AI for things, whether it actually is making me more productive or not over the long term remains to be seen but things do certainly "feel" easier/less cognitively demanding. Which, tbh, is still a benefit even if it doesn't result in large gains in output. Putting in less cognitive load at work just conserves my energy for things that matter - everything else outside of $dayjob.
nromiun · 2d ago
So much money is being pumped into this whole thing that we might get a global economic shock if/when it unravels.
kubb · 2d ago
That won't happen - this money was stolen from the middle class via currency devaluation. Spending it on any economic activity, no matter how pointless, is actually better than just gobbling up assets.
nromiun · 2d ago
An economic recession is not selective. It will affect all of us.
kubb · 2d ago
Rest assured that those having the resources will be least affected.
nromiun · 2d ago
Exactly. The rich take huge bets to get even richer and if they fail the poor people are the most affected. See all the hedge funds margin called, 2000 recession or 2008.
GaggiX · 2d ago
This article is based on the study that had a population of 16 people, most of which never used the tool before.
impure · 2d ago
I find it funny that the first sentence is: I’m not sure if the whole AI thing is falling apart, but we are definitely in a phase where the hype doesn’t match what we are seeing in the real world.
Also I do not think AI is falling apart and I do not think the productivity loss is surprising. It's something I've noticed a lot, you have to check AI output line by line thereby making you slower. AI makes tasks easier, not necessarily faster. And because people are lazy and love making things easier it's still going to be adopted no matter what anyone says.
Mars008 · 2d ago
> “The results surprised us,” research lab METR reported. “Developers thought they were 20pc faster with AI tools, but they were actually 19pc slower when they had access to AI than when they didn’t.”
They forgot about training and practice. Try it with airplanes. "The results surprised us. Given Boeing 777 they thought they will get from London to New York in 7 hours, but they actually couldn't even get off the ground. In fact they couldn't even start the engines."
ikmckenz · 2d ago
> They forgot about training and practice. Try it with airplanes. "The results surprised us. Given Boeing 777 they thought they will get from London to New York in 7 hours, but they actually couldn't even get off the ground. In fact they couldn't even start the engines."
If your "AI Assistant" is as hard to operate as a trans-Atlantic flight on a Boeing 777, perhaps it's not a very good assistant?
Mars008 · 2d ago
It's a tool which is like others. It has limits and needs to get used to. From my experience first attempts are overoptimistic. I tried big chunks at once. Then I learned smaller tasks work better. But the most important thing is it can do simple things in languages that I don't know and don't want to learn (due to natural brain's limitations). Just need to have things done, like simple web front end for application running its own server on some microcontroller.
mckngbrd · 2d ago
I had hardly written any code prior to ChatGPT other than teaching myself some VBA.
Since then, using Gen AI tools to learn + write code, I have deployed a functioning full stack application to the web. NextJS on Vercel, with a backend server also deployed running on Python, and a Supabase DB. Is it the best application ever making loads of money? Definitely not. Are there things wrong with it? Absolutely (although I promise I'm not exposing sensitive env vars and API keys to the web). Did the first versions look like absolute ass as I clumsily figured things out and made bad mistakes? You bet. But it's a functioning app that does some useful things and has real users.
I would never have imagined doing this in a million years prior to Gen AI.
Do some devs see mixed results depending on how they're using the tools? I'm sure. Is Gen AI overhyped broadly speaking? Probably so. But when I see people say it's a delusion, waste of resources, and everyone is wasting their time on it ... for me, it just doesn't line up.
DamnInteresting · 2d ago
> I have deployed a functioning full stack application to the web
Best hope ChatGPT plagiarized that code from people who properly sanitized user input, otherwise it might be vulnerable to SQL injection, XSS, etc. If such holes exist, it may be tough to resolve them without a professional.
> But when I see people say it's a delusion, waste of resources, and everyone is wasting their time on it ... for me, it just doesn't line up.
I don't think anyone would say LLMs are useless, but there are people out there comparing the birth of LLMs to the Industrial Revolution, and the advent of the Internet. Such claims are preposterous.
I think the widespread use of LLMs will ultimately be undone by a few factors:
1) Due to the way LLMs function, they will never be totally cured of 'hallucinations'. They are accurate a lot of the time, and apparently useful in those times. But you can never really trust its output. That is an exhausting problem, and will burn people out.
2) Prices for LLMs are artificially very low right now, these companies are burning investor money to keep going. The 'cost' is going to have to go way up for these LLM providers to be sustainable. I put cost in quotes because it will be some cost in subscription prices, but also cost in terms of ads that will be inserted everywhere, and other degrading forms of monetization.
3) Models are running out of good training data. They'll either hit a wall, or continue training on content that was itself LLM generated, which will go the way of Habsburg Jaw in a jiffy. Ensloppification is real.
I'm sure there will be niches for LLMs in the future, for those who can afford their true cost. But this cram-it-in-everything, available-everywhere frenzy will probably be looked back upon with deserved embarrassment.
ken47 · 2d ago
The overarching problem is that signal:noise ratio in contemporary discussions of AI is absurdly low. This is thanks in large part to a media complex whose primary goal seems to be the pumping of AI investments and stocks, and the associated cacophony of mendacious exaggerations. If only we could have a reasonable discussion without the "contributions" of the non-technical and/or salespeople...
mvdtnz · 2d ago
> This is the question. If AI helps you complete tasks faster, but those tasks generate more work, are you being more productive?
What an utterly bizarre question. Yes, by definition being more productive means doing more work.
mikemac29 · 2d ago
You're misunderstanding the question. It's not that you get more work done, it's that the AI creates work that wasn't there before that is now required to be done. Exanple - AI generates an email faster than I could type it without question, but now I need to go line-by-line to make sure nothing in it is inaccurate before sending it, versus typing it myself and only putting in information I know is accurate. Which scenario takes more time? That's what I think many of the measures are missing. Too many are looking at how quickly the email is typed and calling it a gain, and ignoring the part where you review it. That might still be more productive, but I've yet to see a single AI sales demo that acknowledges this in a realistic way.
nostromo · 2d ago
> Just because you can deal with your inbox more quickly doesn’t mean you’ll spend your afternoon on the beach. The more emails you fire off, the more you’ll receive back, and the never-ending cycle continues.
I've seen this firsthand. It's so easy to produce "content" now, like emails, presentations, etc., that more of my time now is just sifting through AI slop looking for signal.
I'm on a board for a nonprofit, and we've seen a person that is upset with the organization send dozens of AI-generated quasi-legal demands full of hallucinations of rules and laws that don't actually exist. So now we're forced to pay a real human lawyer to deal with a denial of service attack of ChatGPT slop and are spending lots of time just dealing with the onslaught. Our productivity has crashed as a board due to a single bad actor enabled by ChatGPT.
At work, I'm also getting buried in AI sales calls and AI resumes. It's getting more and more difficult to find signal in the noise.
It wouldn't surprise me if there's a lot of people who are experiencing lower productivity for similar reasons.
returnInfinity · 2d ago
AI is improving my productivity on simple tasks. It only needs to improve productivity of the world by 1-10% to be worth it
10% would mean a 10 trillion in productivity, 1% is 1 trillion.
lucaspauker · 2d ago
I think the productivity gains greatly depend on the coding task at hand. You could reasonably design a study that shows AI gains and slowdown depending on the task
elicash · 2d ago
I agree with this. But also, part of what made the original study interesting is the DISCREPANCY between perceived vs actual productivity, as opposed to which method "won."
It implies that even if some tasks are better with AI that it might not be so simple for us to judge which method is better for which task.
OldGreenYodaGPT · 2d ago
Honestly, at this point, I'll just work 2 hours a day since I'm already producing 5-10x the output of my coworkers who aren't leveraging Claude Code at the medium-sized startup where I work. It really makes you wonder what's holding people back if they haven't at least doubled their productivity by now
codingwagie · 2d ago
AI is way underhyped, its just going to take society a long time to integrate this tectonic shift into our existing ways of working
dr_dshiv · 2d ago
I agree it is underhyped. Many are like “oh, this is like blockchain, web3, ar/vr, quantum”— but whatever the promise of those technologies, AI is the only one with hundreds of millions of people productively using it everyday.
compiler-guy · 2d ago
Err, billions of people use the web everyday in ways that make them vastly more productive. Looking up random docs is now trivial whereas before the web, if you were luck your library might have an old copy.
And even then, the web was over-hyped at various points in its life. And it took a terrible crash to sort out what was real from what wasn’t. It’s not that the web wasn’t absolutely transformative (it was!) but it wasn’t what a huge number of proponents thought it would be either.
slowmovintarget · 1d ago
People don't want car-washes that consist of camels licking the car?
AI is NOT going away. Spend serious time to learn how leveraging AI can multiply your output ( quality and quantity ) or get left behind. Practice. Improve. Deliver. Period.
You’ve seen the posts and articles: “I’m inherently better than AI, coding faster and smarter.” Sure, maybe for now—especially if you’re half-assing your AI game. But that edge? It’s vanishing fast.
Stop being so self indulgent.
Become proficient with the tools available to you or get left behind.
NoGravitas · 2d ago
Sounds the same as when NFT bros used to reply to skeptics with "enjoy being poor".
nayshins · 2d ago
Many have seen the writing on the wall, but many are refusing to believe it.
some_random · 2d ago
I'm excited for the AI-cope delusion to fall apart.
It's not a bad paper, but it's also turning into a fantastic illustration of how much thirst there is out there for anything that shows that AI productivity doesn't work.
I just learned there was a 4 minute TV news segment about it on CNBC! https://www.youtube.com/watch?v=WP4Ird7jZoA
Maybe. I think it's a fantastic illustration of anyone doing anything to provide something other than hype around the subject. An actual RCT? Too good to be true! The thirst is for fact vs. speculation, influencer blogposts and self-promotion.
That this RCT provided evidence opposing the hype, is, of course, irresistible.
I think that part is understated, while the part about the "steep learning curve" has been considerably overstated. Ive watched plenty of good developers with hours of cursor practice, using plenty of relevant context giving clear instructions wasting chunks of 5-10 minutes at a time because 4/5 times claude produced crap that needed to be undone.
That story doesnt sell though. Nobody will pay me consulting fees to say it. Hype has more $$$ attached.
It’s also probably not coming from a place of “I’m scared of AI so I want it to fail” but more like “my complex use case doesn’t work with AI and I’m really wondering why that is”.
There’s this desire it seems to think of people who aren’t on the hype train as “against” AI but people need to remember that these are most likely devs with a decade of experience who have been evaluating the usefulness of the tools they use for a long time.
Yes, exactly!
I've spent way too much time trying to get anything remotely close to an LLM writing useful code. Yeah, I'm sure it can speed up writing code that I can write in my sleep, but I want it to write code I can learn from, and so far, my success rate is ~0 (although the documentation along the bogus code is sometimes a good starting point).
Having my timelines filled by people who basically claim that I'm just an idiot for failing to achieve that? Yeah, it's craze-inducing.
Every time I see research that appears to confirm the hype, I see a huge hole in the protocol.
Now finally, some research confirming my observations? It feels so good!
The time is the early 2000s, and the Segway™ is being suggested as the archetype of almost all future personal transportation in cities and suburbs. I don't hate the product, there's neat technology there, they're fun to mess with, but... My bullshit sensor is still going off.
I become tired of being told that I'm just not using enough imagination, or that I would understand if only I was plugged into the correct social-groups of visionaries who've given arguments I already don't find compelling.
Then when somebody does a proper analysis of start/stop distance, road throughput, cargo capacity, etc, that's awesome! Finally, some glimmer of knowledge to push back the fog of speculation.
Sure, there's a nonzero amount of confirmation bias going on, but goshdangit at least I'm getting mine from studies with math, rather than the folks getting it from artistic renderings of self-balancing vehicles filling a street in the year 2025.
Why should we be eager to find out that some new tech is going to undercut us and replace us, devaluing us even more than we already are?
Open source packages are the biggest productivity boost of my entire career, at no point did I think "wow, I wish these didn't exist, they're a threat to my livelihood".
What if that happened across the entire industry overnight?
What if every single time you sat down to write a piece of software, someone immediately produced the Open Source version of it?
That's what people are hoping AI is going to do to us
Why would I support that
WordPress did that to the commercial CMS industry. Linux did that to the commercial Unix industry (RIP, Sun Microsystems).
> That's what people are hoping AI is going to do to us
I think those people are full of crap, and I mostly ignore them.
I have no leg to stand on, I owe my career to proprietary software building on top of open source code, but lately I have been wondering what that alternate world would look like. As it stands, if you want to write software and sell it, you've got at most a couple of months before someone open sources a clone of your project. Sure, there's no guarantee it'll be anywhere near as good. Photoshop is still king. But one can't help but wonder what the proprietary IBM world would have been like.
Back then the answer was supposed to be OOP. With hindsight, the answer was open source - in particular robust library ecosystems like PyPI and NPM.
It looked more fun on paper, FWIW.
Could it be there'd be less jobs because it would have prevented some businesses to bootstrap and stay afloat? That's also possible.
And this is really the thing with AI, some think it'll grow the pie, therefore increasing the demand for developers. Some think it'll just shrink the size of the piece of pie that's for developers.
If we could predict the former, and be sure of it, I doubt anyone would be against AI.
Aside that though, as a developer, you are also concerned with, is this actually making my job better/faster/more-fun. Or am I just forced into an annoying process of using a tool that doesn't help and will just cause more problems I'll have to deal with.
Prior to open source software development was awful. We spent all of our time re-inventing wheels - or, if we were lucky, licensing expensive badly shaped wheels from vendors and crossing our fingers that they would work (because we couldn't fix them if they didn't).
As a result, very few companies developed software - it's not worth hiring a software engineer if it takes them six months to deliver that initial login screen.
Open source started taking off about 25 years ago, and software productivity skyrocketed. The result was a multi-decade boom in jobs and salaries for us software engineers.
Still do.
At least back then it probably felt more like we were doing something new because 8000 versions of the same thing wasn't right at your fingertips.
Open-source does replace entire companies, but typically with better products that other companies can build upon, and in the end, this expands the demand for software.
AI is expected to replace developers, and I believe that it will empower non-techies to develop highly-functional mockups, which will in turn increase the amount of (crappy) software on the market, which I believe will in turn increase the demand for people who can fix this software – if we multiply the number of apps by 1,000, even if only 1% of the apps need quality, this still means that we've increased the demand for software developers 10x.
So I actually think that, within a few years, AI will increase our job security as software developers. It's going to be a rocky few years, of course, but I think we'll be better off, at least until the rest of our dystopian timeline catches up.
Now, what I hate is the dishonesty of AI companies. I mean, as software developers, we strive on rigor: good documentation, compilers and JITs that speed up our code without changing its behavior, APIs and tools that work as claimed, etc. AI companies and their PR, on the other hand, peddle magic thinking and gaslighting. That's the exact opposite of rigor. And this keeps assaulting my sanity.
Should be looking for ways to work slower? I can go back to just one monitor.
If productivity gains mean that either me or a coworker is laid off, and my boss makes more money, why should I care about the that?
Having a job is not a favor that people with money are doing for employees, it is an exchange of services for money
I provide a service. So do most people. Anything that gives the people with money a way to avoid paying for my service is bad for most of us
"The tragedy is that although these groups of workers did suffer from the deployment of new technology, over time, the general population benefitted from cheaper clothes, lower prices and new jobs in other sectors of the economy."
Obviously, as a developer, I don't want to lose my career. But am I a hypocrite? Are you a hypocrite for buying anything that is now manufactured by automation, putting people out of jobs? I'd say yes.
What you’re seeing is a thirst for objective reporting. The average person only has the ability to provide anecdotes - many of which are in stark contrast to the narrative pushed by the billionaires pumping AI.
I don’t think anyone serious thinks AI isn’t useful in some capacity - but it’s more like a bloom filter than a new branch of mathematics. Magically powerful in specific use cases, but not a paradigm shift.
My team does two person PR reviews for example. We'd go a lot faster if we didn't or even just allowed a single reviewer. Similarly, we have no idea what the quality impact would be if we stopped, and what we gain by doing so. We are we not having a 3 reviewer rule for example, why not, two is an arbitrary number?
Unit tests... We'd surely go a lot faster if we didn't bother with them. Teams used to have some dedicated QA members and you'd rely entirely on manual testing. You can push a lot more code out. Was software in the 90s when unit tests and integ tests wasn't used buggier than today's software?
Now take AI, what is the impact of its use? It's not even obvious if it reduced the time it takes to launch a feature, my team isn't suddenly ahead of schedule on all our projects, even though we all use Agentic tools actively now. Ask any one of us and "I think it makes us faster" will be the answer. But ask us why we have a 2 person review rule and we'd similarly say: "I think it prevents bugs and improves the code quality".
The difference with AI now is that you pay for it, it's not free. Having unit tests or doing a 2 person review is just a process change. AI is something you pay for, so there's more desire to know for sure. And it also is something people would like to know if they can lower their headcount without impacting their competitive edge and ability to deliver fast and with good enough quality. Nobody wants to lower the headcount and find out the hard way.
Yep. It's been a "problem" for decades at this point. Business types constantly trying, and failing, to find some way to measure dev productivity, like they can with other types of office drone work.
We've been through Lines of Code, Function Points, various agile metrics, etc. None of these have given business types their holy grail of a perfectly objective measure of productivity. But no one wants to accept an answer of "You just can't effectively measure productivity in software development" because we now live in a data-driven business culture where every little thing must be measured and quantified.
They're obviously talking about the METR paper, but the main takeaway according to the authors themselves was that self-reporting productivity increases is unreliable, not that you should cancel your subscription.
Nothing in that paper said that AI can't speed up software engineering.
Why are we responding to hype with nonsense?
I mean, the paper did provide tangible data that at least in their experiments, AI slowed down software engineering.
What they said is that it's not a proof that there's isn't a scenario or a mechanism where AI could result in speeding up software engineering. For that more research would be needed in measuring productivity of AI in more varied contexts.
For me at least, their experiment seem to describe the average developer's use of AI. So it's probably telling you that currently on average AI might be slowing things down.
Now the question is, can we find good data of outliers, and is it a simple matter of figuring out how to use it effectively, so we can upskill people and get the average to now be faster. Or will the outlier be conditioned on like, only for newbies, only for prototypes, only for the first X weeks on a greenfield code base, etc.
Edit: That said, the most fascinating data point of that study is how software engineers are not able to determine if AI makes them faster or slower, because they all thought they were 20% faster but were 19% slower in reality. So now you have to become really skeptical of anyone who claims they found a methodology or a workflow where their use of AI makes them faster. We need better measurement than just "I feel faster".
As it happens, this meant when candidates started throwing AI at the task, instead of performing that magic it usually can when you make it build a todo app or solve some done-to-death irrelevant leetcode problem it flailed and left the candidate feeling embarrassed.
I really hope AI signals the death knell of fucking stupid interview problems like leetcode. Alas many companies are instead knee jerking and "banning" AI from interview use instead (even claude, hilariously).
What's the goal of this? What are you looking for?
In the real world, you hit problems that the LLM doesn't know what to do with. When that happens, are you stuck, or can you write the code?
This sounds like in there will be a race between this kind of booby trap tests and AIs learning them.
In quite a few interviews in the last year I have come away convinced that they would have performed far better if they had relied on their own knowledge/experience exclusively. Fumbling with windows/tabs, not quite reading what they are copying, if I ask why they chose something, some of them would fold immediately and opt for something way better or more sensible, implying they would have known what to do had they bothered to actually think for a moment.
I put down "no hire" for all of them of course.
How exactly did you outperform? Show, don't talk.
How is anyone supposed to understand what this means?
Given the ambiguity in your description and lack of actual code it’s hard to take you seriously.
But then when I really think about it usually they're just bullshitting out being purposefully vague, using terms that don't mean anything precise in order to avoid actual criticism.
Doing bad things faster might feel more productive to you, but it doesn’t mean that you are delivering more value. You might be, but the metrics you have shared to not prove that.
There are many things in this world that could be fairly described as "more productive" or "faster" than the norm, yet few people would argue that it makes those things a net benefit. You can lie and cheat your way to success, and that tends to be successful too. There are good reasons society frowns on this.
To me, focusing only on "I'm more productive" while ignoring the systemic and societal factors impacted by that "productivity" is completely missing the forest for the trees.
The fact that you further feel that there isn't even a point in engaging on the topic is disturbing considering those ignored factors.
vs.
--- start quote ---
In a randomised controlled trial – the first of its kind – experienced computer programmers could use AI tools to help them write code.
--- end quote ---
Your quote is very representative of the magical wishful thinking most people have about AI: https://dmitriid.com/everything-around-llms-is-still-magical...
Your comment here is very representative of how quickly people who are AI skeptics will jump on anything that supports their skepticism.
In my youth, I would have argued this was bad. Now, I tend to agree. Not that studies are worthless; but they are just part of the accumulation of evidence, and when they contradict a clear result you are directly seeing, you need to weight the evidence appropriately.
(Obviously, replicated studies showing clear effects should be more heavily weighted.)
Everything is just shifting odds.
we don't demand every developer pop Adderall though
Me: The person above literally pitches an unsupported belief against a study
You: it's pretty easy to believe your own experience over even a well-constructed "randomised controlled trial".
Really? Really?!!
As for "boosting your productivity", it's also what I'm talking about in the article I linked:
--- start quote ---
For every description of how LLMs work or don't work we know only some, but not all of the following:
- Do we know which projects people work on? No
- Do we know which codebases (greenfield, mature, proprietary etc.) people work on? No
- Do we know the level of expertise the people have? No. Is the expertise in the same domain, codebase, language that they apply LLMs to? We don't know.
- How much additional work did they have reviewing, fixing, deploying, finishing etc.? We don't know.
Even if you have one person describing all of the above, you will not be able to compare their experience to anyone else's because you have no idea what others answer for any of those bullet points.
--- end quote ---
So what happens when we actually control and measure those variables?
Wait, don't answer: "no, it's easier to believe yourself over a study".
See? Skeptics don't even have to "jump on anything that supports their skepticism." Even you supply them with material.
I've been banging this drum for over a year now: LLMs are deceptively difficult and uninituitive to use. Just one example: https://simonwillison.net/2025/Mar/11/using-llms-for-code/
What I'm willing to assert as fact, based not just on my own experiences (though they're a major role) but on observing this space for several years and talking to literally hundreds of people, is that LLMs can provide you a very real productivity boost in coding if you take the time to learn how to use them - or if you get lucky and chance upon the most productive patterns.
EDIT: I just saw you're the author of https://dmitriid.com/#everything-around-llms-is-still-magica... - that was a great piece! I think I may actually agree with you. I misinterpreted "magical thinking" as referring to something else.
Thank you!
> I think I may actually agree with you.
I was just going to write "see, you actually agree with me", but got hit by the reply rate limit :)
And I agree with >90% of what you write, so I was surprised that this bout took us to weird places.
Im sure they were completely genuine in how they felt, just as i am sure you are too.
Either way, it's hard to interlocute if your misreading is an absolute conclusion that cannot be argued, then transmutated warranted skepticism.
Edit: SimonW? Really? I didn’t see the name but I didn’t expect you to be like that.
I don't think the response from troupo that nayshins's personal experience is invalidated by a "randomised controlled trial" was well argued, so I imitated what I saw as their snarky wording with my own reworded version of it.
I do take the "AI isn't actually a productivity boost" thing a little bit personally these days, because the logical conclusion for that is that I've been deluding myself for the past two years and I'm effectively a victim of "magical thinking".
(That said, I did actually go to delete my comment shortly after posting it because I didn't think it added anything to the conversation, but it had already drawn a reply so I left it there.)
Working in security I often feel the same way and let’s be fair in the grand scheme of things it’s not that big of a deal.
You may just as well have. I, for one, am absolutely ready to re-evaluate any and all approaches I have with AI to see if I am actually more productive or not.
But moreover, your own singular experience with your own code and projects may make you more productive. We don't know if it does because we don't have a baseline against which to measure.
But even moreover over that moreover is that we don't even have a question "does a single senior engineer's experience with his own code and approaches can be generalised over the entire population of programmers?" Skeptics say: no (and now have some proof of that). Optimists loudly say: yes, of course, and dismiss everyone who dares contradict out of hand.
- Snark?
- Is "the issue" that anyone who claims any productivity gain is using magical thinking?
- How does the linked article "deal with" "the issue"?
- What title did they mention?
- What did they link to that has that title?
> Edit: SimonW? Really? I didn’t see the name but I didn’t expect you to be like that.
Like what? I think you're getting a bit emotional & personal here, I don't read anything remotely inappropriate into Simon's comment. Been here 15 years. OP's was odd for HN in that it admits 0 argument: if you think you have productivity gains, it's magical thinking.
My comment was in response to Simon’s reply to a user who posted an article. The title of the article they posted addresses magical thinking in AI.
Now whether that’s an opinion you share or not is not the point. Simon responded as if the user was only calling any perceived gains from AI as magical thinking which is not the case.
I’ll let you come back to that when you feel like it. Altogether, though it’s just disappointing to see someone who’s work I read often jumping to an emotional response when it’s not warranted.
Gosh, I was conflicted, then you pulled out that sentence and I was convinced. :)
Alternatively: When faced with a contradiction, first, check your premises.
I don't want to belabor the point too much, there's little common ground if we're at all or nothing thinking - "the study proved AI is net-negative because of this pull quote" isn't discussion.
the psychological effect reminds me a bit of slot machines, which provide you with enough intermittent wins to make you feel like you're winning while youre lose.
I think this might be linked to that study that found experienced oss devs who thought they were faster when they were in actual fact 20% slower.
There is nothing in it for me, if I am more productive but earn the same and don't get any more time off
Why should I bother at that point?
1) If you are a salaried employee, if you are seen as less productive than your colleagues that use AI, at the very least you won't be valued as much. Either you will eventually earn less than your colleagues or be made redundant.
2) If you are a consultant, you'll be able to invoice more work in the same amount of time. Of course, so will your competitors, so that rates for a set amount work will probably decrease.
3) If you are an entrepreneur, you will be able to create a new product hiring less people (or on your own). Of course, so will your competitors, so that the expectations for viable MVPs will likley be raised.
In short, if AI coding assistants actually make a programmer more productive, you will likely have to learn to live with it in order to not be left behind.
That is to say: "Productivity" is notoriously extremely hard to measure with accuracy and reliability. Other factors such as different (and often terrible) productivity measures, nepotism/cronyism, communication skills, self-marketing skills, and what your manager had for breakfast on the day of performance review are guaranteed to skew the results, and highly likely, in what I would guess is the vast majority of cases, to make any productivity increases enabled by LLMs nearly impossible to detect on a larger scale.
Many people like to operate as if the workplace were a perfectly efficient market system, responding quickly and rationally to changes like productivity increases, but in fact, it's messy and confusing and often very slow. If an idealized system is like looking through a pane of perfectly smooth, clear glass, then the reality is, all too often, like looking through smudgy, warped, clouded bullseye glass into a room half-full of smoke.
Because productivity is hard to measure, if we just assume that using AI tools is more productive we're likely to be making stupid choices
And since I strongly think that AI coding is not making me personally more productive it puts me in a situation where I have to behave irrationally in order to show employers that I'm a good worker bee
I am increasingly feeling trapped between a losers choice. I take the mental anguish of using AI tools against my vetter judgment or I take the financial insecurity (and associated mental anguish) of just being unemployed
actually based on your own admission this is not what you're doing...
People who boast about AI enhanced productivity seem to always forget to mention.
At the game of producing garbage slop? Probably yeah.
I’ve watched coding change from Cursor-esque IDEs to terminal based agentic tools within months.
I still suspect the vast silent majority of professional software devs haven’t integrated any, even Cursor-style, AI tools in to their main gig.
And I reckon that’s completely rational, for those that have made this choice explicitly.
Early adopters of AI tools are making a speculative bet, but so far most of them seem happy with the return.
This piece, however, only focuses on time spent on a task that could be done both ways. Even there, it falls short. Let's assume this study is correct and a specific coding task does take me 19% more time with AI. I can still be more productive because the AI doing some of the work allows me to do other tasks during that time.
I do worry about atrophy of my mind outsourcing too many tasks, admittedly. But that's a different issue.
It's similar to the story of the development of vehicles and how even though we move much faster we spend a greater amount of time in transit. My mom used to lament how annoying it was to have to drive to the grocery store because when she was younger and not everyone had cars the store came to you. Twice a day, in the morning and the evening the "rolling store" would drive through the neighborhood and if they didn't have what you needed right then, they would bring it on the next trip. We are finally coming back full circle with things like Instacart but it's taken a solid ~60 years of largely wasted inefficient travel times.
I think you're onto something with this take, based on my own experience. I definitely agree that my RPE seems lower when I'm using AI for things, whether it actually is making me more productive or not over the long term remains to be seen but things do certainly "feel" easier/less cognitively demanding. Which, tbh, is still a benefit even if it doesn't result in large gains in output. Putting in less cognitive load at work just conserves my energy for things that matter - everything else outside of $dayjob.
Also I do not think AI is falling apart and I do not think the productivity loss is surprising. It's something I've noticed a lot, you have to check AI output line by line thereby making you slower. AI makes tasks easier, not necessarily faster. And because people are lazy and love making things easier it's still going to be adopted no matter what anyone says.
They forgot about training and practice. Try it with airplanes. "The results surprised us. Given Boeing 777 they thought they will get from London to New York in 7 hours, but they actually couldn't even get off the ground. In fact they couldn't even start the engines."
If your "AI Assistant" is as hard to operate as a trans-Atlantic flight on a Boeing 777, perhaps it's not a very good assistant?
Since then, using Gen AI tools to learn + write code, I have deployed a functioning full stack application to the web. NextJS on Vercel, with a backend server also deployed running on Python, and a Supabase DB. Is it the best application ever making loads of money? Definitely not. Are there things wrong with it? Absolutely (although I promise I'm not exposing sensitive env vars and API keys to the web). Did the first versions look like absolute ass as I clumsily figured things out and made bad mistakes? You bet. But it's a functioning app that does some useful things and has real users.
I would never have imagined doing this in a million years prior to Gen AI.
Do some devs see mixed results depending on how they're using the tools? I'm sure. Is Gen AI overhyped broadly speaking? Probably so. But when I see people say it's a delusion, waste of resources, and everyone is wasting their time on it ... for me, it just doesn't line up.
Best hope ChatGPT plagiarized that code from people who properly sanitized user input, otherwise it might be vulnerable to SQL injection, XSS, etc. If such holes exist, it may be tough to resolve them without a professional.
> But when I see people say it's a delusion, waste of resources, and everyone is wasting their time on it ... for me, it just doesn't line up.
I don't think anyone would say LLMs are useless, but there are people out there comparing the birth of LLMs to the Industrial Revolution, and the advent of the Internet. Such claims are preposterous.
I think the widespread use of LLMs will ultimately be undone by a few factors:
1) Due to the way LLMs function, they will never be totally cured of 'hallucinations'. They are accurate a lot of the time, and apparently useful in those times. But you can never really trust its output. That is an exhausting problem, and will burn people out.
2) Prices for LLMs are artificially very low right now, these companies are burning investor money to keep going. The 'cost' is going to have to go way up for these LLM providers to be sustainable. I put cost in quotes because it will be some cost in subscription prices, but also cost in terms of ads that will be inserted everywhere, and other degrading forms of monetization.
3) Models are running out of good training data. They'll either hit a wall, or continue training on content that was itself LLM generated, which will go the way of Habsburg Jaw in a jiffy. Ensloppification is real.
I'm sure there will be niches for LLMs in the future, for those who can afford their true cost. But this cram-it-in-everything, available-everywhere frenzy will probably be looked back upon with deserved embarrassment.
What an utterly bizarre question. Yes, by definition being more productive means doing more work.
I've seen this firsthand. It's so easy to produce "content" now, like emails, presentations, etc., that more of my time now is just sifting through AI slop looking for signal.
I'm on a board for a nonprofit, and we've seen a person that is upset with the organization send dozens of AI-generated quasi-legal demands full of hallucinations of rules and laws that don't actually exist. So now we're forced to pay a real human lawyer to deal with a denial of service attack of ChatGPT slop and are spending lots of time just dealing with the onslaught. Our productivity has crashed as a board due to a single bad actor enabled by ChatGPT.
At work, I'm also getting buried in AI sales calls and AI resumes. It's getting more and more difficult to find signal in the noise.
It wouldn't surprise me if there's a lot of people who are experiencing lower productivity for similar reasons.
10% would mean a 10 trillion in productivity, 1% is 1 trillion.
It implies that even if some tasks are better with AI that it might not be so simple for us to judge which method is better for which task.
And even then, the web was over-hyped at various points in its life. And it took a terrible crash to sort out what was real from what wasn’t. It’s not that the web wasn’t absolutely transformative (it was!) but it wasn’t what a huge number of proponents thought it would be either.
https://sourcegraph.com/blog/the-brute-squad
You’ve seen the posts and articles: “I’m inherently better than AI, coding faster and smarter.” Sure, maybe for now—especially if you’re half-assing your AI game. But that edge? It’s vanishing fast.
Stop being so self indulgent.
Become proficient with the tools available to you or get left behind.