HN.zip

CTO Says 93% of Developers Use AI, but Productivity Is Still 10%

54 points by taubek - 60 comments
jdlshore [3 hidden]5 mins ago
This is self-reported productivity, in that devs are saying AI saves them about 4 hours per week. But let’s not forget the METR study that found a 20% increase in self-reported productivity but a 19% decrease in actual measured productivity.

(It used a clever and rigorous technique for measuring productivity differences, BTW, for anyone as skeptical of productivity measures as I am.)

samuelknight [3 hidden]5 mins ago
https://metr.org/blog/2025-07-10-early-2025-ai-experienced-o...

That info is from mid 2025, talking about models released in Oct 2024 and Feb 2025. It predates tools like Claude Code and Codex, Lovable was 1/3 current ARR, etc.

This might still be true but we desperately need new data.

williamcotton [3 hidden]5 mins ago
Has the METR study been replicated?
overgard [3 hidden]5 mins ago
You're only as fast as your biggest bottleneck. Adding AI to an existing organization is just going to show you where your bottlenecks are, it's not going to magically make them go away. For most companies, the speed of writing code probably wasn't the bottleneck in the first place.
blibble [3 hidden]5 mins ago
the amount of people that work in technology and have never heard of amdahl's law always shocks me

https://en.wikipedia.org/wiki/Amdahl's_law

a 100% increase in coding speed means I then I get to spend an extra 30 minutes a week in meetings

while now hating my job, because the only fun bit has been removed

"progress"

hnuser847 [3 hidden]5 mins ago
So if I'm understanding you correctly, prior to AI tools you spent 1 hour per week coding? And now you spend 30 minutes per week?
zht [3 hidden]5 mins ago
the number of people who have heard of Amdahl's law but don't know when to use "amount of X" vs "number of Y" always shocks me as well
qudat [3 hidden]5 mins ago
Agreed. The bottleneck is QA/Code review and that is never going away from most corps. I've never worked at a job in tech that didn't require code review and no, asking a code agent to review a PR is never going to be "good enough".

And here we are, the central argument for why code agents are not these job killing hype beasts that are so regularly claimed.

Has anyone seen what multi-agent code workflows produce? Take a look at openclaw, the code base is an absolute disaster. 500k LoC for something that can be accomplished in 10k.

ben_w [3 hidden]5 mins ago
> I've never worked at a job in tech that didn't require code review

I have. Sometimes the resulting code was much worse than what you get from an LLM, and yet the project itself was still a success despite this.

I've also worked in places with code review, where the project's own code quality architecture-and-process caused it to be so late to the market it was an automatic failure.

What matters to a business is ideally identical to the business metrics, which are usually not (but sometimes are) the code metrics.

8note [3 hidden]5 mins ago
i dont think thats actually the bottleneck?

the bottleneck is aligning people on what the right thing to do is, and fiting the change into everyone's mental models. it gets worse the more people are involved

oblio [3 hidden]5 mins ago
> Take a look at openclaw, the code base is an absolute disaster. 500k LoC for something that can be accomplished in 10k.

Mission accomplished: acquhire worth probably millions and millions.

I agree with you, by the way.

MYEUHD [3 hidden]5 mins ago
It was a hire not an acquihire. There was no acquisition.
co_king_5 [3 hidden]5 mins ago
I'm sorry but consider how many more edge cases and alternatives can be handled in 500k LoC as compared to that tiny 10k.

In the days of AGI, higher LoC is better. It just means the code is more robust, more adaptable, better suited to real world conditions.

vorticalbox [3 hidden]5 mins ago
one thing that aways slowed me down was writing jsdocs and testing.

Now i can write one example of a pass and then get codex to read the code and write a test for all the branches in that section saves time as it can type a lot faster than i can and its mostly copying the example i already have but changing the input to hit all the branches.

otabdeveloper4 [3 hidden]5 mins ago
> let's have LLMs check our code for correctness

Lmao. Rofl even.

(Testing is the one thing you would never outsource to AI.)

idle_zealot [3 hidden]5 mins ago
Outsourcing testing to AI makes perfect sense of you assume that tests exist out of obligation to meet some code coverage requirements, rather than to ensure correctness. Often I'll write a module and a few tests that cover its functionality, only for CI to complain that line coverage has decreased and reject my merge! AI to the rescue! A perfect job for a bullshit generator.
8note [3 hidden]5 mins ago
outsourcing testing the AI also gets its code to be connected to deterministic results, and show let the agent interact with the code to speculate expectations and check them against the actual code.

it could still speculate wrong things, but it wont speculate that the code is supposed to crash on the first line of code

ben_w [3 hidden]5 mins ago
> (Testing is the one thing you would never outsource to AI.)

I would rephrase that as "all LLMs, no matter how many you use, are only as good as one single pair of eyes".

If you're a one-person team and have no capital to spend on a proper test team, set the AI at it. If you're a megacorp with 10k full time QA testers, the AI probably isn't going to catch anything novel that the rest of them didn't, but it's cheap enough you can have it work through everything to make sure you have, actually, worked through everything.

LoganDark [3 hidden]5 mins ago
You don't use the LLM to check your code for correctness; you use the LLM to generate tests to exercise code paths, and verify that they do exercise those code paths.
onion2k [3 hidden]5 mins ago
And that test will check the code paths are run.

That doesn't tell you that the code is correct. It tells you that the branching code can reach all the branches. That isn't very useful.

menaerus [3 hidden]5 mins ago
In high-performance teams it is. In bike-shedding environments of course it is not.
outside1234 [3 hidden]5 mins ago
This. The key bottleneck in many organizations is the "socialize and align" on what to build. Or just "socialize and align" in general. :)
nasretdinov [3 hidden]5 mins ago
I think that over time people will start looking at AI-assisted coding the same way we now look at loosely typed code, or at (heavy) frameworks: it saves time in the short term, but may cause significant problems down the line. Whether or not this tradeoff makes sense in a specific situation is a matter of debate, and there's usually no obviously right or wrong answer.
doomslayer999 [3 hidden]5 mins ago
Once the free money runs out, the AI cos may shift to making heavily verified code snippets with more direct language control. This will heavily simplify a lot of boilerplate instead of fairytales of some AGI coding wiz.
co_king_5 [3 hidden]5 mins ago
Isn't the boilerplate that "AI" is capable of generating becoming more and more dated with each passing day?

Are the AI firms capable of retraining their models to understand new features in the technologies we work with? Or are LLMs going to be stuck generating C.A. 2022 boilerplate forever?

doomslayer999 [3 hidden]5 mins ago
No to the first question, and maybe with a lot of money for the second question.
mattmanser [3 hidden]5 mins ago
In the 20 years I've been in the industry, boiler plate has dropped dramatically in the backend.

Right now, front end has tons of boiler plate. It's one of the reasons AI hassle such a wow factor for FE, trivial tasks require a lot of code.

But even that is much better than it was 10 years ago.

That was a long way of saying I disagree with your no.

skydhash [3 hidden]5 mins ago
FE has a lot of boilerplate only if you’re starting from scratch every single time. That’s why we had template systems and why we invented view libraries. Once you’ve defined your libraries, you just copy-paste stuff.
matthewbauer [3 hidden]5 mins ago
It seems like they should be able to “overweight” newer training data. But the risk is the newer training data is going to skew more towards AI slop than older training data.
otabdeveloper4 [3 hidden]5 mins ago
There won't ever be newer training data.

The OG data came from sites like Stackoverflow. These sites will stop existing once LLMs become better and easier to use. Game over.

esclerofilo [3 hidden]5 mins ago
Every time claude code runs tests or builds after a change, it's collecting training data.
co_king_5 [3 hidden]5 mins ago
Has Anthropic been able to leverage this training data successfully?
esclerofilo [3 hidden]5 mins ago
I can't pretend to know how things work internally, but I would expect it to be involved in model updates.
otabdeveloper4 [3 hidden]5 mins ago
You need human language programming-related questions to train on too, not just the code.
8note [3 hidden]5 mins ago
thats what the related chats are for?
ptx [3 hidden]5 mins ago
Apparently "AI is speeding up the onboarding process", they say. But isn't that because the onboarding process is about learning, and by having an AI regurgitate the answers you can complete the process without learning anything, which might speed it up but completely defeats the purpose?
mjfisher [3 hidden]5 mins ago
I think there's definite scope for that being true; not because you can start doing stuff before you understand it (you can), but because you can ask questions of a codebase your unfamiliar with to learn about it faster.
raphman [3 hidden]5 mins ago
Yes, that's how I'd interpret it, too.

According to the article, onboarding speed is measured as “time to the 10th Pull Request (PR).”

As we have seen on public GitHub projects, LLMs have made it really easy to submit a large number of low-effort pull requests without having any understanding of a project.

Obviously, such a kind of higher onboarding speed is not necessarily good for an organization.

OptionOfT [3 hidden]5 mins ago
Correct. Reading code is important. The details are in the minutia, and the way code works is that the minutia are important.

Summarizing this with AI makes you lose that context.

snsjzhhz [3 hidden]5 mins ago
This has been my experience as a dev, and it always confuses me when people say they prefer to work at a “higher level”. The minutiae are often just as important as some of the higher level decisions. Not everything, but not an insignificant portion either. This applies to basic things like correctness, performance, and security - craft, style, and taste are not involved.
co_king_5 [3 hidden]5 mins ago
> This has been my experience as a dev, and it always confuses me when people say they prefer to work at a “higher level”.

> The minutiae are often just as important as some of the higher level decisions.

Frankly, a failure to understand this is a tell that someone is not equipped to evaluate code quality.

xeiotos [3 hidden]5 mins ago
Unsurprising for multiple reasons. Most organizations have other bottlenecks and limiting factors than “how fast can you develop”.

Regardless, if you’re a dev who is now 2x as productive in terms of work completed per day, and quality remains stable, why should this translate to 2x the output? Most people are paid by the hour and not for outcomes.

And yes, I am suggesting that if you complete in 4 hours that which took you 8 hours in 2019, that you should consider calling it a day.

ilovetux [3 hidden]5 mins ago
I think some AI companies are just now starting to feel the pressure to profit.

Soon, I predict we will see a pretty significant jump in price that will make a 10% productivity gain seem tiny compared to the associated bills.

For now, these companies are trying to reach critical mass so their users are so dependant on their tech that they have to keep paying at least in the short term.

bluejekyll [3 hidden]5 mins ago
I found the title for this post misleading. To clarify it a bit, AI has only improved productivity by 10% even though 93% of devs are using it.
dandanua [3 hidden]5 mins ago
Yeah, the title may suggest that productivity is still 10% out of 100% after CEOs fired half of developers believing that the rest will do all the job with the help of AI.
onion2k [3 hidden]5 mins ago
A 10% uplift in productivity for the cost of probably 0.001% of the salary budget is an incredible success.
arctic-true [3 hidden]5 mins ago
This is exactly right. And assuming organizations use the gains to cut headcount rather than boost total productivity, a 10% reduction in white collar employment would still be an era-defining systemic shock to the economy.
emp17344 [3 hidden]5 mins ago
Productivity improvements from automation actually result in an increase in jobs, not fewer jobs. Basic economics.
agentifysh [3 hidden]5 mins ago
I read this article as the CTO being the bottleneck if he's only seeing 10% productivity boost at his organization.

I dont think this is a purely AI problem more with the legacy costs of maintaining many minds that can't be solved by just giving people AI tools until the AI comes for the CTO role (but not CEO or revenue generating roles) too and whichever manager is bottlenecking.

I imagine a future where we have Nasdaq listed companies run by just a dozen people with AI agents running and talking to each other so fast that text becomes a bottleneck and they need another medium that can only be understood by an AI that will hold humans hand

This shift would also be reflected by new hardware shifts...perhaps photonic chips or anything that lets AI scale up crazy without the energy cost....

Exciting times are ahead AI but it's also accelerating digital UBI....could be good and bad.

Nezteb [3 hidden]5 mins ago
> it's also accelerating digital UBI

Do you have sources for this claim?

chvid [3 hidden]5 mins ago
As far as I can tell from my workplace the total impact on productivity is neutral to negative.
rcfox [3 hidden]5 mins ago
The title is misleading. Productivity isn't at 10%, it's at 110%.
moralestapia [3 hidden]5 mins ago
AI adoption has reduced productivity at my workplace, and by a noticeable amount!
randomtoast [3 hidden]5 mins ago
This will lead to natural selection. As AI becomes increasingly integrated into all areas, companies that manage it less effectively than others will face greater selection pressure.
otabdeveloper4 [3 hidden]5 mins ago
That's expected for any new "low-code" solution du jour.
mytailorisrich [3 hidden]5 mins ago
Blunt opinion: Most devs are not that good and really only execute what they are told to do.

The threat of AI for devs, and the way to drastically improve productivity is there: keep the better devs who can think systemically, who can design solutions, who can solve issues themselves and give them all the AI help available, cut the rest.

tyleo [3 hidden]5 mins ago
That’s how I feel too. When I was an architect at a ~300-person company, a big chunk of my job shifted to reviews, technical design docs, and guidance. I’m getting great results by feeding context like that into Claude Code, then reviewing and steering what it produces.

It really does feel like a multiplier on me and I understand things enough to get my hands dirty where Claude struggles.

Lately I’ve been wondering if that role evolves into a more hierarchical review system: senior engineers own independent modules end-to-end, and architects focus on integration, interfaces, and overall coherence. Honestly, the best parts of our product already worked like that even before AI.

gedy [3 hidden]5 mins ago
I can see where productivity could be higher if all I did was type in programs to some spec, or bootstrapping new apps all day - but that's like not the reality of "programming", at least for me past 25 years. Sorting through what to even make and interpreting "requirements" is what takes the most time
downrightmike [3 hidden]5 mins ago
Yeah, industry has told them that devs aren't valuable and AI can do their job. Who TF has motivation after that?
co_king_5 [3 hidden]5 mins ago
No motivation? I'm sorry buddy but your ass is getting replaced by Claude Code in the next 3-6 weeks.