I see the same with AI. Some companies will use AI quietly and productively without much fuzz. Others are just using it as a marketing tool or an ego trip by execs but no real understanding.
First, I find that I'm using a lot fewer libraries in general because I am less constrained by the mental models imposed by library authors upon what I'm actually trying to do. Libraries are often heavy and by nature abstract low-level calls from API. These days, I'm far more likely to have 2-3 functions that make those low-level calls directly without any conceptual baggage.
Second, I am generalizing but a reasonable assertion can be made that publishing a package is implicitly launching an open source project, however small in scope or audience. Running OSS projects is a) extremely demanding b) a lot of pain for questionable reward. When you put something into the universe you're taking a non-zero amount of responsibility for it, even just reputationally. Maintainers burn out all of the time, and not everyone is signed up for that. I don't think there's going to be anything remotely like a 1:1 Venn for LLM use and package publishing.
I would counter-argue that in most cases, there might already be too many libraries for everything under the sun. Consolidation around the libraries that are genuinely amazing is not a terrible thing.
Third, one of the most recurring sentiments in these sorts of threads is that people are finally able to work through the long lists of ideas they had but would have never otherwise gotten around to. Some of those ideas might have legs as a product or OSS project, but a lot of them are going to be thought experiments or solve problems for the person writing them, and IMO that's a W not an L.
Fourth, once most devs are past the "vibe" party trick phase of LLM adoption, they are less likely to squat out entire projects and far, far more likely to return to doing all of the things that they were doing before; just doing them faster and with less typing up-front.
In other words, don't think project-level. Successful LLM use cases are commit-level.
a) there are likely many more active, eager contributors all of a sudden, and
b) there's suddenly a huge amount of new papers published every week about algorithms and techniques that said contributors then eagerly implement (usually of dubious benefit).
More cynically, one might also hypothesize that
c) code quality has dropped, so more frequent releases are required to fix broken programs.
I have a news feed, work tab for managing issues/PRs, markdown editor with folders, calendar, AI powered buttons all over the place (I click a button, it does something interesting with Claude code I can't do programmatically).
Why don't I share it? Because it's highly personal, others would find it doesn't fit their own workflow.
Before AI for the last 8 or so years now first at a startup then working in consulting mostly with companies new to AWS or they wanted a new implementation, it’s been:
1. Gather requirements
2. Do the design
3. Present the design and get approval and make sure I didn’t miss anything
4. Do the infrastructure as code to create the architecture and the deployment pipeline
5. Design the schema and write the code
6. Take it through UAT and often go back to #4 or #5
7. Move it into production
8. Monitoring and maintenance.
#4 and #5 can be done easily with AI for most run of the mill enterprise SaaS implementations especially if you have the luxury of starting from the ground up “post AI”. This is something you could farm off to mid level ticket takers before AI.
But there's not one tool there that triggered a major boost in output or number of apps / libraries / products created - unless I missed something.
Sure, total output has increased, especially since the early 2010's thanks to both Github becoming the social network of software development, and (arguably) Node / JS becoming one of the most popular languages/runtimes out there attracting a lot of developers to publish a lot of tools. But that's not down to productivity or output boosting developments.
But it requires A LOT of work to make sure it is actually safe for people and organizations. And no, an .md file saying “PLEASE DONT PWN ME, KTHX” isn’t it at all. “Alignment” is only part of the equation.
If you’re not afraid to dive into rabbitholes, here is how it works: http://community.safebots.ai/t/layer-4-browser-extensions-pe...
And you can even see the number of new games that disclosed using generative AI (~21% in 2025). [2]
And that's probably significantly undercounting because I doubt everyone voluntarily discloses when they use tools like Claude Code (and it's not clear how much Valve cares about code-assistance). [3]
Also no one is buying or playing a lot of these games.
[1] https://steamdb.info/stats/releases/
[2] https://steamdb.info/stats/releases/?tagid=1368160
[3] https://store.steampowered.com/news/group/4145017/view/38624...
Yup. And for most purposes, that's enough. An app does not have to be productized and shipped to general audience to be useful. In fact, if your goal is to solve some specific problem for yourself, your friends/family, community or your team, then the "last step" you mention - the one that "takes majority of time and effort" - is entirely unnecessary, irrelevant, and a waste of time.
The productivity boost is there, but it's not measured because people are looking for the wrong thing. Products on the market are not solutions to problems, they're tools to make money. The two are correlated, because of bunch of obvious reasons (people need money, solving a problem costs money, people are happy to pay for solutions, etc.), but they're still distinct. AI is dropping the costs of "solving the problem" part, much more than that of "making a product", so it's not useful to use the lack of the latter as evidence of lack of the former.
Having worked extensively with vibe-coded software, the main problem for me is that I have tuned-off from the ai-code, and I dont see any skin-in-the-game for me. This is dangerous because it becomes increasingly harder to root-cause and debug problems because that muscle is atrophying. use-it or lose-it applies to cognitive skills (coding/debugging). Now, I lean negatively to ai-code because, while it seduces us with fast progress in the first 80%, the end outcome is questionable in terms of quality. Finally, ai-coding encourages a prompt-and-test or trial-and-error approach to software engineering which is frustrating and those with experience would prefer to get it right by design.
Just doesn't have the same ring to it.
This is true, and I bet there are thousands of people who are in this stage right now - having gotten there far faster than they would have without Claude Code - which makes me predict that the point made in the article will not age well. I think it’s just a matter of a bit more time before the deluge starts, something on the order of six more months.
-- we had a terrible time building something so now we're only going to buy things
-- we had a terrible time buying something so now we're only going to build things
-- repeat...
Either way you can have a brilliant success and either way you fail abjectly, usually you succeed at most but not all of the goals and it is late and over budget.
If you build you take the risks of building something that doesn't exist and may never exist.
If you buy you have to pay for a lot of structure that pushes risks around in space and time. The vendor people needs marketing people not to figure out what you need, but what customers need in the abstract. Sales people are needed to help you match up your perception of what you need with the reality of the product. All those folks are expensive, not just because of their salaries but because a pretty good chunk of a salesperson's time is burned up on sales that don't go through, sales that take 10x as long they really should because there are too many people in the room, etc.
When I was envisioning an enterprise product in the early 2010s for instance I got all hung up on the deployment model -- we figured some customers would insist on everything being on-premise, some would want to host in their own AWS/Azure/GCP and others would be happy if we did it all for them. We found the phrase "hybrid cloud" would cause their eyes to glaze over and maybe they were right because in five years this became a synonym for Kubernetes. Building our demos we just built things that were easy for us to deploy and the same would be true for anything people build in house.
To some extent I think AI does push the line towards build.
1. Debugging my own code or obvious behavior with other libraries.
2. Debugging pain-in-the-ass behavior with other libraries.
My patience with the latter is significantly less now, and so is perhaps my skill in debugging them. Libraries that change their apis for no apparent reason, libraries which use nonstandard parameter names, libraries which aren’t working as advertised.
I've started tons of scratch my own itch projects. There's adoption, UX, onboarding costs even if you're the only audience.
TLDR: i don't even use my own projects. I churn.
IMHO (this may not apply to you!) a lot of people launch a "competitor" of a product which seems to be a clone of the product without improving something that the other product misses/is very bad at.
I do think there would be value in sharing your setup at some point if you get around to it, I think a lot of builders are in the same boat and we're all trying to figure out what the right interface for this is (or at least right for us personally).
0. It runs way too fast and far ahead. You need to slow it down, force planning only and explicitly present a multi-step (i.e. numbered plan) and say "we'll do #1 first, then do the rest in future steps".
take-away: This is likely solved with experience and changing how I work - or maybe caring less? The problem is the model can produce much faster than you can consume, but it runs down dead ends that destroy YOUR context. I think if you were running a bunch of autonomous agents this would be less noticeable, but impact 1-3 negatively and get very expensive.
1. lots of "just plain wrong" details. You catch this developing or testing because it doesn't work, or you know from experience it's wrong just by looking at it. Or you've already corrected it and need to point out the previous context.
take-away: If you were vibe coding you'd solve all these eventually. Addressing #0 with "MORE AI" would probably help (i.e. AI to play/validate, etc).
2. Serious runtime issues that are not necessarily bugs. Examples: it made a lot of client-side API endpoints public that didn't even need to exist, or at least needed to be scoped to the current auth. It missed basic filtering and SQL clauses that constrained data. It hardcoded important data (but not necessarily secrets) like ports, etc. It made assumptions that worked fine in development but could be big issues in public.
take-away: AI starts to build traps here. Vibe coders are in big trouble because everything works but that's not really the end goal. Problems could range from 3am downtime call-outs to getting your infrastructure owned or data breaches. More serious: experienced devs who go all-in on autonomous coding might be three months from their last manual code review and be in the same position as a vibe coder. You'd need a week or more to onboard and figure out what was going on, and fix it, which is probably too late.
3. It made (at least) one huge architectural mistake (this is a pretty simple project so I'm not sure there's space for more). I saw it coming but kept going in the spirit of my experiment.
take-away: TBD. I'm going to try and use AI to refactor this, but it is non trivial. It could take as long as the initial app did to fix. If you followed the current pro-AI narrative you'd only notice it when your app started to intermittently fail - or you got you cloud provider's bill.
To a point, but I think this overstates it by quite a bit. At the moment I'm weighing some tradeoffs around this myself. I'm currently making an app for a niche interest of mine. I have a few acquaintances who would find it useful as well but I'm not sure if I want to take that on. If I keep the project for personal use I can make a lot of simplifying decisions like just running it on my own machine and using the CLI for certain steps.
To deploy this to for non-tech users I need to figure out a whole deployment approach, make the UI more polished, and worry more about bugs and uptime. It sucks to get invested in some software that then constantly starts breaking or crashing. GenAI will help with this somewhat, but certainly won't drop the extra coding time cost down to zero.
But based on the hype (100x productivity!), there should be a deluge of high quality mobile apps, Saas offerings, etc. There is a huge profit incentive to create quality software at a low price.
Yet, the majority of new apps and services that I see are all AI ecosystem stuff. Wrappers around LLMs, or tools to use LLMs to create software. But I’m not really seeing the output of this process (net new software).
Emacs with Hyperbole[0]?
Do you never open a code editor?
1) I'm a former SWE in a business role at a small-market publishing company. I've used Claude Code to automate boring processes that previously consumed weeks of our ops and finance teams' time per year. These aren't technically advanced, but previously would have required in-house dev talent that would not have been within reach of small businesses. I wouldn't have had the time to code these things on my own, but with AI assistance the time investment is greatly reduced (and mostly focused on QA). The only needle moved here is on a private Github repo, but it's real shipped code with small but direct impact.
2) I used to often find myself writing simple Perl wrappers to various APIs for personal or work use. I'd submit these to CPAN (Perl's equivalent to PyPI) in case anyone else could use them to save the 30-60 minutes of work involved. These days I don't bother -- most AI tools can build these in a matter of seconds; publishing them to CPAN or even Github now feels like unnecessary cruft, especially when they're likely to go without active maintenance. So, my LOC published to public repos is down, even though the amount of software produced is the same. It's just that some of that software has become less useful to the world writ large.
3) The code that's possible to ship quickly with pure AI (vibe coding) is by definition not the kind of reusable code you'd want to distribute on PyPI. So, I'd expect that any productivity impact from AI on OSS that's designed to be reusable would be come very slowly, versus "hockey stick" impact.
200 years ago text was much more expensive, and more people memorized sayings and poems and quotations. Now text is cheap, and we rarely quote.
In the last few months, Gemini (and I) have written for highly personal, very niche apps that are perfect for my needs, but I would never dream of releasing. Things like cataloguing and searching my departed mom‘s recipe cards, or a text message based budget tracker for my wife and I to share.
These things would never be released or available as of source or commercial applications in the way that I wanted them, and it took me less time to have them built with AI then it would have taken me to Research existing alternatives and adapt my workflow/use case to fit whatever I found.
So yeah, there are more apps but I would venture to say you’ll never see most of them…
I think this might be more of an comment on software as a business than AI not coding good apps.
Pre-ChatGPT, in ~2020, there were about 5,000 new packages per month. Starting in 2025 (the actual year agents took off), there is a clear uptick in packages that is consistently about 10,000 or 2X the pre-ChatGPT era.
In general, the rate of increase is on a clear exponential. So while we might not see a step change in productivity, there comes a point where the average developer is in fact 10X productive than before. It just doesn't feel so crazy because it can about in discrete 5% boosts.
I also disagree with the dataset being a good indicator of productivity. I wouldn't actually suspect the number of packages or the frequency of updates to track closely with productivity. My first order guess would that AI would actually be deflationary. Why spend the time to open source something that AI can gen up for anyone on a case by case basis specific to the project. it takes a certain level of dedication and passion for a person to open source a project and if the AI just made it for them, then they haven't actually made the investment of their time and effort to make them feel justified in publishing the package.
The metrics I would expect to go up are actually the size of codebases, the number of forks of projects that create hyper customized versions of tools and libraries, and other metrics like that.
Overall, I'd predict AI is deflationary on the number of products that exist. If AI removes the friction involved with just making a custom solution, then the amount of demand for middleman software should actually fall as products vertically integrate and reduce dependencies.
But I have been absolutely flooded with trailers for new and upcoming indie games. And at least one indie developer has admitted that certain parts of their game had used the aide of AI.
I also noticed sometimes when I think of writing something, I ask AI first if it exists, and AI throws up some link and when I check the link it says "made with <some AI>".
So I'm not sure what author is trying to say here but I definitely feel like I am noticing a rise in software output due to AI.
But with that said, I also am noticing the burden of taking care of those open source projects. Sometimes it feels like I took on a 2nd job.
I think a lot of software is being produced with AI and going unnoticed, they don't all end up on the front page of HN for harassing developers.
It seems like all tech executives are saying they are seeing big increases in productivity among engineering teams. Of course everyone says they're just [hyping, excusing layoffs, overhired in 2020, etc], but this would be the most relevant metric to look at I think.
But the mistake is to expect a huge productivity boost.
This is highly related to Amdahl's law, also The Mythical Man-Month.
Some tasks can be accomplished so fast that it seems magical, but the entire process is still very serial, architecture design and debug are pretty weak on the AI side.
And even “product engineers” often do not have experience going from zero to post sales support on a saas on their own.
It is a skill set of its own to make product decisions and not only release but stick with it after the thing is not immediately successful.
The ability to get some other idea going quickly with AI actually works against the habits needed to tough through the valley(s).
Yes, you do produce more code. But LoC produced is never a healthy metric. Reviewing the LLM generated code, polishing the result and getting it to production-level quality still very much requires a human-in-the-loop with dedicated time and effort.
On the other hand, people who vibe code and claims to be 10x productive, who produces numerous PRs with large diffs usually bog down the overall productivity of teams by requiring tenuous code reviews.
Some of us are forced to fast-track this review process so as to not slow down these "star developers" which leads to the slow erosion in overall code quality which in my opinion would more than offset the productivity gains from using the AI tools in the first place.
A lot of the real value shows up as workflow compression instead. Internal tools, one-off automations, bespoke research flows, coding helpers, things that would never have justified becoming a product in the first place.
Besides, it's working for me. If it isn't working for others I don't want to convince them of anything. I do want to hear from other people for whom it's working, though, so I'm happy to share when things work for me.
The tools for generating AI code aren't yet capable of producing code that is decent enough for general purpose use cases, with good robust tests, and clean and quality.
Arguably makes the remaining 20% even harder to handle.
I'm sure that AI can be a huge boost to great, mature developers. Which are insanely rare in an industry that has consistently promoted brainless ivy league coders farming algo quizzes for months.
But those with a huge sensibility and experience can definitely be enabled to produce more.
But the 20% is still there and again, it's easy to make it way harder because you're less intimate with the brittle 80%.
While it’s interesting to see that in open source software the increase is not dramatic, this ignores however many people are now gen-coding software they will never publish just for them, or which winds up on hosting platforms like Replit.
This is more useful for discussing what kind of projects AI is being used for than whether it's being used.
Try it out and don't stop trying. If something improves at this rate, even if you think it's not there right now, don't assume it is going to stop. Be honest about the things we were always obviously bad at, that the ai has been getting quickly better at, and assume that it will continue getting better. If this were true, what would that mean for you?
Are there any agentic sales and marketing offerings?
Because being able to reliably hand off that part of the value chain to an agent would close a real gap. (Not sure this can be done in reality)
Of course these are specific workplaces designed around moving tickets on a board, not high-agentic, fast-moving startups or independent projects—but they might represent a lot of the developer workforce.
I also know this is not everyone's experience and probably a rare favorable outcome of productivity gain captured by a worker that is not and won't stay the norm.
I have no plans of publishing them or making the open source, so it will not be a part of this metric. I believe others are doing this too.
1. People aren't creating new apps, but enhancing existing ones
2. Companies are less likely to pay for new offerings when the barrier to entry is lowered due to AI. They'll just vibe code what they need.
If someone did make a mobile app, how would it get up take? Coding has never been the hard part about a successful software product.
Because it's better to sell shovels than to pan for gold.
In the current state of LLMs, the average no-experience, non-techy person was never going to make production software with it, let alone actually launch something profitable. Coding was never the hard part in the first place, sales, marketing & growth is.
LLMs are basically just another devtool at this point. In the 90s, IDEs/Rapid App Development was a gold rush. LLMs are today's version of that. Both made developer's life's better, but neither resulted in a huge rush of new, cheap software from the masses.
Problem is that all these companies trying to push AI experiences know that giving users unfettered access to their data to build further customization is corporate suicide.
Oh and sadly, llm’s are useless for the imaginative part too. Shucks eh.
I had to create a bunch of GitHub and Linear apps. Without me even asking Codex whipped up a web page and a local server to set them up, collecting the OAuth credentials, and forward them to the actual app.
Took two minutes, I used it to set up the apps in three clicks each, and then just deleted the thing.
Code as transient disposable artifacts.
This mix of doing my main work on complex stuff (healthcare) with heavy AI input, and then having 1-2 agents building lighter tools on the side, has been surprisingly effective.
Not OP, just guessing.
The real output is happening in private repos, internal tools, and single-purpose apps that never get published anywhere. I've been building a writing app as a side project. AI got me from zero to a working PWA with offline support, Stripe integration, and 56 SEO landing pages in about 6 weeks of part-time work. Pre-AI that's easily a 6-month project for one person.
But I'm never going to publish it as a PyPI package. It's a deployed web app. The productivity gain is real, it just doesn't show up in the datasets this article is looking at.
The iOS App Store submission data (24% increase) that someone linked in the comments is a much better signal. That's where the output is actually landing.
It's like looking at tire sales to wonder about where the EV cars are.
Apparently new iOS app submissions jumped by 24% last year:
> According to Appfigures Explorer, Apple's App Store saw 557K new app submissions in 2025, a whopping 24% increase from 2024, and the first meaningful increase since 2016's all-time high of 1M apps.
The chart shows stagnant new iOS app submissions until AI.
Here's a month by month bar chart from 2019 to Feb 2026: https://www.statista.com/statistics/1020964/apple-app-store-...
Also, if you hang out in places with borderline technical people, they might do things like vibe-code a waybar app and proudly post it to r/omarchy which was the first time they ever installed linux in their life.
Though I'd be super surprised if average activity didn't pick up big on Github in general. And if it hasn't, it's only because we overestimate how fast people develop new workflows. Just by going by my own increase in software output and the projects I've taken on over the last couple months.
Finally, December 2025 (Opus 4.5 and that new Codex one) was a big inflection point where AI was suddenly good enough to do all sorts of things for me without hand-holding.
YoloSwag (13 commits)
[rocketship rocketship rocketship]
YoloSwag is a 1:1 implementation of pyTorch, written in RUST [crab emoji]
- [hand pointing emoji] YoloSwag is Memory Safe due to being Written in Rust
- [green leaf emoji] YoloSwag uses 80% less CPU cycles due to being written in Rust
- [clipboard emoji] [engineer emoji] YoloSwag is 1:1 API compatible with pyTorch with complete ops specification conformance. All ops are supported.
- [recycle emoji] YoloSwag is drop-in ready replacement for Pytorch
- [racecar emoji] YoloSwag speeds up your training workflows by over 300%
Then you git clone yoloswag and it crashes immediately and doesn't even run. And you look at the test suite and every test just creates its own mocks to pass. And then you look at the code and it's weird frankenstein implementation, half of it is using rust bindings for pytorch and the other half is random APIs that are named similarly but not identical.
Then you look at the committer and the description on his profile says "imminentize AGI.", he launched 3 crypto tokens in 2020, he links an X profile (serial experiments lain avatar) where he's posting 100x a day about how "it's over" for software devs and how he "became a domain expert in quantum computing in 6 weeks."
Like I've been making things, and making changes to things, but I haven't published any of that because, well they're pretty specific to my needs. There are also things which I won't consider publishing for now, even if generally useful because, well the moat has moved from execution effort to ideas, and we all want to maintain some kind of moat to boost our market value (while there's still one). Everyone has reasonable access to the same capabilities now, so everyone can reasonably make what they need according to their exact specs easily, quickly and cheaply.
So while there are many things being made with AI, there is ever-decreasing reasons to publish most of it. We're in an era of highly personalized software, which just isn't worth generalizing and sharing as the effort is now greater than creating from scratch or modifying something already close enough.
IE, using agents to iterate through many possible approaches, spike out migrations, etc might save a project a year of misadventures, re-designs, etc, but that productivity gain _subtracts_ the intermediate versions that _didn't_ end up being shipped.
As others have mentioned, I think yak-shaving is now way more automated. IE, If I want to take a new terminal for a spin, throw together a devtool to help me think about a specific problem better, etc, I can do it with very low friction. So "personal" productivity is way higher.
Also using PyPI as a benchmark is incredibly myopic. Github's 2025 Octoverse[0] is more informative. In that report, you can see a clear inflection point in total users[1] and total open source contributions[2].
The report also notes:
> In 2025, 81.5% of contributions happened in private repositories, while 63% of all repositories were public
[0]: https://github.blog/news-insights/octoverse/octoverse-a-new-...
[1]: https://github.blog/wp-content/uploads/2025/10/octoverse-202...
[2]: https://github.blog/wp-content/uploads/2025/10/octoverse-202...
Personally, I see the paid or adware software market shrinking, not growing, as a testament to the success of LLMs in coding.
I’ve done some experiments with reading gedcom files, and I think I’m quite close to a demoable version of a genealogy app.
Biggest thing is a tool for remotely working musicians. It’s about 10000 lines of well written rust, it is a demoable state and I wish I could work more on it but I just started a new job.
But yeah, this wouldn’t have been possible if I hadn’t been a very experienced dev who knows how to get things live. Also I’ve found a way to work with LLMs that works for me, I can quickly steer the process in the right way and I understand the code thats written, again it’s possible that a lot of real experience is needed for this.
> So, let’s ask again, why? Why is this jump concentrated in software about AI?...Money and hype
The AI field right now is drowning in hype and jumping from one fad to another.Don't get me wrong: there are real productivity gains to be had, but the reality is that building small one-offs and personal tools is not the same thing as building, operationalizing, and maintaining a large system used by paying customers and performing critical business transactions.
A lot of devs are surrendering their critical thinking facilities to coding agents now. This is part of why the hype has to exist: to convince devs, teams, and leaders that they are "falling behind". Hand over more of your attention (and $$$) to the model providers, create the dependency, shut off your critical thinking, and the loop manifests itself.
The providers are no different from doctors pushing OxyContin in this sense; make teams dependent on the product. The more they use the product, the more they build a dependency. Junior and mid-career devs have their growth curves fully stunted and become entirely reliant on the LLM to even perform basic functions. Leaders believe the hype and lay off teams and replace them with agents, mistaking speed for velocity. The more slop a team codes with AI, the more they become reliant on AI to maintain the codebase because now no one understands it. What do you do now? Double down; more AI! Of course, the answer is an AI code reviewer!. Nothing that more tokens can't solve.
I work with a team that is heavily, heavily using AI and I'm building much of the supporting infrastructure to make this work. But what's clear is that while there are productivity gains to be had, a lot of it is also just hype to keep the $$$ flowing.
Fans of vibecoding and agentic tools say they are 2x as productive, 10x as productive – maybe 100x as productive! Someone built an entire web browser from scratch. Amazing!
So, skeptics reasonably ask, where are all the apps? If AI users are becoming (let’s be conservative) merely 2x more productive, then where do we look to see 2x more software being produced? Such questions all start from the assumption that the world wants more software, so that if software has gotten cheaper to make then people will make more of it. So if you agree with that assumption, then where is the new software surplus, what we might call the “AI effect”?
We’ll look at PyPI, the central repository for Python packages. It’s large, public, and consistently measured, so we should expect to see some AI effect there.
There it is, see it? The release of ChatGPT. Does it look like an epochal revolution of software productivity on the upper chart? No.
There are a few spikes in the lower chart showing new packages/month, in what you might call the “AI era” of 2020 onward. But those reflect spam and malware floods, not genuine package creation.1

Two-panel chart showing PyPI total packages growing exponentially to 800k and new packages per month fluctuating around 5-15k, with ChatGPT release marked showing no obvious inflection point
This is curious. If AI is making software engineers more productive, why aren’t they producing more software?
But, you might say, package creation is not the right measure. Anyone can create and upload a “package” which is nothing but a hello world demo. This is always easier than creating something durable which people actually use. We want to look at “real” packages, packages which are actually downloaded, used, and maintained over time.
Okay, so let’s consider a different chart. We start by gathering the 15,000 most downloaded Python packages on PyPI in December 2025.2 Then we split the packages into cohorts based on their birth-year, and for each cohort we plot their median release frequency over time.3 This seems like a reasonable proxy measure of the production of real, actively-used software.
To show one cohort’s release frequency over time, we draw a line. So in the chart below, every line starts with a point showing the number of update releases within the first 12 months of the life of a package born in that year. The line proceeds as the package ages.
So what do we see? Do packages get updated more frequently after the advent of ChatGPT?

Well … sort of?
We clearly see that packages born after ChatGPT were updated more frequently within their first year (13 releases/year) than packages born back in 2014 (6 releases/year). This is seen in the fact that the cohort life lines start higher over time.
But this looks like it’s continuing a trend which starts too early to be attributed to an AI productivity boost. First-year release frequency started increasing in 2019 (at 10 releases/year), well before modern AI coding tools appeared. This seems just as likely to be due to growing adoption of continuous integration tools like GitHub Actions, which have been around longer.
Another reason to doubt this increase is entirely due to AI is the other effect visible in this chart, which is that packages are released less frequently as they get older. This is seen in the fact that all of the cohort life lines decrease over time. That has not changed. In other words, people are not using AI in a way that leads them to update a package more frequently as it ages.
But surely some of that increase in initial release frequency is due to an AI boost? Let’s look deeper.
Let’s split packages by whether they’re about AI or not, by classifying based on the package’s description.4 There can we see an AI effect?

There it is! Or at least, there’s something!
Packages which are not about AI look much more like their pre-ChatGPT era cohorts, in that they show the same modest secular trend of increasing releases per year.
But in contrast, the packages which are about AI show a dramatic increase in release frequency. For example, the packages first-released in 2023 about AI reached a median of 20 releases in their first 12 months. Almost 2x their non-AI counterparts in the same year.
In short, for some reason, newly created packages about AI are being updated much more frequently.
Of course, AI is very popular right now. When we see that packages about AI are updated more frequently, are we merely observing that popular packages are updated more frequently?
To address that question, let’s do one more split. Let’s take our initial group of the top 15,000 packages by download in December 2025, and split it into two groups, the more popular 7,500 and the less popular 7,500.
Was our observation regarding packages “about AI” merely an observation regarding popularity?

No. The top-right quadrant jumps out: popular AI packages jumped to 21-26 median releases per year post ChatGPT, more than double the ~10 that popular non-AI packages have held steady at (and also significantly more than the less popular AI packages).
So we do see a >2x effect in release frequency, and it’s concentrated in the most popular packages about AI specifically.
But of course the interesting question is, why?
Before considering what’s causing this, let’s recap the evidence:
There is no obvious increase in the rate of package creation as a whole, post-ChatGPT, and only a marginal increase in the rate of package updates as a whole.
There is a small, steady increase in update frequency over the years, but this trend predates ChatGPT.
There is a large (>2x) increase in update frequency for popular AI packages, and a smaller bump for less popular AI packages.
If we ask why we see this pattern of evidence, we discover that it’s actually adequate to let us conclude that some things are not happening, and to suggest some plausible interpretations for what is going on.5
Is AI massively boosting developer productivity across the board?
No. We are not seeing indications that developers as a whole are 100x or even 10x more productive. The bumper crop of new packages, or new package updates, just does not exist!
Relax. You are not missing a party that literally everyone else was invited to.
Are some developers building much faster, by using AI?
Perhaps? But the visible aggregate effect is still so modest, that if some devs are getting this big boost, there certainly aren’t many of them. Or else the purported boost is not really that big. What we see in aggregate is hardly any uptick in package update frequency.
However, we do see a boost in newly-created popular packages about AI.
Are people building an enormous amount of software for using AI?
Yes, yes they are. The jump in update frequency for recent packages about AI is really the headline effect here. The narrowness of this effect is the puzzle that needs to be explained.
So, let’s ask again, why? Why is this jump concentrated in software about AI? We do have two hypotheses:
AI “skill issue”. Maybe people building AI tools are also the ones most likely to know how to use AI effectively. This would produce a bigger productivity boost for AI packages. But if skill alone explained the jump, we’d expect it across all AI packages. Instead, the 2x2 chart shows it’s concentrated in the most popular ones, which suggests something else is also at play.
Money and hype 🤑💰. An enormous amount of funding and enthusiasm has flowed into AI, and it is being converted into (amongst other things) PyPI packages. Maybe it’s not that developers working on these packages have gotten more productive. It’s just that they work more, because there is more money to pay for that work. The cohort sizes in figure 3 illustrate this: the 2021 cohort has a non-AI to AI ratio of over 6:1 (1211 to 185). While the 2024 cohort ratio is under 2:1 (727 to 423)! On this view, it’s not so much that AI is making developers superhuman, but that supercharged interest in AI is paying for a higher rate of creation and iteration of packages about AI.
Alas, the data do not tell us which of these effects is larger.
But what we can say is that the main measurable impact of the generative AI revolution, so far, at least on the PyPI ecosystem, is not a Cambrian explosion in all software. But a sharp and concentrated burst in the updating of packages that are themselves part of the AI ecosystem.
See the official pypi blog: Inbound Malware Volume Report↩︎
This data was downloaded from hugovk’s monthly dump of 15,000 top-pypi-packages January 19th 2026.↩︎
We count releases in 12-month windows from each package’s first upload, not calendar years. This avoids having to annualize partial first-year figures. Non-final versions (alpha, beta, rc, dev, post) are excluded.↩︎
We used GPT5.2 to classify packages as “AI-related” or not based on their PyPI description. We agreed on 93% after labeling 100 packages ourselves. The classifications are imperfect but directionally useful.↩︎
All analysis code and data is available at https://github.com/AnswerDotAI/pypi-analysis.↩︎
They can only be interested in one thing, self-advancement. No other explanation works! If they were interested in self-improvement, they might try reading or writing something themselves! Wouldn't it show if they had?
I recognize that models are getting better, but consider: if you already don't understand how programming or LLMs work, and you use LLMs precisely to avoid knowing how to do things, or how they work (the "CEO" mode), each incremental improvement will impress you more than it impresses others. There's no AI exception to Dunning-Kruger.
I recognize that "this" is a difficult thing to pin down in real time. But in the end we know it when we see it, and it has the fascinating and useful quality of not really being explainable by anything else.
Unless and until the culture gets to a place where no one would risk embarrassing themselves by doing something like this, we're stuck with it.
You can get a throw away app in 5 mins, before I wouldn't even bother.
Before that, it single-shot an app for me where I can copy-paste a table (or a subsection of it) from Excel and print it out perfectly aligned on label sticker paper; it does instantly what used to take me an hour each time, when I had to fight Microsoft Word (mail merge) and my Canon printer's settings to get the text properly aligned on labels, and not cut off because something along the way decided to scale content or add margins or such.
Neither of these tools is immediately usable for others. They're not meant to, and that's fine.
I vibe'd a basic ticketing system in just under an hour that does what we need. So not 20 mins, but more like 45-60.
For 20 minutes of time, I had a simple TTS/STT app that allows me to have a voice conversation with my AI assistant.
The amount of useless slop in the app store doesn't matter. There are no new and useful apps made with AI - apps that contribute to productivity of the economy as whole. The trade and fiscal deficits are both high and growing as is corporate indebtedness - these are the true measures for economic failure and they all agree on it.
AI is a debt and energy guzzling endeavor which sucks the capital juice out of the economy in return for meager benefits.
I can't think of a reason for the present unjustified AI rush and hype other than war, but any success towards that goal is a total loss for the economy and environment - that's the relation between economics and deadly destruction in a connected world, reality is the proof.
Really the one thing that conclusively has changed is that the 'ask it on stackoverflow' has become 'ask it an LLM'. Around 95% of the stackoverflow questions can be answered by an LLM with access to the documentation, not sure what will happen to the other 5%. I don't think stackoverflow will survive a 20-fold reduction in size, if only because their stance on not allowing repeat questions means that exponential growth was the main thing preventing them from becoming stale.
In that they obviously have no real utility, sure. There hasn't been a paradigm shift, they still suck at programming, and anyone trying to tell you otherwise almost certainly has something to sell you.
The premise is that AI has already fundamentally changed the nature of software engineering. Not some specific, personal use case, but that everything has changed and that if you're not embracing these tools, you'll perish. In light of this, I don't think your rebuttal works. We should be seeing evidence of meaningful AI contributions all over the place.
Detractors of AI are often accused of moving the goalposts, but I think your comment is guilty of the same. Before Claude Code, we had Cursor, Github Copilot, and more. Each of these was purportedly revolutionizing software engineering.
Further, the core claim for AI coding is that it lets you ship code 10x or 100x faster. So why do we need to wait years to see the result? Shouldn't there be an explosion in every type of software imaginable?
There's a very real problem of low effort AI slop, but throwing out the baby with the bathwater is not the solution.
That said, I do kind of wonder if the old model of open source just isn't very good in the AI era. Maybe when AI gets a lot better, but for now it does take real human effort to review and test. If contributors were reviewing and testing like they should be doing, it wouldn't be an issue, but far too many people just run AI and don't even look at it before sending the PR. It's not the maintainers job to do all the review and test of a low-effort push. That's not fair to them, and even discarding that it's a terrible model for software that you share with anyone else.
I have a list of ideas a mile long that gets longer every day, and LLMs help me burn through that list significantly faster.
However, the older I get, the more distraught I get that most people I meet "IRL" are simply not sitting on a list of problems they simply lack time to solve. I have... a lot of emotions around this, but it seems to be the norm.
If someone doesn't see or experience problems and intuitively start working out how they would fix them if they only had time, the notion that they could pair program effectively ideas that they didn't previously have with an LLM is absurd.
There are many small, different, and one-time tasks that don’t fit full blown apps. Which I would characterize an AI building a novel app as building a house out of random bits of lumber. It will work but will have no cohesive process and sounds like a nightmare.
Would it be harder? Sure. And perhaps the difficulty adds an additional cost of passion being a necessary condition to embark on the innovation. Passion leads to really good stuff.
My personal fear is we get landfill sites of junk software produced. To some extent it should be costly to convert an idea to a concept - the cost being thinking carefully so what you put out there is somewhat legible.
I'd rather not pay monthly for something (like water) that I'm turning on and off and may not even need for weeks. But paying per-liter is currently more expensive so that's what we currently do.
I think the future is going to be local models running on powerful GPUs that you have on-prem or in your homelab, so you don't need your wallet perpetually tethered to a company just to turn the hose on for a few minutes.
Right.
I don't think you even need cynicism or whatever you felt you were having impolite thoughts about:
I'd expect the top mature libraries to be the most resistant to AI tool use for various reasons. They already have established processes, they don't accept drive-by PR spam, the developers working on them might be the least likely to be early adopters, and -- perhaps most importantly -- the todo list of those projects might need the most human comms, like directional planning rather than the sort of yolo feature impl you can do in a one-man greenfield.
All to further bury signals you might find elsewhere in broader ecosystems.
I get that people are upset that making a cool six figures off of stitching together React components is maybe not a viable long-term career path anymore. For those of us on the user side, the value is tremendous. I’m starting to replace what were paid enterprise software and plug-ins and tailoring them to my own taste. Our subject matter experts are translating their knowledge and work flows, which usually aren’t that complicated, into working products on their own. They didn’t have to spend six months or a year negotiating an agreement to build the software or have to beg our existing software vendors, who could not possibly care less, for the functionality to be added to software we are, for some reason, expected to pay for every single year, despite the absence of any operating cost to justify this practice.
Maybe I should have said "obvious to me," but I guess I just struggle to see how a serious crack at using modern opus in claude code doesn't make it obvious at this point.
I'd really recommend trying the "spike out a self-contained minimal version of this rearchitecture/migration and troubleshoot it iteratively until it works, then make a report on findings" use-case for anyone that hasn't had luck with them thus far and is serious about trying to reach conclusions based on direct experience.
Yep, also a huge factor. Why publish something you built with an AI assistant if you know it's going to be immediately dunked on not because the quality may be questionable, but because someone sees an em-dash, or an AI coauthor, and immediately goes on a warpath? Heck I commented[0] on the attitude just a few hours ago. I find it really irritating.
[0] https://github.com/duriantaco/fyn/issues/4#issuecomment-4117...
What's sauce for the goose is sauce for the gander. If you make that argument that 'I don't believe in kinks or discontinuities in code release due to AI, because so many AI coding systems have come out incrementally since 2020', then OP does provide strong evidence for an AI acceleration - the smooth exponential!
This is flat-earther level. It's like an environmentalist saying that nothing made with fossil fuels contributes to productivity. But they don't say that because they know it's not true.
There are so many valid gripes to have with LLMs, pick literally any of them. The idea that a single line of generated code can't possibly be productivity net positive is nonsensical. And if one line can, then so can many lines.
Having your code snatched and its copyright disregarded, to the benefit of some rando LLM vendor. People can just press "pause" and wait until they see whether they fuel something that brings joy to the world. (Which it might in the end. Or not.)
curl 'https://cdn.statcdn.com/Statistic/1020000/1020964-blank-754.png' \
-H 'Origin: https://www.statista.com' \
-H 'Referer: https://www.statista.com/' \
--output chart.png
Assuming it's a real chart, that will give you the image with the uptick in the last year.Relevant! If the maximalist interpretation of AI capabilities were close to real, and if people tend to point their new super powers at their biggest pain points.. wouldn't it be a big blow for all things advertising / attention economy? "Create a driver or wrapper app that skips all ads on Youtube/Spotify" or "Make a browser plugin that de-emphasizes and unlinks all attention-grabbing references to pay-walled content".
If we're supposed to be in awe of how AI can do anything, and we notice repeatedly that nope, it isn't really empowering users yet, then we may need to reconsider the premise.
Seems like the rest of the whole AI business, the only things going to the top are the AI tools themselves but not the things they are supposed to built.
1. To benefit myself with features/projects
2. To benefit others with my work
1 by itself would mean no bothering with PR, modifications, etc. It's way easier to hoard your changes than to go through the effort getting them merged upstream. 2 by itself isn't enough motivation to spend the effort getting up to speed on the codebase, testing, etc. Together though, it's powerful motivation for me.
I have to remind myself that both things are a net positive with AI training on my stuff. It's certainly not all pros (there's a lot of cons with AI too), but on the whole I think we're headed for a good destination, assuming open models continue to progress. If it ends up with winner-takes-all Anthropic or OpenAI, then that changes my calculus and will probably really piss me off. Luckily I've gotten positive value back from those companies, even considering having to pay for it.
The moat used to be effort, nobody wants to rewrite this from scratch (especially when it's free). What's left is actually understanding why the thing works the way it does. Not sure that's enough to sustain open source long-term? I guess we all have to get used to it?
Indeed, I've got a few applications I've built or contributed too that are (A)?GPL, and for those I do worry about this AI washing technique. For libraries that are MIT or permissive anyway, I don't really care. (I default to *GPL for applications, MIT/Apache/etc for libraries)
The AI wouldn't have been able to do it by itself, but I wouldn't have been arsed to do it alone either.
If that would be true then all these AIs are useless. Who needs them to built something that already exists?
Ah my favorite, entirely made up quote.
Apocraphyly attributed to the U.S. Patent Office Commissioner in 1899.