Hi all,
I want to raise something that may be a little sensitive, but I am doing so in good faith and with genuine curiosity rather than criticism.
Firstly, I wanted to point out that I work heavily with AI driven solutions myself across a range of projects, including some fairly complex systems. I am very familiar with what current large language models can and cannot do in real world production environments.
They are impressive tools, and it is a rapidly growing technology space. I am a big fan of where it can actually aid through acceleration and helpful autocomplete assistance, refactoring support, and certain debugging workflows. Used correctly, they are valuable in our space.
However, they are not autonomous engineers. They still require clear direction, and the person using them having the majority of the understanding to correctly prompt and guide the A.I to produce not just working but the correct solutions. Their output should always have careful review.
In my own experience, when AI generated code is used without strong oversight or understanding, certain patterns tend to emerge:
-
Features or methods unintentionally removed while fixing something else
-
Logic rewritten in a way that subtly breaks existing behaviour
-
Previously fixed bugs resurfacing in slightly altered form
-
Code that “looks” structurally correct but does not fully respect the architectural context
-
Inconsistent patterns introduced into otherwise mature codebases
-
And just those odd issues that can happen with human coding but are kind of different. I am sure people will know what I mean. Odd issues.
We are seeing this nearly on a weekly basis at the moment.
While some companies do not want to admit it some major patch and update deployments have clearly been A.I developed, debugged and deployed and have caused a variety of large issues.
Windows is in a shocking mess because it is no secret that bugs and patch deployment is driven by their huge A.I Push and its NOT GOOD.
We have impressive, improving image and video generation but still limited and flawed in practical use. Coca Cola spent way to much and had too many people using A.I prompting to generate a really bad Xmas advert. The winter Olympic allowed A.I video generation that fundamentally broke the Olympic rings representation and allowed that to role out the door! It is the “Well, we spent this much money and time on it, we got to use it”.
Over the past several Umbraco releases, I have noticed what appear to be some unusual regressions and reintroductions of issues. Occasionally functionality seems to disappear and later reappear after being reported and just some of these patterns I have started to notice with A.I managed projects.
I have also noticed references to Claude and Claude bot related tooling in the Git repositories.
To be clear, I am not opposed to AI assistance in development. Quite the opposite. An AI assistant helping engineers with legwork, scaffolding, or repetitive tasks is entirely sensible. A faster more direct assistant aiding debugging is really useful!
What concerns me is when there are signs of over reliance without rigorous human review and architectural accountability.
An increased workload in the last year or so for myself has been involving inheriting projects that were clearly heavily AI generated, and I see these consistent patterns in how issues arise as well as seeing and knowing when my Junior staff have blindly used A.I with prompts vs, aiding them or doing it themselves.
Some of the bugs and regressions I have seen myself, seen reported recently in Umbraco feel similar to those patterns. That may be coincidence, but it has prompted me to have concerns and questions.
So I am simply asking, openly and respectfully:
-
Is AI tooling being used in core development workflows?
-
If so, what safeguards or review processes are in place?
-
Are others observing similar regression patterns, or am I over analysing this?
This is not an accusation. It is a genuine question about process and quality control as the industry evolves.
If this is a direction Umbraco is taking I think it is important to have those flags go up. I care about Umbraco as a platform, and I want to understand how these new tools are being integrated into development practices.
Thanks