I think, for a lot of people, solving the problem was always the fun part. There is immense pleasure in a nice piece of code - something that is elegant, clever and simple at the same time. Grinding out code to get something finished - less fun…
Sure, but, in the real world, for the software to deliver a solution, it doesn't really matter if something is modelled in beautiful objects and concise packages, or if it's written in one big method. So for those that are more on the making /things/ side of the spectrum, I guess they wouldn't care if the LLM outputs code that has each iteration written separately. It's just that if you really like to work on your craftsmanship, you spend most of the time rewriting/remodelling because that's where the fun is if you're more on the /making/ things side of the spectrum, and LLMs don't really assist in that part (yet?). Maybe LLMs could be used to discuss ways to model a problem space?
As I've gotten more experience I've tended to find more fun in tinkering with architectures than tinkering with code. I'm currently working on making a secure zero-trust bare metal kubernetes deployment that relies on an immutable UKI and TPM remote attestation. I'm making heavy use of LLMs for the different implementation details as I experiment with the architecture. As far as I know, to the extent I'm doing anything novel, it's because it's not a reasonable approach for engineering reasons even if it technically works, but I'm learning a lot about how TPMs work and the boot process and the kernel. I still enjoy writing code as well, but I see them as separate hobbies. LLMs can take my hand-optimized assembly drag racing or the joy of writing a well-crafted library from my cold dead hands, but that's not always what I'm trying to do and I'll gladly have an LLM write my OCI layout directory to CPIO helper or my Bazel rule for putting together a configuration file and building the kernel so that I can spend my time thinking about how the big pieces fit together and how I want to handle trust roots and cold starts.
On top of that there's a not insignificant chance you've actually just stolen the code through an automated copyright whitewashing system. That these people believe they're adding value while never once checking if the above is true really disappoints me with the current direction of technology. LLMs don't make everyone better, they make everything a copy. The upwards transfer of wealth will continue.
I enjoy coding for the ability to turn ideas into software. Seeing more rapid feature development, and also more rapid code cleanup and project architecture cleanup is what makes AI assisted coding enjoyable to me
Or, given that OP is presumably a developer who just doesn't focus fully on front end code they could skip straight to checking MDN for "center div" and get a How To article ( https://developer.mozilla.org/en-US/docs/Web/CSS/How_to/Layo... ) as the first result without relying on spicy autocomplete. Given how often people acknowledge that ai slop needs to be verified, it seems like a shitty way to achieve something like this vs just checking it yourself with well known good reference material.
Aren't you afraid it's gonna be a race to the bottom ? the software industry is now whoever pays gemini to deploy something prompted in a few days. Everybody can, so the market will be inundated by a lot of people, and usually this makes for a bad market (a few shiny one gets 90% of the share while the rest fight for breadcrumbs) I'm personally more afraid that stupid sales oriented will take my job instead of losing it to solid teams of dedicated expert that invested a lot of skills in making something on their own. it seems like value inversion
Yes, I worry about this quite a bit. Obviously nobody knows yet how it will shake out, but what I've been noticing so far is that brand recognition is becoming more important. This is obviously not a good thing for startup yokels like me, but it does provide an opportunity for quality and brand building. The initial creation and generation is indeed much easier now, but testing, identifying, and fixing bugs is still very much a process that takes some investment and effort, even when AI assisted. There is also considerable room for differentiation among user flows and the way people interact with the app. AI is not good at this yet, so the prompter needs to be able to identify and direct these efforts. I've also noticed in some of my projects, even ones shipped into production in a professional environment, there are lots of hard to fix and mostly annoying bugs that just aren't worth it, or that take so much research and debugging effort that we eventually gave up and accepted the downsides. If you give the AI enough guidance to know what to hunt for, it is getting pretty good at finding these things. Often the suggested fix is a terrible idea, but The AI will usually tell you enough about what is wrong that you can use your existing software engineering skills and experience to figure out a good path forward. At that point you can either fix it yourself, or prompt the AI to do it. My success rate doing this is still only at about 50%, but that's half the bugs that we used to live with that we no longer do, which in my opinion has been a huge positive development.
‘Why were they long term?’ is what you need to ask. Code has become essentially free in relative terms, both in time and money domains. What stands out now is validation - LLMs aren’t oracles for better or worse, complex code still needs to be tested and this takes time and money, too. In projects where validation was a significant percentage of effort (which is every project developed by more than two teams) the speed up from LLM usage will be much less pronounced… until they figure out validation, too; and they just might with formal methods.
As a customer, I don't want to pay for vibe-coded products, because authors also don't have a time (and/or skills) to properly review, debug and fix products.
Definitely. I'm not disparaging the process of assembling IKEA furniture, nor the process of producing software using LLMs. I've done both, and they have their time and place. What I'm pushing back on is the idea that these are equivalent to carpentry and programming. I think we need new terminology to describe this new process. "Vibe coding" is at the extreme end of it, and "LLM-assisted software development" is a mouthful. Although, the IKEA analogy could be more accurate: the assembly instructions can be wrong; some screws may be missing; you ordered an office chair and got a dining chair; a desk may have five legs; etc. Also, the thing you built is made out of hollow MDF, and will collapse under moderate levels of stress. And if you don't have prior experience building furniture, you end up with no usable skills to modify the end result beyond the manufacturer's original specifications. So, sure, the seemingly quick and easy process might be convenient when it works. Though I've found that it often requires more time and effort to produce what I want, and I end up with a lackluster product, and no learned skills to show for it. Thus learning the difficult process is a more rewarding long-term investment if you plan to continue building software or furniture in the future. :)
- Providing boilerplate/template code for common use cases - Explaining what code is doing and how it works - Refactoring/updating code when given specific requirements - Providing alternative ways of doing things that you might not have thought of yourself YMMV; every project is different so you might not have occasion to use all of these at the same time.
I think it depends what you are doing. I’ve had Claude right the front end of a rust/react app and it was 10x if not x (because I just wouldn’t have attempted it). I’ve also had it write the documentation for a low level crate - work that needs to be done for the crate to be used effectively - but which I would have half-arsed because who like writing documentation? Recently I’ve been using it to write some async rust and it just shits the bed. It regularly codes the select! drop issue or otherwise completely fails to handle waiting on multiple things. My prompts have gotten quite sweary lately. It is probably 1x or worse. However, I am going to try formulating a pattern with examples to stuff in its context and we’ll see. I view the situation as a problem to be overcome, not an insurmountable failure. There may be places where an AI just can’t get it right: I wouldn’t trust it to write the clever bit tricks I’m doing elsewhere. But even there, it writes (most of) the tests and the docs. On the whole, I’m having far more fun with AI, and I am at least 2x as productive, on average. Consider that you might be stuck in a local (very bad) maximum. They certainly exist, as I’ve discovered. Try some side projects, something that has lots of existing examples in the training set. If you wanted to start a Formula 1 team, you’re going to need to know how to design a car, but there’s also a shit ton of logistics - like getting the car to the track - that an AI could just handle for you. Find boring but vital work the AI can do because, in my experience, that’s 90% of the work.
Mmm, I do a lot of frontend work but I find writing the frontend code myself is faster. That seems to be mostly what everyone says it's good for. I find it useful for other stuff like writing mini scripts, figuring out arguments for command line tools, reviewing code, generating dumb boilerplate code, etc. Just not for actually writing code.
My problem is that code review has always been the least enjoyable part of the job. It’s pure drudgery, and is mentally taxing. Unless you’re vibe coding, you’re now doing a lot of code review. It’s almost all you’re doing outside of the high-level planning and guidance (which is enjoyable). I’ve settled on reviewing the security boundaries and areas that could affect data leaks / invalid access. And pretty much scanning everything else. From time to time, I find it doing dumb things- n+1 queries, mutation, global mutable variables, etc, but for the most part, it does well enough that I don’t need to be too thorough. However, I wouldn’t want to inherit these codebases without an AI agent to do the work. There are too many broken windows for human maintenance to be considered.
Worse, you’re doing code review of poorly written code with random failure modes no human would create, and an increasingly big ball of mud that is unmaintainable over time. It’s just the worst kind of reviewing imaginable. The AI makes an indecipherable mess, and you have to work out what the hell is going on.
There's been so much pressure to use AI at work. My codebase is a zen garden I've been raking for 6 years. I have concerns about what's going to happen after a few months of "we're using AI cause they told us to."
It's amazing to be able to try a bunch of ideas with very minimal cost. That being said, AI code assistants don't have eyeballs and they often make things that don't look very good. Craft, polish and judgement still matter.
I find they can make some things look objectively "good", but they just look generic and it feels very easy to spot a site that was made without the vision, polish and judgement. You can get LLMs to create some truly unique sites, but it takes a lot more work than a few prompts.
In the context of "fun again", debugging slop, finding imaginary dependencies, and discovering unimaginably fragile code isn't fun , even if it's not important. But past bad output, I worry for our creative fulfillment. The old timers are right. That feeling of accomplishment, a keystone of happiness is a product of work. Probably beyond the scope of the thread.
Setting up build system and prototyping sure. As a replacement for Figma it’s great. But I would throw away all the code and start from scratch if I wanted to be able to maintain the code in the long term.
Either the projects he's working on are side projects, and in that case I don't see why he would need to use the complex pipelines, just Vanilla JS and PHP still work super fine, even better nowadays actually, or the projects are professional ones and then to ship code written by AI is extremely dangerous and he should have resources (time and people) to do things properly without AI. So, I'm clearly not convinced.
Going in 2026, the frontend has many good options, but AI is not one of them. We have many typesafe (no, not TypeScript!) options with rock solid dev tooling, and fast compilers. AI is just a badaid, its not the road you want to travel.
Of course its fun. Making slop _is_ very fun. Its a low-effort dopamine-driven way of producing things. Learning is uncomfortable. Improving things using only your braincells can be very difficult and time consuming.
Agree with this. Like the author, I've been keeping ajour with web development for multiple decades now. If you have deep software knowledge pre-LLM, you are equipped with the intuition and knowledge to judge the output. You can tell the difference between good and bad, if it looks and works the way you want, and you can ask the relevant questions to push the solution to the actual thing that you envisioned in your mind. Without prior software dev experience people may take what the LLM gives them at face value, and that's where the slop comes from imho.
It’s because business demand speed and shipping over other concerns. We had to fight hard for proper quality controls in the face of the LLM coding assistance boom where I work. These are great tools but they have limits and can lead to poor engineering hygiene quite quickly. It took a major issue being attributed to having too much trust in these tools before we were able to enforce better hygiene with them
Customers don't buy software based on quality first, they buy on features. Until customers in mass, or regulations demand quality, money will be made on deliveries. If your lucky and can program how you want and take the time you need, then you can focus on the attributes you feel best about.
If you have customers that will put up with things being slow as molasses and crashing al the time, well….can you send some my way because mine won’t STFU about it.
please, developers are making terrible slop way before AI, look at the javascript infested frameworks in use on the web. they make NO sense. they are not making things simpler
And now you can generate javascript infested slop frameworks for $5 per million tokens. Such an improvement. And it's so easy to just ask Claude to make one for you, why even bother standardizing anything when you can just use bespoke slop for anything anymore. Libraries and frameworks? Not needed. Just shove everything into CC/Codex and let it figure it out.