Summarizer

Vibe Coding Quality Concerns

← Back to Claude Code On-the-Go

The rise of "vibe coding" has sparked a polarizing debate between enthusiasts who celebrate its rapid, phone-based workflow and skeptics who fear it produces a "mountain of sloppy, unmaintainable code." While some developers find success treating LLMs as autonomous agents for toy projects, others argue that professional quality still demands a "human-in-the-loop" to review output with the same excruciating scrutiny one would give a "dyslexic intern." Concerns persist regarding the loss of local source code visibility and the security risks inherent in "web sandbox" setups, where developers may inadvertently ship vulnerabilities alongside their features. Ultimately, the consensus suggests that while AI can handle the heavy lifting, maintaining high standards requires moving beyond the "vibe" to focus on rigorous testing, manual poking, and disciplined oversight.

15 comments tagged with this topic

View on HN · Topics
Would you describe this product as a whole application suite (blogging, calendar, commerce) plus its own backend infrastructure that is capable of serving these apps to the public internet and functioning offline via ad-hoc wireless peer-to-peer, with a cryptographic layer providing identity, security and censorship resistance, and that runs on phone, laptop or raspberry pi? Quite ambitious. Is this an LLM hallucinating? taking a break from coding? or leaking your personal desktop session? https://github.com/geograms/geogram/blob/main/.cli_history
View on HN · Topics
> It won't matter if I'm washing the dishes, walking the dog, driving to the supermarket, picking up my kids from school. I'll always be switched on, on my phone, continuously talking to an LLM, delivering questionable features and building meaningless products, destroying in the process the environment my kids are going to have to grow in. I remember hearing similar criticisms of continuous delivery. On one end of the spectrum people who had to wait months to get changes out now got them out relatively quickly. On the other end of the spectrum, some person was going to push changes at midnight. A decade on forward I've never actually worked at a shop that at scale did continuous delivery in its truest sense where changes go straight to production. Simply, nothing beats a human in the loop; it's always about balancing the costs of automation and a lower barrier to entry. I imagine this kind of thing, if it ever actually takes hold and can be adopted by a larger subset of engineers, will follow a similar path. Long way of saying, I don't think you're Chicken Little but also don't start breathing into a bag just yet.
View on HN · Topics
I've been really impressed with https://github.com/BloopAI/vibe-kanban to do this. Really really impressed.
View on HN · Topics
I'm surprised to see people getting value from "web sandbox"-type setups, where you don't actually have access to the source code. Are folks really _that_ confident in LLMs as to entirely give up the ability to inspect the source code, or to interact with a running local instance of the service? Certainly that would be the ideal, but I'm surprised that confidence is currently running that high.
View on HN · Topics
Oh fascinating - so you're reviewing "your own" code in-PR, rather than reviewing it before PR submission? I can see that working! Feels weird, but I can see it being a reasonable adaptation to these tools - thanks! What about running services locally for manual testing/poking? Do you open ports on the Anthropic VM to serve the endpoints, or is manual testing not part of your workflow?
View on HN · Topics
The output from Jules is a PR. And then it's a toss-up between "spot on, let's merge" and "nah, needs more work, I will check out the branch and fix it properly when I am the keyboard". And you see the current diff on the webpage while the agent is working.
View on HN · Topics
> it was weird to me that people were comfortable operating on a codebase that they don't have locally, that they can't directly interact with. I have a project where I've made a rule that no code is written by humans. It's been fun! It's a good experience to learn how far even pre-Opus 4.5 agents can be pushed. It's pretty clear to me that in 12 months time looking at the code will be the exception, not the rule.
View on HN · Topics
I have my very fast macbook pro at my desk in my office, and I use tmux and tailscale and git worktrees and I’ve built a notification setup like this author. Thanks to tailscale and ssh I can vibecode on the go from my phone with this setup. While it’s great to leave a task running, no matter what I do I can’t achieve the type of high quality work on the go that I can when I’m sitting at my desk. For me working on a full SaaS.. I just can’t do quality work on my phone. The only way I can do quality work is to sit at my desk where I’m focused on the work. To play with the result of a prompt, take copious notes, feed them back to the agent, not ship until the thing is polished to a shine. To feature flag the changes, review all code in excruciating detail as though it was written by a dyslexic intern, add all the metrics and logs one can think of (VictoriaMetrics), add user-behavior logging (Amplitude/Posthog) and monitor the thing like your livelihood depends on it. Because it’s a product and you have pride in your work. All of that needs loads of screen real estate and a keyboard. So I guess the more things change, the more they stay the same.
View on HN · Topics
Same here, I’m vibecoding a toy project where I never looked at the code from my phone, but I always seat for work. I’m using happy app and that’s good enough for now, I have the desktop in tailscale but I access it that way just for testing
View on HN · Topics
I wonder when/how to test and review the code though? I mean, how do you know Claude Code hasn't entered a completely different path than you had imagined?
View on HN · Topics
That sounds nice, but what happens when there's something Claude messes up, doesn't know how to do something, or when you have to review the thousand lines it added to your project ? Unless it's a totally vibe coded side project without any tests or quality control of some sort. I'm just curious what you can build with this setup. It just seems to be the way to create a mountain of sloppy, unmaintainable code.
View on HN · Topics
What kind of things people are building that can be almost completely automatically built like this?
View on HN · Topics
Yes that works too. But this way I can open the firewall, npm run dev and send the link of my new vibe coded security vulnerability/app to my friends without my computer running. Plus a VM for this, a container for that and soon my 32GB memory isn't enough. I offload aggressively.
View on HN · Topics
Setup is still rough around the edges (use an agent to set it up), but clawdbot (prev clawdis) from Peter Steinberger works phenomenally well for agent orchestration and personal assistance. The community for clawd is exploding right now, and I think this is purely based on merit. It’s been a game changer for my vibe coding workflow, and lots of fun. https://github.com/clawdbot/clawdbot
View on HN · Topics
> Now I code from my phone. Except that you are doing anything else but coding here. Coding involves writing code, which isn't actually done by the author here.