Summarizer

Review Gate Workflows

Alternative approaches using multiple models for code review, fresh context agents, Codex and Gemini reviewers, and loops between agents to improve quality

← Back to Welcome to Gas Town

While the provided article champions high-throughput "vibe coding," commenters argue that the real bottleneck in agentic workflows is maintaining quality and managing the human developer's limited focus. They advocate for a "Review Gate" architecture that prioritizes precision over speed, utilizing independent, fresh-context models like Gemini and Codex to scrutinize code produced by a primary agent. By forcing these agents into iterative loops where a task is only considered complete once a diverse panel of reviewers is satisfied, developers can catch subtle bugs and prevent the compounding errors typical of stochastic systems. Ultimately, these perspectives suggest that for non-trivial or PhD-level work, the most effective strategy is not more parallel agents, but a rigorous, multi-model verification process that minimizes human intervention.

2 comments tagged with this topic

View on HN · Topics
The article seems to be about fun, which I'm all for, and I highly appreciate the usage of MAKER as an evaluation task (finally, people are actually evaluating their theories on something quantitative) but the messaging here seems inherently contradictory: > Gas Town helps with all that yak shaving, and lets you focus on what your Claude Codes are working on. Then: > Working effectively in Gas Town involves committing to vibe coding. Work becomes fluid, an uncountable that you sling around freely, like slopping shiny fish into wooden barrels at the docks. Most work gets done; some work gets lost. Fish fall out of the barrel. Some escape back to sea, or get stepped on. More fish will come. The focus is throughput: creation and correction at the speed of thought. I see -- so where exactly is my focus supposed to sit? As someone who sits comfortably in the "Stage 8" category that this article defines, my concern has never been throughput, it has always been about retaining a high-degree of quality while organizing work so that, when context switching occurs, it transitions me to near-orthogonal tasks which are easy to remember so I can give high-quality feedback before switching again. For instance, I know Project A -- these are the concerns of Project A. I know Project B -- these are the concerns of Project B. I have the insight to design these projects so they compose, so I don't have to keep track of a hundred parallel issues in a mono Project C. On each of those projects, run a single agent -- with review gates for 2-3 independent agents (fresh context, different models! Codex and Gemini). Use a loop, let the agents go back and forth. This works and actually gets shit done. I'm not convinced that 20 Claudes or massively parallel worktrees or whatever improves on quality, because, indeed, I always have to intervene at some point. The blocker for me is not throughput, it's me -- a human being -- my focus, and the random points of intervention which ... by definition ... occur stochastically (because agents). Finally: > Opus 4.5 can handle any reasonably sized task, so your job is to make tasks for it. That’s it. This is laughably not true, for anyone who has used Opus 4.5 for non-trivial tasks. Claude Code constantly gives up early, corrupts itself with self-bias, the list goes on and on. It's getting better, but it's not that good.
View on HN · Topics
I've tried most of the agentic "let it rip" tools. Quickly I realized that GPT 5~ was significantly better at reasoning and more exhaustive than Claude Code (Opus, RL finetuned for Claude Code). "What if Opus wrote the code, and GPT 5~ reviewed it?" I started evaluating this question, and started to get higher quality results and better control of complexity. I could also trust this process to a greater degree than my previous process of trying to drive Opus, look at the code myself, try and drive Opus again, etc. Codex was catching bugs I would not catch with the same amount of time, including bugs in hard math, etc -- so I started having a great degree of trust in its reasoning capabilities. I've codified this workflow into a plugin which I've started developing recently: https://github.com/evil-mind-evil-sword/idle It's a Claude Code plugin -- it combines the "don't let Claude stop until condition" (Stop hook) with a few CLI tools to induce (what the article calls) review gates: Claude will work indefinitely until the reviewer is satisfied. In this case, the reviewer is a fresh Opus subagent which can invoke and discuss with Codex and Gemini. One perspective I have which relates to this article is that the thing one wants to optimize for is minimizing the error per unit of work. If you have a dynamic programming style orchestration pattern for agents, you want the thing that solves the small unit of work (a task) to have as low error as possible, or else I suspect the error compounds quickly with these stochastic systems. I'm trying this stuff for fairly advanced work (in a PhD), so I'm dogfooding ideas (like the ones presented in this article) in complex settings. I think there is still a lot of room to learn here.