llm/9db4e77f-8dd5-46da-972e-40d33f3399ef/batch-6-3f2adc1d-6525-41c3-9519-e17cb33886a3-input.json
The following is content for you to classify. Do not respond to the comments—classify them.
<topics>
1. Feasibility of Parallel Agent Workflows
Related: Skepticism regarding the human capacity to supervise multiple AI agents simultaneously, utilizing analogies like washing dishes vs. laundry, and debating the cognitive load required for context switching between 10 active coding streams.
2. Code Quantity versus Quality
Related: Discussions on whether generating 50-100 Pull Requests a week represents true productivity or merely 'token-maxxing', with concerns about code churn, technical debt, and the inability of humans to properly review such high volumes of generated code.
3. The One-Person Unicorn Startup
Related: Debates on whether AI enables solo founders to build billion-dollar companies, arguing that while coding is easier, business bottlenecks like sales, marketing, and product-market fit remain unsolved by LLMs, despite rumors of stealth successes.
4. Claude Code Product Feedback
Related: User feedback on the Claude Code CLI tool, mentioning specific bugs like terminal flickering and context loss, comparisons to tools like Codex and Cursor, and complaints about reliability and lack of basic features.
5. Cost and Access Disparities
Related: Analysis of the financial feasibility of running Opus 4.5 agents in parallel, noting that while Anthropic employees may have unlimited access, the cost for average users would be prohibitive due to token limits and API pricing.
6. Marketing Hype and Astroturfing
Related: Accusations that the original post and similar recent content represent a coordinated marketing campaign by Anthropic, with users expressing distrust of 'influencer' style posts and potential conflicts of interest from the tool's creator.
7. Future of Software Engineering
Related: Existential concerns about the devaluation of coding skills, the shift from creative building to managerial reviewing of AI output, and fears that junior developers will lose the opportunity to learn through doing.
8. Technical Workflow Configurations
Related: Specific details on managing AI agents, including the use of git worktrees for isolation, planning modes, 'teleporting' sessions between local CLI and web interfaces, and using markdown files to define agent behaviors.
9. AI Code Review Strategies
Related: Approaches for handling AI-generated code, such as using separate AI instances to review PRs, the necessity of rigorous CI/CD guardrails, and the danger of blindly trusting 'green' tests without human oversight.
10. The Light Mode Terminal Debate
Related: A humorous yet contentious side discussion sparked by the creator's use of a light-themed terminal, leading to arguments about eye strain, readability, astigmatism, and developer cultural norms regarding dark mode.
11. SaaS Commoditization and Moats
Related: Predictions that AI will drive the marginal cost of software to zero, eroding traditional SaaS business models, and that future business value will rely on proprietary data, domain expertise, and distribution rather than code.
12. Agentic Limitations and Reliability
Related: Criticisms of current AI agents acting like 'slot machines' requiring constant steering, their struggle with complex concurrency bugs, and the observation that they often produce boilerplate rather than solving deep architectural problems.
13. Corporate Adoption and Budgeting
Related: Anecdotes about colleagues burning through massive amounts of API credits with varying degrees of success, and the disconnect between management's desire for AI productivity and the reality of review bottlenecks.
14. Context Management Techniques
Related: Discussions on how to optimize context for AI agents, including the use of CLAUDE.md or AGENTS.md to establish rules, and the technical challenges of context limits and pruning during long sessions.
15. Vibe Coding vs. Engineering
Related: The distinction between 'vibe coding' (iterating until it feels right without deep understanding) and traditional engineering, with experienced developers using AI as a force multiplier rather than a replacement for understanding.
0. Does not fit well in any category
</topics>
<comments_to_classify>
[
{
"id": "46522950",
"text": "In the reddit thread Boris says they’re adding the ability to call skills via slash commands in an upcoming release and that he uses the term skill and slash commands interchangeably."
}
,
{
"id": "46522928",
"text": "\"Boris: Skills = slash commands, I use them interchangeably\"\n\nhttps://www.reddit.com/r/ClaudeAI/comments/1q2c0ne/comment/n..."
}
,
{
"id": "46522800",
"text": "I believe slash commands are all loaded into the initial context and executed when invoked by the user. Skills on the other hand only load the name and description into initial context, and the agent (not user) determines when to invoke them, and only then is the whole skill loaded into context. So skills shift decision making to the agent and use progressive disclosure for context efficiency."
}
,
{
"id": "46523674",
"text": "How much Codex and Claude Code are different from each other?\nI have been using Codex for few weeks doing experiments related to data analysis and training models with some architecture modifications. I wouldn't say I have used it extensively, but so far my experience has been good. Only annoying part has been not able to use GPU in the Codex without using `--sandbox danger-full-access` flag. Today, I started using Claude Code, and ran similar experiments as Codex. I find the interface is quite similar to Codex. However, I hit the limit quite quickly in Claude Code. I will be exploring its features further. I would appreciate if anyone can share their experience of using both tools."
}
,
{
"id": "46525202",
"text": "Codex is heavily inspired by Claude Code. They aren't that different. They might diverge more in future."
}
,
{
"id": "46525317",
"text": "They are quite different, Claude Code with Opus 4.5 is miles better than Codex"
}
,
{
"id": "46527118",
"text": "What is it you find makes it much better?"
}
,
{
"id": "46523414",
"text": "Why stop at 5-10?\nMake it 5 billion - 10 billion parallel agents.\nPR number go up"
}
,
{
"id": "46523133",
"text": "Frankly Claude code is painfully slow. To the point I get frustrated.\n\nOn large codebases I often find it taking 20+ minutes to do basic things like writing tests.\n\nWay too often people are like it takes 2 minutes for it to do a full pr. Yeah how big is the code base actually.\n\nI also have a coworker who is about 10x more then everyone else. Burning through credits yet he is one of the lowest performers.{closing in on around 1k worth of credits a day now)."
}
,
{
"id": "46523143",
"text": "$1,000.00 of credits per-day?? $200,000 per year? Those are bonkers numbers for someone not performing at a high level (on-top of their salary). Do you know what they are doing?"
}
,
{
"id": "46523166",
"text": "Yup. The way he works is all tasks he is issued in a sprint he just fires them through opus in parallel hoping to get a hit on Claude magically solving the ticket having them constantly be iterated on them. He doesnt even try using proper having plans be created.\n\nOften time tickets get fleshed out or requirements change. He just throws everything out and reshoves it into Claude.\n\nI weep for the planet."
}
,
{
"id": "46525325",
"text": "They should just be on the $200 a month Max plan"
}
,
{
"id": "46522721",
"text": "I don't understand how these setups scale longterm, and even more so for the average user. The latter is relevant because, as he points out, his setup isn't that far out of reach of the average person - it's still fairly close to out of the box claude code, and opus.\n\nBut between the model qualities varying, the pricing, the timing, the tools constantly changing, I think it's really difficult to build the institutional knowledge and setup that can be used beyond a few weeks.\n\nIn the era of AI, I don't tink it's good enough to \"have\" a working product. It's also important to have all the other things that make a project way more productive, like stellar documentation, better abstractions, clearer architecture. In terms of AI, there's gotta be something better than just a markdown file with random notes. Like what happens when an agent does something because it's picking something up from some random slack convo, or some minor note in a 10k claude.md file. It just seems like the wild west where basic ideas like additional surface area being a liability is ignored because we're too early in the cycle.\n\ntl;dr If it's just pushing around typical mid-level code, then... I just think that's falling behind."
}
,
{
"id": "46522828",
"text": "I'm a bit jealous. I would like to experiment with having a similar setup, but 10x Opus 4.5 running practically non stop must amount to a very high inference bill. Is it really worth the output?\n\nFrom experimentation, I need to coach the models quite closely in order to get enough value. Letting it loose only works when I've given very specific instructions. But I'm using Codex and Clai, perhaps Claude code is better."
}
,
{
"id": "46523145",
"text": "I have a coworker who is basically doing this right now he leads our team and is second place overall. Regularly runs opus in parallel he alone is burning through 1k worth of credits a day.\n\nHe is also one of our worst performers."
}
,
{
"id": "46523571",
"text": "Wait, what is he second place at?"
}
,
{
"id": "46524268",
"text": "Credit usage."
}
,
{
"id": "46523150",
"text": "I've tried running a number of claude's in paralell on a CRUD full stack JS app. Yes, it got features made faster, yes it definitely did not leave me enough time to acutally look at what they did, yes it definitely produced sub-par code.\n\nAt the moment with one claude + manually fixing crap it produces I am faster at solving \"easier\" features (Think add API endpoint, re-build API client, implement frontend logic for API endpoint + UI) faster than if I write it myself.\n\nThings that are more logic dense, it tends to produce so many errors that it's faster to solve myself."
}
,
{
"id": "46523637",
"text": "> manually fixing crap it produces\n\n> it tends to produce so many errors\n\nI get some of the skepticism in this thread, but I don't get takes like this. How are you using cc that the output you look at is \"full of errors\"? By the time I look at the output of a session the agent has already ran linting, formatting, testing and so on. The things I look at are adherence to the conventions, files touched, libraries used, and so on. And the \"error rate\" on those has been steadily coming down. Especially if you also use a review loop (w/ codex since it has been the best at review lately).\n\nYou have to set these things up for success. You need loops with clear feedback. You need a project that has lots of clear things to adhere to. You need tight integrations. But once you have these things, if you're looking at \"errors\", you're doing something wrong IMO."
}
,
{
"id": "46524444",
"text": "I don't think he meant like syntax errors, but thinking errors. I get these a lot with CC. Especially for example with CSS. So much useless code it produces, it blows my mind. Once I deleted 50 lines of code and manually added 4 which was enough to fix the error."
}
,
{
"id": "46522860",
"text": "Yeah, doesn't this guy work for Anthropic? He'd get to use 10x Opus 4.5 for free."
}
,
{
"id": "46524621",
"text": "I’ve found that experienced devs use agentic coding in a more “hands-on” way than beginners and pure vibe-coders.\n\nVibecoders are the best because they push the models in humorous and unexpected ways.\n\nJunior devs are like “I automated the deploy process via an agent and this markdown file”\n\nSeasoned devs will spend more time writing the prompt for a bug fix, or lazily paste the error and then make the 1-line change themselves.\n\nThe current crop of LLMs are more powerful than any of these use cases, and it’s exciting to see experienced devs start to figure that out (I’m not stanning Gas Town[0], but it’s a glimpse of the potential).\n\n[0] https://steve-yegge.medium.com/welcome-to-gas-town-4f25ee16d..."
}
,
{
"id": "46525184",
"text": "Partially related: I really dislike the vibe of Gas Town, both the post and the tool, I really hope this isn't what the future looks like. It just feels disappointing."
}
,
{
"id": "46525979",
"text": "Boris is a power user's power user.\n\nI would highly recommend every project maintain a simple, well-written AGENTS.md file. At first it may seem like a more nitpicky README, but you will quickly see how much coding agents benefit from this added context. Imo, the two most important things to include in AGENTS.md are frequent commands and verification methods.\n\nA third thing I've started adding to my projects is a list of related documentation and libraries that may not be immediately obvious. Things like confluence pages and other repos associated with the project."
}
,
{
"id": "46523569",
"text": "I had this ShowHN yesterday, which didn't grab much attention, so i'm using this opportunity as it seems relevant (it is a solution for running CC in parallel)\n\nif you folks like to run parallel claude-code sessions, and like native terminal like Ghostty, i have a solution for using Git Worktree natively with Ghostty, it is called agentastic.dev, and it has a built-in Worktree/IDE/Diff/Codereview around Ghostty (macos only for now).\n\nWould be happy to answer any questions\n\nShowHN post: https://news.ycombinator.com/item?id=46501758"
}
,
{
"id": "46524372",
"text": "Warm take for sure, but I feel that LLMs and agents have made me a worse programmer as a whole. I am enjoying the less mental strain as I do my hobbies like sketching/art while the LLM is running. But definitely it isn't making me any faster.\n\nI'm at the point of considering another job but I just fear that my skills have deteriorated to the point that I can't pass any (manual) coding assessments anymore"
}
,
{
"id": "46525834",
"text": "How come that Claude Code isn't open source?\n\nI can't imagine that it has some kind of special access to Anthropic's servers and that it does things an API-user can't do, maybe except for the option to use the Claude.ai credits/quota.\n\nEven their Agent SDK just wraps the `claude`-executable, IIRC."
}
,
{
"id": "46523633",
"text": "That's just unrealistic. If i were to use it like this as an actual end user i would get stopped by rate limits/those weekly / session limits instantly"
}
,
{
"id": "46524791",
"text": "10-15 parallel Opus 4.5 instances running almost full-time? Even if you could get it, what would be the monthly bill for that?"
}
,
{
"id": "46528942",
"text": "One of my side projects has been to recover a K&R C computer algebra system from the 1980's, port to modern 64-bit C. I'd have eight tabs at a time assigned files from a task server, to make passes at 60 or so files. This nearly worked; I'm paused till I can have an agent with a context window that can look at all the code at once. Or I'll attempt a fresh translation based on what I learned.\n\nWith a $200 monthly Max subscription, I would regularly stall after completing significant work, but this workflow was feasible. I tried my API key for an hour once; it taught me to laugh at the $200 as quite a deal.\n\nI agree that Opus 4.5 is the only reasonable use of my time. We wouldn't hire some guy off the fryer line to be our CTO; coding needs best effort.\n\nNevertheless, I thought my setup was involved, but if Boris considers his to be vanilla ice cream then I'm drinking skim milk."
}
,
{
"id": "46525067",
"text": ">Most sessions start in Plan mode\n\nI don't get why they added a Plan mode. Even without it, you can just ask claude to \"make a plan\" from where you can iterate on it.\n\nAlso is it really that much faster to type \"/commit-push-pr\" than it is to type \"commit, push and make a pr\" ?"
}
,
{
"id": "46525311",
"text": "It’s really a convenience to force the model to use the planning tool and it prevents edit/write tools until the user approves the plan, like an inverse of “auto accept edits” mode."
}
,
{
"id": "46523659",
"text": "One thing that’s helped me is creating a bake-off. I’ll do it between Claude and codex. Same prompt but separate environments. They’ll both do their thing and then I’ll score them at the end. I find it helps me because frequently only one of them makes a mistake, or one of them finds an interesting solution. Then once I declare a winner I have scripts to reset the bake-off environments."
}
,
{
"id": "46523918",
"text": "you can have them review each other's work, too."
}
,
{
"id": "46523468",
"text": "I spent a whole day running 3x local CC sessions and about 7 Claude code web sessions over the day. This was the most heavy usage day ever for me, about 30 pull requests created and merged over 3 projects.\n\nI got a lot done, but my brain was fried after that. Like wired but totally exhausted.\n\nHas anyone else experienced this and did you find strategies to help (or find that it gets easier)?"
}
,
{
"id": "46523739",
"text": "Did you manage to make proper reviews of all the 30 PR ?"
}
,
{
"id": "46529750",
"text": "I have formal requirements for all implemented code. This is all on relatively greenfield solo developed codebases with tools I know inside out (Django, click based cli etc) so yes. Thanks so much for your concern, internet person!"
}
,
{
"id": "46530572",
"text": "I was genuinely interested in knowing if you did it properly or not, since I read a lot of tales like this but don't understand how it can be true."
}
,
{
"id": "46526551",
"text": "I think we all know the answer to that."
}
,
{
"id": "46530029",
"text": "what cc plan do you use? the20$?"
}
,
{
"id": "46533780",
"text": "Max 5x ($100)"
}
,
{
"id": "46523031",
"text": "Having the 5 instances going at once sounds like Google Antigravity.\n\nI haven't used Claude Code too much. One snag I found is the tendency when running into snags to fix them incorrectly by rolling back to older versions of things. I think it would benefit from an MCP server for say Maven Central. Likewise it should prefer to generate code using things like project scaffolding tooling whenever possible."
}
,
{
"id": "46523305",
"text": "This feels like the desperate, look at me! post, which is the exact opposite of Andrej Karpathy's recent tweet[0] about feeling left behind as a programmer, as covered on Hacker News[1].\n\nI guess would want to see how sustainable this 5 parallel AI effort is, and are there demonstrably positive outcomes. There are plenty of \"I one-shotted this\" examples of something that already (mostly) existed, which are very impressive in their own right, but I haven't seen a lot of truly novel creations.\n\n[0] https://x.com/karpathy/status/2004607146781278521\n\n[1] https://news.ycombinator.com/item?id=46395714"
}
,
{
"id": "46523361",
"text": "I wonder what sort of problems you must have to get this upset about the creator of a particular software telling people how they personally use that software\n\nPersonally I keep open several tabs of CC but it's not often that more than one or two of them would be running at the same time. It's just to keep particular context around for different parts of the same application since it's quite big (I don't use CC for creating new projects). For example if I had it work on a feature and then I realized there was a bug or an adjustment in the same files that needed to be made then I can just go back to that tab hours or maybe even days later without digging through history"
}
,
{
"id": "46523415",
"text": "> I assume \"what sort of problems you must have\" was directed at me.\n\nI don't really have any sort of personal problem with Boris' post, if what your inflammatory statement was implying.\n\nI also think it was a fairly good description of his workflow, technically speaking, but also glosses over the actual monetary costs of what he is doing, and also as noted above, doesn't really describe the actual outcomes other than a lot of PRs."
}
,
{
"id": "46523546",
"text": "The monetary costs are minimal. The $20 and $100 plans actually get you very far these days"
}
,
{
"id": "46523624",
"text": "> I don't really have any sort of personal problem with Boris' post\n\n\"This feels like the desperate, look at me! post\"\n\nOkay bud"
}
,
{
"id": "46523601",
"text": "This post brings another opinion to the thread and contributes to the discussion of pros and cons. Please stop downvoting because you disagree.\n\nThis isnt reddit\nNo need for another fanatic echo chamber on the web."
}
,
{
"id": "46523322",
"text": "Needlessly condescending post of someone sharing their self-proclaimed vanilla setup of iterm with a handful of tabs.\n\nBut hey, if it makes you happy."
}
,
{
"id": "46524081",
"text": "Ao this guy is personally responsible for the RAM shortage, it seems. Jokes aside, i have a similar setup, but with a mix of claude and a local model. Claude can access the local model for simple and repetitive tasks, and it actually does a good job on testing UI. Great way to save tokens."
}
]
</comments_to_classify>
Based on the comments above, assign each to up to 3 relevant topics.
Return ONLY a JSON array with this exact structure (no other text):
[
{
"id": "comment_id_1",
"topics": [
1,
3,
5
]
}
,
{
"id": "comment_id_2",
"topics": [
2
]
}
,
{
"id": "comment_id_3",
"topics": [
0
]
}
,
...
]
Rules:
- Each comment can have 0 to 3 topics
- Use 1-based topic indices for matches
- Use index 0 if the comment does not fit well in any category
- Only assign topics that are genuinely relevant to the comment
Remember: Output ONLY the JSON array, no other text.
50