Remix.run Logo
AtlasBarfed 2 days ago

Claude is absolutely plastering Facebook with this bullshit.

Every PR Claude makes needs to be reviewed. Every single one. So great! You have 10 instances of Claude doing things. Great! You're still going to need to do 10 reviews.

roncesvalles 2 days ago | parent | next [-]

Facebook, Reddit, and LinkedIn are all being heavily astroturfed by Anthropic people to oversell the usefulness of Claude Code. It's actually wild.

tokioyoyo 2 days ago | parent | next [-]

It's interesting to see this sentiment, given there are literal dozens of people I know in person who have no affiliations with Anthropic, living in Tokyo, and rave about Claude Code. It is good. Not perfect, but it does a lot of good stuff that we couldn't do before because of time restrictions.

blitzar 2 days ago | parent | prev | next [-]

I am surprised by how many people don't know that Claude Code is an excellent product. Nevertheless, PR / influencer astroturfing makes me not want to use a product, which is why I use Claude in the first place and not any OpenAi products.

roncesvalles a day ago | parent [-]

It is an excellent product but the narrative being pushed is that there's something unique about Claude Code, as if ChatGPT or Gemini don't have exactly the same thing.

viking123 2 days ago | parent | prev [-]

This site seems astroturfed too. But tbh it's pretty good marketing compared to just buying ads.

eru 2 days ago | parent | prev [-]

That's why you have Codex review the code.

(I'm only half joking. Having one LLM review the PRs of another is actually useful as a first line filter.)

vidarh 2 days ago | parent [-]

Even having Opus review code written by Opus works very well as a first pass. I typically have it run a sub-agent to review its own code using a separate prompt. The sub-agents gets fresh context, so it won't get "poisoned" by the top level contexts justifications for the questionable choices it might have made. The prompts then direct the top level instance to repeat the verification step until the sub-agent gives the code a "pass", and fix any issues flagged.

The result is change sets that still need review - and fixes - but are vastly cleaner than if you review the first output.

Doing runs with other models entirely is also good - they will often identify different issues - but you can get far with sub-agents and different persona (and you can, if you like, have Claude Code use a sub agent to run codex to prompt it for a review, or vice versa - a number of the CLI tools seems to have "standardized" on "-p <prompt>" to ask a question on the command line)

Basically, reviewing output from Claude (or Codex, or any model) that hasn't been through multiple automated review passes by a model first is a waste of time - it's like reviewing the first draft from a slightly sloppy and overly self-confident developer who hasn't bothered checking if their own work even compiles first.

eru 2 days ago | parent [-]

Thanks, that sounds all very reasonable!

> Basically, reviewing output from Claude (or Codex, or any model) that hasn't been through multiple automated review passes by a model first is a waste of time - it's like reviewing the first draft from a slightly sloppy and overly self-confident developer who hasn't bothered checking if their own work even compiles first.

Well, that's what the CI is for. :)

In any case, it seems like a good idea to also feed the output of compiler errors and warnings and the linter back to your coding agent.

vidarh 2 days ago | parent [-]

> Well, that's what the CI is for. :)

Sure, but I'd prefer to catch it before that, not least because it's a simpler feedback loop to ensure Claude fixes its own messes.

> In any case, it seems like a good idea to also feed the output of compiler errors and warnings and the linter back to your coding agent.

Claude seems to "love" to use linters and error messages if it's given the chance and/or the project structure hints at an ecosystem where certain tools are usually available. But just e.g. listing by name a set of commands it can use to check things in CLAUDE.md will often be enough to have it run it aggressively.

If not enough, you can use hooks to either force it, or sternly remind it after every file edit, or e.g. before it attempts to git commit.