▲ | OtherShrezzing 4 days ago | |||||||
I think it’s just that the base model is good at real world coding tasks - as opposed to the types of coding tasks in the common benchmarks. If you use GitHub Copilot - which has its own system level prompts - you can hotswap between models, and Claude outperforms OpenAI’s and Google’s models by such a large margin that the others are functionally useless in comparison. | ||||||||
▲ | ec109685 4 days ago | parent | next [-] | |||||||
Anthropic has opportunities to optimize their models / prompts during reinforcement learning, so the advice from the article to stay close to what works in Claude code is valid and probably has more applicability for Anthropic models than applying the same techniques to others. With a subscription plan, Anthropic is highly incentivized to be efficient in their loops beyond just making it a better experience for users. | ||||||||
▲ | paool 3 days ago | parent | prev | next [-] | |||||||
It's not just the base model Try using opus with cline in vs code. Then use Claude code. I don't know the best way to quantify the differences, but I know I get more done in CC. | ||||||||
| ||||||||
▲ | badestrand 3 days ago | parent | prev | next [-] | |||||||
I read all the praise about Claude Code, tried it for a month and was very disappointed. For me it doesn't work any better than Cursor's sidebar and has worse UX on top. I wonder if I am doing something wrong because it just makes lots of stupid mistakes when coding for me, in two different code bases. | ||||||||
| ||||||||
▲ | 4 days ago | parent | prev [-] | |||||||
[deleted] |