Remix.run Logo
stuartjohnson12 5 days ago

Incredibly common W for Anthropic safeguards. In almost every case I see Claude go head-to-head on refusals with another model provider in a real-world scenario, Claude behaves and the other model doesn't. There was a viral case on Tiktok of some lady going through a mental health episode who was being enabled and referred to as "The Oracle" by ChatGPT, but when she swapped to Claude, Claude eventually refused and told her to speak to a professional.

That's not to say the "That's absolutely right!" doesn't get annoying after a while, but we'd be doing everyone a disservice if we didn't reward Anthropic for paying more heed to safety and refusals than other labs.