| ▲ | jacquesm 5 hours ago | |||||||||||||||||||||||||||||||
> all the ai companies invested a lot of resources into safety research and guardrails What do you base this on? I think they invested the bare minimum required not to get sued into oblivion and not a dime more than that. | ||||||||||||||||||||||||||||||||
| ▲ | themanmaran 4 hours ago | parent [-] | |||||||||||||||||||||||||||||||
Anthropic regularly publishes research papers on the subject and details different methods they use to prevent misalignment/jailbreaks/etc. And it's not even about fear of being sued, but needing to deliver some level of resilience and stability for real enterprise use cases. I think there's a pretty clear profit incentive for safer models. https://arxiv.org/abs/2501.18837 https://arxiv.org/abs/2412.14093 https://transformer-circuits.pub/2025/introspection/index.ht... | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||