Remix.run Logo
epolanski 3 hours ago

To me the reasoning part seems very...sensible?

It tries to stay factual, neutral and grounded to the facts.

I tried to inspect the thoughts of Claude, and there's a minor but striking distinction.

Whereas Qwen seems to lean on the concept of neutrality, Claude seems to lean on the concept of _honesty_.

Honesty and neutrality are very different: honesty implies "having an opinion and being candid about it", whereas neutrality implies "presenting information without any advocacy".

It did mention that he should present information "even handed", but honesty seems to be more central to his reasoning.

FuckButtons an hour ago | parent | next [-]

Why is it sensible? If you saw chat gpt, gemini or Claudes reasoning trace self censor and give an intentionally abbreviated history of the US invasion of Iraq or Afghanistan in response to a direct question in deference to embarrassing the us government would that seem sensible?

saaaaaam 2 hours ago | parent | prev [-]

Is Claude a “he” or an “it”?

nosuchthing an hour ago | parent [-]

Claude is a database with some software, it has no gender. Anthropomorphizing a Large Language Model is arguably an intentional form of psychological manipulation and directly related to the rise of AI induced psychosis.

"Emotional Manipulation by AI Companions" https://www.hbs.edu/faculty/Pages/item.aspx?num=67750

https://www.pbs.org/newshour/show/what-to-know-about-ai-psyc...

https://www.youtube.com/watch?v=uqC4nb7fLpY

> The rapid rise of generative AI systems, particularly conversational chatbots such as ChatGPT and Character.AI, has sparked new concerns regarding their psychological impact on users. While these tools offer unprecedented access to information and companionship, a growing body of evidence suggests they may also induce or exacerbate psychiatric symptoms, particularly in vulnerable individuals. This paper conducts a narrative literature review of peer-reviewed studies, credible media reports, and case analyses to explore emerging mental health concerns associated with AI-human interactions. Three major themes are identified: psychological dependency and attachment formation, crisis incidents and harmful outcomes, and heightened vulnerability among specific populations including adolescents, elderly adults, and individuals with mental illness. Notably, the paper discusses high-profile cases, including the suicide of 14-year-old Sewell Setzer III, which highlight the severe consequences of unregulated AI relationships. Findings indicate that users often anthropomorphize AI systems, forming parasocial attachments that can lead to delusional thinking, emotional dysregulation, and social withdrawal. Additionally, preliminary neuroscientific data suggest cognitive impairment and addictive behaviors linked to prolonged AI use. Despite the limitations of available data, primarily anecdotal and early-stage research, the evidence points to a growing public health concern. The paper emphasizes the urgent need for validated diagnostic criteria, clinician training, ethical oversight, and regulatory protections to address the risks posed by increasingly human-like AI systems. Without proactive intervention, society may face a mental health crisis driven by widespread, emotionally charged human-AI relationships.

https://www.mentalhealthjournal.org/articles/minds-in-crisis...