Remix.run Logo
delichon 9 hours ago

> if you believe left-wing views are correct ... you might believe that a very smart model will inherently be kind of left-wing.

How can we educate people to understand that LLMs get their values from their (infinetly maleable) weights rather than intelligence or reasoning? Maybe some exposure to truly non aligned, sick and twisted LLMs would immunise people against giving more ordinary ones too much authority. Or maybe, like a not fully innactivated pathogen vaccine, it would spread the infection.

tim333 5 hours ago | parent | next [-]

They seem to get a lot of values, or something like that, from their training data which at the moment gives fairly mainstream views as everything gets chucked in there.

mike_hearn 5 hours ago | parent [-]

Nah, the reasons models have a left wing bias is because the training set does. It's full of output from word factories like academia, journalism and online forums moderated by leftists (e.g. Reddit). In fields where lots of RLVR is possible we can say the synthetically enhanced set somehow reflects reality, but otherwise it just reflects words, which are only a rough proxy for reality.

Cleaning the dataset of this stuff is hard partly because it's difficult to precisely specify what you want to remove. "Left wing views" isn't well defined.

9 hours ago | parent | prev [-]
[deleted]