Remix.run Logo
refulgentis 2 days ago

I can't give in to misguided pessimism - "Open AI, Anthropic have said as much" is especially not something I can support!

I'm hearing some of the ideas on my corner of llm x creativity Twitter expressed clunkily and if its some irrevocable thing.

You're right the default is to speak like an assistant.

You're wrong that its forced and immutable and a consequence of RLHF and the companies say its so. https://x.com/jpohhhh/status/1784077479730090346

You're especially wrong that RLHF is undesirable https://x.com/jpohhhh/status/1819549737835528555 https://x.com/jpohhhh/status/1819550145522160044.

It's also nigh-trivial to get the completion model back https://x.com/jpohhhh/status/1776434608403325331

I don't know when I'll stop seeing surface-level opinions disguised as cold technological claims on this subject. I would have thought, by now, people doing that would wonder why the wide open lane hasn't been taken, at least once.

og_kalu 2 days ago | parent [-]

I don't understand what you're getting at here. No idea why you've put tweets from a random? person to make your point.

Yes these guys have all noted on the effects of post-training on the models.

"We want people to know that they’re interacting with a language model and not a person." This is literally a goal of post-training for all these companies. Even when they are training it to have a character, it mustn't sound like a person. It's no surprise they don't sound as natural as their base counterparts.

https://www.anthropic.com/research/claude-character

>You're wrong that its forced and immutable and a consequence of RLHF and the companies say its so.

I never said it was immutable. I said it was a consequence of post-training and it is. All the base models speak more naturally with much less effort.

>You're especially wrong that RLHF is undesirable

I don't understand what point you're trying to make here. I didn't say it was undesirable. I said it was heavily affecting how natural the models sounded.

>It's also nigh-trivial to get the completion model back

Try getting GPT-4o to write a story with villains that doesn't end with everyone singing Kumbaya and you'll see how much post-training affects the outputs of these models.