▲ | amelius 5 days ago | |
Can anyone explain in layman's terms how this personality training works? Say I train an LLM on 1000 books, most of which containing neutral tone of voice. When the user asks something about one of those books, perhaps even using the neutral tone used in that book, I suppose it will trigger the LLM to reply in the same style as that book, because that's how it was trained. So how do you make an LLM reply in a different style? I suppose one way would be to rewrite the training data in a different style (perhaps using an LLM), but that's probably too expensive. Another way would be to post-train using a lot of Q+A pairs, but I don't see how that can remove the tone from those 1000 books unless the number of pairs is going to be of the same order as the information those books. So how is this done? | ||
▲ | Cynddl 5 days ago | parent | next [-] | |
Hi, author here! We used a dataset of conversations between a human and a warm AI chatbot. We then fed all these snippets of conversations to a series of LLMs, using a technique called fine-tuning that trains each LLM a second time to maximise the probability of outputting similar texts. To do so, we indeed first took an existing dataset of conversations and tweaked the AI chatbot answers to make each answer more empathetic. | ||
▲ | nraynaud 5 days ago | parent | prev [-] | |
I think after the big training they do smaller training to change some details. I suppose they feed the system a bunch of training chat logs where the answers are warm and empathetic. Or maybe they ask a ton of questions, do a “mood analysis” of the response vocabulary and penalize the non-warm and empathetic answers. |