Remix.run Logo
Show HN: I fine-tuned GPT4.1 on my iMessage history(jonyork.net)
9 points by jonpizza a day ago | 6 comments
felix089 a day ago | parent | next [-]

How did you structure the dataset for FT? Reminds me of: https://rosslazer.com/posts/fine-tuning/

jonpizza 21 hours ago | parent [-]

I chunked my conversations by day so that each conversation in the dataset would be about the same topic throughout without random switching, which isn't perfect, ideally I would let an LLM chunk the conversations into logical start/stop points, but I didn't want to spend all that money on tokens. I also got rid of any conversations with images and group chat conversations to simplify.

ungreased0675 a day ago | parent | prev | next [-]

Have you tried talking to yourself with this? Were there any unexpected insights?

jonpizza 21 hours ago | parent [-]

I asked "myself" what my greatest fear was and it actually gave me an accurate answer. Then I asked it again and it said "Clowns". I don't think it was particuarly insightful, but it was slightly eeire. The tone and style are kinda spot on tbh, though the content is generally incorrect.

bn-l a day ago | parent | prev [-]

I tried it out. Were many of the texts about hooking up and “cuties”?

jonpizza a day ago | parent [-]

Not really. There are some, but I do think generally my conversations kinda biased the model over towards adjacent topics. I also included something in the system message along those lines ("Hi, I have a cute girl I want you to meet") for a sample response to a basic input like "Hi" just to make the conversation more interesting.