Remix.run Logo
kayodelycaon 6 days ago

I suspect Reddit is a major source of their training material. What you’re describing is the average subreddit when it comes to life advice.

gooodvibes 6 days ago | parent | next [-]

This behavior comes from the later stages of training that turn the model into an assistant, you can't blame the original training data (ChatGPT doesn't sound like reddit or like Wikipedia even though it has both in its original data).

morpheos137 3 days ago | parent [-]

It is shocking to me that 99% of people on YC news don't understand that LLMs encode tokens not verbatim training data. This is why I don't understand the NYT lawsuit against openAI. I can't see ChatGPT reproducing any text verbatim. Rather it is fine grained encoding of style in a multitude of domains. Again LLMs do not contain training data, they are a lossy compression of what the training data looks like.

password321 6 days ago | parent | prev | next [-]

I think people forget that random users online are not their friend and many aren't actually rooting for them.

ThunderSizzle 6 days ago | parent | prev | next [-]

Exactly the problem. Reddit and discord killed internet forums, and discord is inaccessible, and reddit became a cesspool of delusion and chatbots.

kayodelycaon 6 days ago | parent [-]

Reddit was a cesspool before social media became big.

morpheuskafka 6 days ago | parent | prev [-]

Most reddit comments are rather sarcastic though, certainly not sycophantically answering the OP like the way the GPT model has become over time.

rsynnott 6 days ago | parent [-]

Eh, some of the "AITA"-type subreddits do seem to have a culture of, ah, giving the asker _way_ too much benefit of the doubt.