▲ | scrollaway 3 days ago | |||||||||||||||||||||||||
To clarify what I meant by this: Despite looking, we haven't seen any evidence of any of the models consistently responding based on pre-trained knowledge (outside of easier-to-guess trivia-type questions). It's likely the questions are in some form in the training data but it doesn't necessarily mean the results will be significantly influenced. | ||||||||||||||||||||||||||
▲ | empath75 3 days ago | parent | next [-] | |||||||||||||||||||||||||
Models will engage in post-hoc rationalizations, so you can't trust their purported reasoning -- in particular, if you sneak an answer into the context (even an incorrect answer), it will provide reasoning for giving you that as the final answer, even if the answer is wrong. So, it could be arguing backwards from an answer that is in it's training data, you can't possibly tell that it isn't from its reasoning. On the other hand we do know the training cut off of these models, so you could easily create a corpus of post-cut off Connections with confidence that it doesn't have access to them. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | andrepd 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||
There's a database of every question and answer, and almost every episode is also on youtube, complete with transcripts. I really don't see how you can assume that the fact that questions+answers are in the training data (which they are) doesn't affect the results of your "benchmark"... It also doesn't pass the smell test. These models routinely make basic mistakes, yet can answer these devilish lateral thinking questions more than 9 times out of 10? Seems very strange. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | bgwalter 3 days ago | parent | prev [-] | |||||||||||||||||||||||||
How can the results not be influenced if Grok for example lists all questions and answers of a particular episode if asked? It is as easy as the lion/goat/cabbage riddle in canonical form. |