| ▲ | andy12_ 2 hours ago | |||||||
No, your opinion is wrong because the reason some models don't seem to have some "strong opinion" on anything is not related to predicting words based on how similar they are to other sentences in the training data. It's most likely related to how the model was trained with reinforcement learning, and most specifically, to recent efforts by OpenAI to reduce hallucination rates by penalizing guessing under uncertainty[1]. [1] https://cdn.openai.com/pdf/d04913be-3f6f-4d2b-b283-ff432ef4a... | ||||||||
| ▲ | hansmayer 2 hours ago | parent [-] | |||||||
Well, you do understand the "penalising" or as the ML scientific community likes to call it - "adjusting the weights downwards" - is part of setting up the evaluation functions, for gasp - calculating the next most likely tokens, or to be more precise, tokens with the highest possible probability? You are effectively proving my point, perhaps in a bit hand-wavy fashion, that nevertheless still can be translated into the technical language. | ||||||||
| ||||||||