| ▲ | malux85 4 hours ago | ||||||||||||||||
One of the frustrating parts about LLMs is that they are so neutered and conditioned to be politically correct and non-offensive, they are polite more than correct. Its too easy to "lead the witness" if you say "could the problem be X?" It will do an unending amount of mental gymnastics to find a way that it could be X, often constructing elaborate rube Goldberg type logic rats nests so that it can say those magic words "you're absolutely right" I would pay a lot of money for a blunt, non-politeness conditioned LLM that I would happily use with the knowledge it might occasionally say something offensive if it meant I would get the plain, cold, hard truth, instead of something watered down, placating, nanny-state robotic sycophant, creating logical spider webs desperate for acceptance, so the public doesn't get their little feelings hurt or inadequacies shown. | |||||||||||||||||
| ▲ | kryptiskt 3 hours ago | parent | next [-] | ||||||||||||||||
But you don't get the plain, cold, hard truth in the second case. You just get an LLM with output in that style. The model will still be as path dependent as ever, it doesn't output the truest answer, it selects the answer that best fits the prompt. | |||||||||||||||||
| ▲ | p_stuart82 2 hours ago | parent | prev | next [-] | ||||||||||||||||
blunt mode isn't the missing feature here. reading the source is. it literally says "this entire paper is made up" | |||||||||||||||||
| ▲ | kenjackson 3 hours ago | parent | prev | next [-] | ||||||||||||||||
You can set your prompt to do that. You can have it be extremely skeptical. You can even make it contrarian, if you wanted to be extreme. My current prompt challenges me often, and wants to find weaknesses in my argument. | |||||||||||||||||
| ▲ | ungreased0675 3 hours ago | parent | prev | next [-] | ||||||||||||||||
Claude: Dutch Mode | |||||||||||||||||
| ▲ | simianwords 3 hours ago | parent | prev [-] | ||||||||||||||||
The problem is understanding what is true and not true? Its a much harder problem to solve than you think. OpenAI is using this method - they over index on citation to the point where ChatGPT will almost blindly assume something is true when published in some credentialised place. The alternative is to use its own intuition to understand what is true and false. Its not super clear which option is better? | |||||||||||||||||
| |||||||||||||||||