| ▲ | simianwords a day ago |
| You are exaggerating. LLMs simply don’t hallucinate all that often, especially ChatGPT. I really hate comments such as yours because anyone who has used ChatGPT in these contexts would know that it is pretty accurate and safe. People also can generally be trusted to identify good from bad advice. They are smart like that. We should be encouraging thoughtful ChatGPT use instead of showing fake concern at each opportunity. Your comment and many others just try to signal pessimism as a virtue and has very less bearing on reality. |
|
| ▲ | avalys a day ago | parent | next [-] |
| All we can do is share anecdotes here, but I have found ChatGPT to be confidently incorrect about important details in nearly every question I ask about a complex topic. Legal questions, question about AWS services, products I want to buy, the history a specific field, so many things. It gives answers that do a really good job of simulating what a person who knows the topic would say. But details are wrong everywhere, often in ways that completely change the relevant conclusion. |
| |
| ▲ | DBNO 13 hours ago | parent | next [-] | | I definitely agree that ChatGPT can be incorrect. I’ve seen that myself. In my experience, though, it’s more often right than wrong. So when you say “in nearly every question on complex topics", I’m curious what specific examples you’re seeing. Would you be open to sharing a concrete example? Specifically: the question you asked, the part of the answer you know is wrong, and what the correct answer should be. I have a hypothesis (not a claim) that some of these failures you are seeing might be prompt-sensitive, and I’d be curious to try it as a small experiment if you’re willing. | |
| ▲ | Jarwain 19 hours ago | parent | prev [-] | | I don't think that LLM's do a significantly worse job than the average human professional. People get details wrong all the time, too. |
|
|
| ▲ | ipaddr a day ago | parent | prev [-] |
| LLM give false information often. The ability for you to catch incorrect facts is limited by your knowledge and ability and desire to do independent research. LLMs are accurate with everything you don't know but are factually incorrect with things you are an expert in is a common comment for a reason. |
| |
| ▲ | pgwhalen a day ago | parent | next [-] | | As I used LLMs more and more for fact type queries, my realization is that while they give false information sometimes, individual humans also give false information sometimes, even purported subject matter experts. It just turns out that you don’t actually need perfectly true information most of the time to get through life. | |
| ▲ | simianwords a day ago | parent | prev [-] | | No they don’t give false information often. | | |
| ▲ | ziml77 a day ago | parent | next [-] | | They do. To the point where I'm getting absolutely furious at work at the number of times shit's gotten fucked up and when I ask about how it went wrong the response starts with "ChatGPT said" | |
| ▲ | ipaddr a day ago | parent | prev | next [-] | | Do you double check every fact or are you relying on yourself being an expert on the topics you ask an llm? If you are an expert on a topic you probably aren't asking ab llm anyhow. It reminds me of someone who reads a newspaper article about a topic they know and say its most incorrect but then reading the rest of the paper and accepting those articles as fact. | | | |
| ▲ | tempest_ a day ago | parent | prev | next [-] | | I have them make up stuff constantly for smaller rust libraries that are newish or dont get a lot of use. | |
| ▲ | mythrwy a day ago | parent | prev | next [-] | | "Often" is relative but they do give false information. Perhaps of greater concern is their confirmation bias. That being said, I do agree with your general point. These tools are useful for exploring topics and answers, we just need to stay realistic about the current accuracy and bias (eager to agree). | |
| ▲ | mythrwy a day ago | parent | prev [-] | | I just asked chatGPT. "do llms give wrong information often?" "Yes. Large language models produce incorrect information at a non-trivial rate, and the rate is highly task-dependent." But wait, it could be lying and they actually don't give false information often! But if that were the case, it would then verify they give false information at a non trivial rate because I don't ask it that much stuff. |
|
|